As explained elsewhere, the user creates a class for their DP model. The class should declare a static
data member for each variable of these types. New objects of the appropriate class are created and then assigned to the variables while the model is being built: after calling Initialize() and before calling CreateSpaces().
Creating an object for a variable does not add the variable to the model. It must be added to the model after being created by calling the corresponding function during the build phase:
The action vector α is the vector of discrete actions chosen conditional on the state of the DP. A generic element of α is denoted a. An action variable takes on a.N different values. The action vector includes α.N variables, denoted a0 through a(α.N)‾.
Action variables are represented as an ActionVariable and are added to MyModel
and to α using Actions(). Calls to Actions()
and other model-building routines is part of MyCode
that user writes unless the class they are using includes the definition of the action variables.
In the basic definition of a DDP model, A is the set of possible actions. In DDP the set of actions is built up by adding action variables to α. So A emerges from the properties of the variables added to α.
The Possible Actions Matrix A is the matrix of all possible action vectors. It is the Cartesian product of the actions added to α. The word possible is used instead of feasible, because this notion of A is constructed mechanically from properties of the actions. It has nothing to do with the interpretation of the actions. Feasible actions are defined below.
An action refers to a particular value of the action vector α, which in turn is a row of A. An action variable refers to a column of A, but note that values of an action variable are repeated within the column. DDP constructs A when MyCode
calls CreateSpaces()
.
Discrete choice is described in two ways other than the one above used in DDP. For example, suppose there are two action variables and 6 total action vectors: α=(a0,a1)a0.N=3a1.N=2α.D=6 Some papers would treat this as a single action with six values. And/or papers may define a vector of 6 indicator variables to denote which choice was made: di=I{a=i}.
The three different approaches to coding discrete choices are illustrate in this table:
DDP Approach | Single Action | Indicator Vectors a0 a1 | a | d0 d1 d2 d3 d4 d5 ----------------------+---------------+---------------------------- 0 0 | 0 | 1 0 0 0 0 0 1 0 | 1 | 0 1 0 0 0 0 2 0 | 2 | 0 0 1 0 0 0 0 1 | 3 | 0 0 0 1 0 0 1 1 | 4 | 0 0 0 0 1 0 2 1 | 5 | 0 0 0 0 0 1
One reason to use an action vector like α is that each variable can be interpreted as a dimension of choice. There is no obvious interpretation of a=3
in the table, and the interpretation would change with the dimensions of the action variables. This approach would force MyModel
to decode into an action vector to make their code look like a mathematical model.
Another reason to use the DDP approach: the action vector approach makes it natural to impose feasibility conditions on the choice set, as discussed below.
Indicator vectors make it possible to write any utility as a sum: U=∑kdkuk. But as with the single action approach the interpretation is not obvious and U() will in no way resemble usual mathematical notation for an objective.
Usually an action variable can be an object of the base ActionVariable class. The user rarely needs to define a derived class.
MyModel
to handle/process the full matrix of feasible actions in utility and transitions. For this reason, MyModel
will rarely if ever access the current value of an action, a.v
. The vector of values of an action variable a is returned by CV(a).
Optionally, the user can create a class derived from ActionVariable and supply a Update function with it. This allows the user to associate a meaningful value for the numbers 0 to N‾, stored as a.actual
. DDP updates an action each time the problem is resolved, so actual values can depend on Parameter values. AV(a) will return the actual values of a
that correspond to each feasible action vector α. The default virtual Update() simply sets actual to the vector < 0 : N-1 >.
Update()
Unless you want to provide actual values for an action you do not need to create a derived class. Let a denote a discrete choice of hours to work each week. The maximum number of hours, H, depends on a parameter which can differ across individuals but is constant across the state space.
struct Hours : ActionVariable { decl H; Hours(const N,const H); Update(); } Hours::Hours(const N,const H) { this.H = H; ActionVariable("h",N); } Hours::Update() { actual = AV(H)* (vals / N); } ⋮ MyModel::MaxH() { return CV(hasch) ? 30 : 50; } ⋮ MyModel::Initialize() { ⋮ GroupVariables(hasch = new FixedEffect("",2)); Actions(h = new Hours(5,MyModel::MaxH)); ⋮ } MyModel::Utility() { return CV(wage)*CV(h); }
A is defined as the set of possible actions, built up by adding action variables to α. All possible actions is the matrix:
A≡×j=0,…,α.N−1{0,1,…,aj.N−1}
An action vector α is a row of the matrix A. An action variable is a column of A. Its position is the property a.pos
.
Feasibility is a property MyModel
imposes on α. The model may say that some logically possible actions cannot be chosen because they do not make sense given the interpretation of α. Or some actions are ruled infeasible for convenience to avoid calculations that are relatively unimportant. We can write feasibility as a property of the state. Or it can be written as a function of the state, which is more standard. Using both versions, the feasible actions at θ is a matrix property:
∀θ∈Θ,A(θ)≡θ.A⊆A.
FeasibleActions(): by default all possible actions are feasible unless MyModel
says otherwise. A(θ)≡A for all endogenous states. This default assumption is produced by including a built-in virtual method, Bellman::FeasibleActions()
. meaning that MyModel
can replace it with its own version. This is explained in Bellman.
CreateSpaces constructs the possible matrix A, stored as the first element of a list: A ↔ A[0]
. It then calls FeasibleActions(A)
for each reachable state θ. Every time a different value is returned by FeasibleActions()
a new matrix is added to the A list. The index into the list is then stored as Aind. That is, there is not a different matrix at each θ. Instead there is a list of different feasible matrices and A(θ) is really an index into that list.
The user's (optional) feasible action method should determine feasibility using the action variables stored in the derived DP model. So if a
contains an ActionVariable
its value are available as CV(a)
. In short, θ.A ↔ A[Aind]
. If MyModel
does not specify feasible actions, then Aind = \(\theta\).j = 0
for all states.
The A list contains the actual values not the current values. If the user provides an Update() for a derived class of action variable, then A[Aind]
is the matrix of actual values. Since the default is that actual values equal current values (a.actual[a.v] = a.v
), then A[Aind]
is always possible to use. The uncoded list of matrices is available as Matrix.
Autonomous
Coevolving
Most aspects of DP models can be handled by autonomous variables because the transition can depend on current values of other variables. Only certain kinds of correlations between two or more variables require coevolving variables in a block.
The model places discrete state variables/blocks into one of four vectors.
IID Transitions | General Transition | No Transitions (Group Variables) | |||
---|---|---|---|---|---|
Exogenous (ϵ) | Semi-Exogenous (η) | Endogenous (θ) | RandomEffect (γr) | FixedEffect (γf) | |
Generic Element | e | h | q | r | f |
Form of Ps | Ps(s′) | Ps(s′|α,ϵ,η,θ) | g′=g | ||
Explanation | IID, exogenous to everything. | Transition can depend on current states and actions. | value is fixed for this problem. | ||
Special form of Pq, q∈θ, q≠s | cannot enter Pq | can enter transition of other variables | can enter but see SetUpdateTime | ||
Explanation | q independent of ϵ, given α | η and θ can affect q | default r can only enter U() | can enter Pq | |
Computing Implication | Compute & store Pϵ and Pη once on each solution. | Compute and store Pθ at each θ, store as a matrix in α and η | Store P⋆(α|…γr) | Reuse space for each fixed effect. |
class MyModel : DPparent { ⋮ static decl e, q; ⋮ static Initialize(); } ⋮ MyModel::Initialize() { DPparent::Initialize(new MyModel()); ⋮ ExogenousStates(e = new SimpleJump("iid",5)); EndogenousStates(q = new LaggedState("lag",e)); ⋮ CreateSpaces(); }
s
has marginal transition probabilities that enter the overall transition probability as multiplicative factors, conditional on current actions, states and parameters. This is in contrast to Coevolving state variables which are part of a StateBlock. The difference with an Autonomous state is that its transition is jointly distributed with one or more other Coevolving states.
Some derived classes of state variables play special roles in DDP. Two major ones are TimeVariables which serve as the DDP clock and TimeInvariants which track variables that are fixed during a DDP and stored in the γ vector. DDP will re-solve the model for each value of γ. Your model may include state variables that are already defined in DDP but require some changes to accommodate the environment. The next section discusses how to create a new derived class for your state variable, but it may be possible to avoid that using the Augmented state variable feature.
Example: Suppose your model has a state variable q
which is of the RandomUpDown() class. That is, each period q
may go up or down by one unit or remain unchanged depending choices and other state variables. However, when another state variable h
goes from 0 to 1 you want the value of q
to become constant and not subject to the RandomUpDown transition. That is, you want to Freeze the current value of q
when h
is 1.
class MyModel : DPparent { ⋮ static decl a, h, q; … static Initialize(); } ⋮ MyModel::Initialize() { DPparent::Initialize(new MyModel()); ⋮ Actions(a = new ActionVariable("a",2) ) ; EndogenousStates( h = new PermanentChoice("h",a); q = new Freeze(new RandomUpDown("q",10,<0.5;0.5>),h); ⋮ CreateSpaces(); }
new
state variable is passed to the Freeze() constructor. In essence this underlying random variable is hidden from the model, which only sees the variables passed to EndogenousStates. What the model sees is the augmented variable q
which acts like a RandomUpDown variable unless h=1
. As always, the other state variables or actions that q
depends on have to be passed to its creator (or constructor).
In fact, Freeze is a special case of a more general type of Augmented state variable, namely a ValueTriggered() state variable.
Below is an example of how to define a new state variable is provided. It shows both a very basic coding which can get the job done but which might be limiting in some ways. So a second version of the code is shown which is more robust. Several advanced elements of the Ox programming language determine how and why you create a state variable this way. These features may be confusing to someone just starting to program in Ox, especially with no prior experience with object-oriented programming in an interpreted and dynamically-typed language.
«VarName»
for whatever name you chose.
niqlow/templates/DynamicPrograms/StateVariable.ox
to a file called «VarName».ox
. Source: niqlow/templates/DynamicPrograms/StateVariable.ox
#include
or #import
. See the Ox documentation for more details. If you use #import "«VarName»"
then you must create two separate files, #import «VarName».h
and #import «VarName».ox
. If you use #include "«VarName».ox"
then the two parts of the code appear in one file.
If your variable is closely related to one of the pre-defined variables listed above then you might be able to choose that as your base class. This may (greatly) reduce the new code you have to write. If you start from scratch then choose as the base class either Random or NonRandom. These two categories have no effect, but allow DDP to summarize models that include your state variable more accurately. That is, DDP treats random and nonrandom state variables the same way.
class
(or struct
) has to know everything necessary to compute the transition of the state variable at any state.
L
and the number of different values it takes on, N
. The constructor for state variables is defined as StateVariable::StateVariable(const L, const N)
to ensure these values are set.
Transit()
and Update()
that compute the transition and update the actual discrete values associated with the state variable. The word optional is somewhat inaccurate, because every state variable must have those functions. The issue is whether your state variable needs their own versions of them or can they inherit the version from a parent class.
new
operator.
StateVariable()
if it is not derived from another class below Random and NonRandom). So this means you must declare and define a constructor for your state variable which is called by new
, if only to ensure that the parent constructor is called.
new
operator calls the constructor of a class. The delete
operator calls the destructor. Again, in Ox, neither of these needs to exist, but in DDP a state variable needs a constructor. Typically it will not need a destructor and here is why: State variables are presumed to exist the whole time the program is running. DDP does provide a method to clean up a model that is no longer needed to save memory and to reinitialize for a new DP problem. DDP will delete each state variable but it cannot remove dynamically allocated variables defined within classes. So if your state variable has any new
commands that are not paired with a delete
in your code, then yes, you should write a destructor which will delete these objects. However, the chance that not doing this creates a significant memory leak is small given the anticipated use of DDP.
Transit()
method
Transit()
is, in most cases, declared as a virtual method. So if your state variable has no transition of its own, the one from the parent variable will be called in its place. This is why it must be called Transit()
, because it replaces another version of a Transit()
that would be called if necessary. Transit
will often need to access matrix of current feasible actions at the current state, which is available in Alpha::C that is set every time the point in the state space is changed or initialized. Each row of Alpha::C
is an action vector α each column is an action variable a. Transit()
must report back the transitions of an instance of the state variable from a given current point in the state space.
Transit()
Transit()
might look like this:
Alpha::C i j 0 0 1 0 0 1 1 1
F
. That is, suppose one value, say 3, may be feasible only if a certain action is taken. Then 3 must be in F
, even if 3 is not feasible if another action were taken. The elements of F
do not need to be sorted. The other output is a matrix of transition probabilities, P
. The rows of the matrix must match the rows of Alpha::C
; the columns must match the columns of the row vector F
. The ij element of P
is the probability that your state variable takes on the jth value in F
given the agent chooses the ith action in the feasible set.
Transit()
might be sent using code that looks like this: return {F,P};
, presuming you have declared local variables F
and P
and filled them with the correct information in the correct format. Notice that the only explicit information sent to Transit()
is the feasible matrix. Yet it must send back the transition at a specific state in the state space. How does your code know what state is the current state? Further, how does your code know the current value of parameters which might affect the probabilities? The answer is that the value or, if the quantity is changing like the current state, the location of the value must be stored in a member of the state. The only way to get this information is through the constructor function discussed above.
To make the example look better, suppose you have decided to set «VarName» to be MyStateVar
. This will be used wherever the name of the new derived class belongs. In math notation we will refer to the new state variable as y. But in code y might not be descriptive enough. Instead, you might add y to the DP model with code that looks like this:
decl mystate; // in the class
definition of MyStateVar
⋮
mystate = new MyStateVar("y",4,…);
EndogenousStates(mystate);
The …is not literal. It is there because in this example the constructor
MyStateVar()
will require other arguments. The state variable is stored in an Ox variable called mystate
, but we will refer to it as y, which is the label given to it.
x
is an occupation code and the choice variable i
indicates a choice to work, then y
equals the occupation the person worked at last period if they did work. Otherwise y
should take on the value 0. Simply put, the value of y next period is y' = ix. First, the y process as the feasible state next period:
y' = ixNow, as a transition probability:
Prob(y' = z) = 1 if z = ix 0 if z ≠ ix.
MyStateVar
should be classified as a NonRandom state variable. This is a special case of an autonomous process because the probabilities do not depend on the next values of the other states, such as x'
. If that were the case, the user has to create a StateBlock which handles the transitions for multiple Coevolving variables. (DDP has no way of knowing for sure that the probabilities are always 0 and 1, at all states and for all possible parameter values, so that is why you manually categorize it as nonrandom because you know that the 0 and 1 are hardcoded into the DNAof the state variable.) Because the transition depends on the action chosen y is nonrandom but it is not deterministic. A deterministic random variable would be something like a seasonal cycle.
mystate
might have these values:mystate { .L = "y" .N = 4 .v = 0 .vals = <0,1,2,3> .pos = 1 ⋮ }That is, y takes on 4 different values (0,1,2,3). At the current state it happens to have value
v=0
. It also happens to be the second state variable in the state vector (pos=1
).
State Vector x y z t 2 0 3 8The variable
t
is the age variable in the finite horizon. Now we can represent the transition of y'
at this particular state as follows:
Alpha::C Prob(y') trim 0 cols → Prob(y') i j y'= 0 1 2 3 y'=0 2 ----------------------------------------------------- 0 0 1 0 0 0 1 0 1 0 0 0 1 0 0 1 0 1 1 0 0 0 1 0 1 1 0 0 1 0 0 1
y
takes on four different values, but given that x=2
, only two of those values are feasible next period, 0 or 2. The overall transition routine in DDP is smart enough to drop the columns of all zeros, but some computation can be reduced by trimming those columns inside Transit()
. Thus, we can focus on the trimmed 2-column representation of the transition probability.
Transit()
returns the vector y' and the matrix Prob(y').
1-i
. And the second column is simply i
.
Transit
. Transit()
know that x=2
currently? And how will it know that i
is in the first column of Alpha::C
not the second? This is where the object-oriented approach to building a dynamic programming model comes in handy compared to using vectors of values to represent the state vector. Transit()
will know about x
and i
because it will be constructed to know them as the following code illustrates:
MyStateVar
requires the user to send the instance of some state variable class that corresponds to x
. occup
to pointto the right variable. Because an argument and a member is both called
occup
we refer to the member with that name using the this.
operator. The constructor also asks for an instance of an action variable which corresponds to i and stored in work
. Once the members are stored in members they are available for use by any method of the class, including Transit()
.
mystate
.
occup
can take on. This property is always stored in the N
member, so the constructor can get mystate
's value from occup
: MyStateVar
inherits the number of different occupations directly from the argument occup
. So the new base constructor passes that value through to the base constructor called by the constructor of a class derived from StateVariable. Since NonRandom
is really a container it does not have its own constructor. So StateVariable()
is called directly.
mystate->Transit()
is called it will knowabout x and i, but how will it know about the current value they have?
v
member, which was shown above. Thus the current value of x equals occup.v
. In the code for Transit the expression 0~occup.v
generates, for the state above, the row vector <0 2>
. These are the only feasible values next period, or possible values of y'
. So the current value of x is not accessed from a vector of numbers, which might be the way DP code in FORTRAN might do. MyStateVar
class. It does not matter to MyStateVar
whether occup
has the label x. It does not matter that
occup
happens to be first in the state vector. It does matter, of course, that occup
is indeed a state variable, and it has to know this before anything is done in the model.
return
statement is simply Ox code to copy in 1-i
and i
. However, these are column vectors and they get their value from values from Alpha::C
.
Alpha::C
will i
be in?
work
which gets copied to the member with the same name. And DDP puts the position of the action in the action matrix in pos
. So whatever column it is any instance of MyStateVar
will know which one it is.
Knowing the feasible values and their transition probabilities for a single state variable is not sufficient to solve a DP problem. The values of this state variable combine the feasible values of all other state variables to determine the possible states next period give the current state. If the value function were stored as a multidimensional array (or multi-dimensional matrix, which is not possible in Ox) then this state's possible values would simply be inserted into the right index for V, as in V[x][y][z][a]
. However, in Ox this would be inefficient, and without knowing ahead of time how many states will be added to the model it is not possible to write the code. (There is no way in Ox or C to write something like V[x]...[a]
to allow a dynamic number of dimensions.)
DDP is designed to avoid this problem by storing V as a one dimensional vector regardless of the number of state variables. The state is stored as a vector (internally), and the size of a vector can be determined dynamically. Associated with the state vector is a vector of offsets or indices. Multiply the state vector and the offsets to determine the index of the state in the V vector. Each state variable has an offset which depends on the other state variables and the number of values they take on.
DDP handles the offset. The Transit()
function only has to return the list of feasible values and their probabilities. Actually, DDP keeps track of several offset vectors to index different aspects of the overall solution algorithm.
When you write the code for your state variable you force the user (possibly yourself) to provide information required to code the transition for y. But you are relying on DDP to process the state variables in the model, for example by looping through all the feasible current values of x and y. Further, DDP must know that i is one of the actions and include a column for it in the feasible matrix. This is done by using methods specific to the DP class for adding elements of the model. Namely, the EndogenousStates and Actions methods. So we can now complete the code that would add a variable in the model that corresponds to y and which can reliably follow the transition y' = ix.
MyStateVar
.
static decl i, x, mystate; x = new StateVariable("x",4); i = new ActionVariable("i",2); mystate = new MyStateVar("y",x,i); Actions(i); EndogenousStates(x,mystate);
MyStateVar
is called. Worse, since Ox is dynamically typed, and since it initializes static
variables at 0 it is possible for incorrect information sent to the constructor to mimic valid information.
isclass()
is very useful here, because it checks that the arguments are derived from the correct base class (or the correct intermediate class).
N-1
, at least when using the v
member.
actual
member of occup
, not v
.
N-1
but 1 to N
(or any other set of values), the ambiguity in coding ix can be resolved. A value of 0 for this state variable indicates the person did not work last period and starts out with no occupation. If there is an occupation actually coded as 0, then this state variable will infer that not working does not reset occupation. That is, occupation is determined by something else (such as the kinds of jobs the person chooses to look for). Note that by default> actual
is the same as v
, but if the state or action variable has its own Update() method it can
set the actual
codes associated with the internal values 0…N‾
.
Note that Ox assignment of arithmetic types and string type implies copying over the contents from the right-hand side to the left-hand side. Ox accesses an object through a reference to the object which is created using the new operator. An object is removed from memory using the delete operator (if there is no matching delete, the object will exist until the program terminates).
Source: niqlow/examples/classreference.ox
Source: niqlow/examples/classreference.txt
b=a
and then changing b
does not change the value of a
. However, changing the member v
does change the corresponding value of a
. In the first three assignments, Ox clones the right hand side of the assignment. But when assigning a variable that is currently an instance of a class a clone is not made. Instead, a reference or pointer to the class is made. So accessing the member of the reference is equivalent to accessing the member of the assigned class.
this.occup = occup;
does not duplicate the state variable passed as an argument. It creates a reference to it. MyStateVar
can mess up the x
variable, which should be controlled by DDP. But it also means that as DDP manipulates x
the other variable mystate
is informed automatically, without DDP needing to know that y
depends on the value of x
.
new
operator allows separate instances of MyStateVar
to be created and passed to different variables to track.occup
and work
are not declared as static
members, each instance has its own space for these variables. So they can be different values not pointers to the same space.
It has been emphasized a few times that the user's code does not need to handle state variables in vectors. So why is it required to handle actions stored as a matrix? Here the issue is computational speed within an interpreted language like Ox and Matlab. Namely, nested Ox loops induce a computational overhead that nested loops in compiled languages like C and FORTRAN do not.
The innermost loop of nearly all discrete dynamic programming solution algorithms is a choice over finite options. Requiring the user-defined utility()
and Transit()
methods to handle the matrix of feasible actions means this inner loop can be handled with native Ox matrix routines. Avoiding a layer of nested loops can speed up code considerably in an interpreted language like Ox.
|is not used in these notes, but here it is used to emphasize that f and g may be correlated if we do not observe or condition on current action and state values.
MyModel
might specify individual transitions of the form Ph(h′|h,age)Pw(w′|h′−h,w,age). This formulation could be used to rule out, for example, a growth spurt (h′>h) and significant weight loss (w′<w). This requires a sequence in the realizations: h′ is realized first and then its value feeds into the transition probabilities for weight. In some ways this sequencing of the realizations makes the model simpler to specify.
These instructions follow the ones provided for creating a StateVariable. They are simpler and more to the point, focussing on the distinct features of state blocks. Most of the additional considerations apply here as well but are not discussed again.
«BlockName»
for whatever name you chose.
niqlow/templates/DynamicPrograms/StateBlock.ox
to a file called «BlockName».ox
.Source: niqlow/templates/DynamicPrograms/StateBlock.ox
There are fewer predefined blocks than autonomous variables. So it is likely that StateBlock
will be the base for your class.
struct Vitals : StateBlock { decl wght, hght; Vitals(); Transit(); Update(); GrowthProbs(); }
optional methods that must be provided for each new state variable.
Transit()
and Update()
that compute the transition and update the actual discrete values associated with the state variable.
GrowthProb()
, which will be presumed to return a vector of transition
probabilities in a format described below.
Vitals::Vitals() { StateBlock("vitals"); wght = new Coevolving("w",10); hght = new Coevolving("h",6); AddToBlock(wght,hght); }
Transit()
methodTransit()
function returns a single row vector of "prime states" but a block must return a matrix of feasible "prime" states, each row is the state of the corresponding member of the block in the order they were added to the block. The transition probabilities still pertain to each column of the matrix. The rows of the transition probabilities correspond to the rows of the feasible actions θ.A.
Transit
can access the feasible actions \(\alpha\)
at the current state. in C.
F
: the values this block could take on next period. Each row corresponds to one of the variables in the block. Each column is a different outcome. This is what allows for correlated transitions within a block. The other output is a matrix of transition probabilities, P
. The rows of the matrix must match the rows of Alpha::C
; the columns must match the columns of the row vector F
. The ij element of P
is the probability that your state block takes on the vector of values in column j of F
given the agent chooses the ith action in the feasible set Alpha::C
.
reshape
to get the right dimensions:
Vitals::Transit() { decl F, P; F = wght.v-1 ~ wght.v ~ wght.v+1 ~ wght.v ~ wght.v+1 | hght.v ~ hght.v ~ hght.v ~ hght.v+1 ~ hght.v+1; P = GrowthProb(); return { F , reshape(P,Alpha::N,columns(F)) }; }So, if weight and height were currently at levels 2 and 3, respectively, then
F
would be:1 2 3 2 3 3 3 3 4 4;Notice that this transition matrix imposes the coevolving condition that weight cannot fall when height increases. There is no way to achieve this with two autonomous variables because their innovations are statistically independent.
GrowthProb()
will return a 1×5 vector of probabilities, corresponding to the five possible outcomes in F
.
reshape()
simply duplicates row-by-row since weight and height change do not depend on current actions.
wght.v-1
or hght.v+1
will be incorrect when the current values are near the bounds and this will eventually cause an error. The formula for F
works fine as long as both variables are away from their boundaries and it illustrates the key issue here of what a block transition looks like.
Update()
function for the block.
.v
values of each variable. If there are M variables in the block, each taking on am.N
values, then the actual matrix will be N ×(∏m am.N)
.
decl stats; ⋮In this case, the vital state process is exogenous to other state variables, but if it influences the transitions of other state variables then the block is only semi-exogenous.SemiExogenousStates(stats = new Vitals());
Since a StateBlock is a type of StateVariable it has a .v
data member. DDP ensures that it always contains the vector of values of the state variables, in the same order as they were sent to AddToBlock. And this means that CV() will return the vector of values for a block just as it returns .v
for a scalar state. In addition, AV() will return the vector of actual
values that correspond to the indices in .v
. That is, it will pull out of the actual
matrix the right value for each of the M variables at their current values.
MyModel
each generic element is a separate state variable.
P(ζ′,ϵ′,η′,θ′,γ′ | α,ζ,ϵ,η,θ,γ )=fζ(ζ′)×Pϵ(ϵ′)×Pη(η′)×Pθ(θ′ | α,η,θ)×I{γ′=γ}.
Pϵ(ϵ′)=∏k=0,…,ϵ.N−1Pek(e′k)Pη(η′)=∏k=0,…,η.N−Phk(h′k)Pθ(θ′|α,η,θ)=∏k=0,…,θ.N−Pq(q′k|α,η,θ).
Time is a key concept in dynamic programming. At least the difference between now (today) and later (tomorrow) underlies Bellman's equation. In a stationary environment that is all that matters while solving the model. In a non-stationary world the clock takes on more values. The primary assumption is that time progresses: t′≥t. This allows the value function to be solved backwards in time, saving storage or computation if, mistakenly the environment is assumed to be stationary. (That is, with stationarity a fixed point in all states must be solved at once, but with non-stationarity only a subset of the state space needs to be consider at each stage as we work backwards.)
So the two simple extremes are stationarity (today is followed by tomorrow which is the same as today) and normal aging: t′ = t+1 until t=T‾. However, timing can be more complicated that those two cases. One case is an agent facing a finite horizon problem and the possibility of early death.
The Clock Block is a single StateBlock that is always in the endogenous vector θ, and is always the rightmost element of it, in the sense that all task that span the endogenous state space will loop over time in the outermost loop.
SetClock()
must take place between the calls to Initialize()
and CreateSpaces()
Initialize(new MyModel()); SetClock(InfiniteHorizon); CreateSpaces();If
MyModel
does not set the clock explicitly, then a stationary infinite horizon clock is set by CreateSpaces().
t′ ≥ tWith anticipation (foresight), V(θ) can/should be solved backwards in t if time is important in the model beyond just today and tomorrow in an infinite horizon.
RandomMortality
clock described below, the next time may be either t+1
or T-1
if death occurs.
t
. However, no other time periods must be stored, so separate coding of the t
process and t″
process conserves memory in complex environments. Because it plays no direct role in the mathematics (as opposed to the computations), t″ is never listed as a member of θ, but it will be seen in output with the value 0. In more complex environments the clock may include other state variables whose values coevolve with t and t″.
MyModel
does not need to refer to it directly.MyModel
as t. That is,I::t ≡ counter.t.vSince
t
is in the index class I MyModel
can use the identifier t
for its own use.
counter.t.N
, also denoted T
, is the number of values that the time variable t
takes on. T ≡ t.N T = 1 for an infinite horizon model (T = ∞).
MyModel
stored as T
When T is finite, N::T ≡ T = counter.t.N, When T = ∞, N::T ≡ 1
N::T=1
) by checking the class of counter
.
Also see ClockTypes
InfiniteHorizon
: t″ = t = 0 = T‾.MyModel
only has to deal with today and the transitions of state variables.
Ergodic
Ergodic
DDP will compute the ergodic or stationary distribution across states, P∞(θ). If the user's state transitions are not themselves stationary then this calculation may fail.
SubPeriods
SubPeriods
tag creates a Divided clock.
s
, all state variables must be endogenous (added to θ). This is checked in CreateSpaces. The user should augment state variables using
NormalAging
: t′ = t+1, up to T‾; t″=0.StaticProgram
, with T&line;=0.
StaticProgram
is a tag associated with the class StaticP, which is derived from the class Aging, DDP cannot confuse this with a Stationary environment.
RandomMortality
: the agent either ages normally or dies before the start of the next period
t
is decremented the just-solved for values are placed where t″ = 0
will reach it. This means that t″=0
is typically associated with "ordinary" time evolution while other values are perturbations such as premature death of the agent. The mortality probability π() can constant or depend on the current state and current actions.
RandomAging
: the agent spends a random amount of time in each age bracket
UncertainLongevity
:
t=T‾
is still the case of death which is still random and occurs with probability π() as above. But now t=T‾-1
is now a stationary problem and t=T‾
is a terminal state. Otherwise, once t=T‾-1
today and tomorrow are the same. DDP iterates on the value function at t=T‾
as if it were a (non-ergodic) stationary problem, continuing until convergence. Then it will proceed backwards as with mortality. The advantage of this approach is that there is a single choice probability for this final phase (conditional on other state variables) rather than computing and storing slightly different choice probabilities as t
approaches T‾
.
RandomAging
which uses AgeBrackets for the state clock with random mortality. But it is not a special case of either one so it is derived as a third class from NonStationary.
SocialExperiment
: Phased treatment and random assignment
RegimeChange
SocialExperiment
except the unexpected environment lasts forever.
TRUE
. This does not itself check for convergence, and other considerations may set setPstar to TRUE.
Time invariants index different DP models (elements of γ). Time invariants are variables that take on different values across agents but are fixed for a given agent.
The basic DP model concerns a single agent in a single environment. However, many applications involve related DP problems that differ in parameter values, which in turn alter the primitives U(), P(), δ. The state vector γ holds variables that are fixed for an agent but differ across agents.
As with other state vectors, anything that can go in γ could be placed in θ. However, since the state does not vary it is inefficient to included invariant states in Θ.
Instead, DDP resuses Θ for each value of γ and stores only a minimum amount of information for previously solved models. Only state variables derived from the TimeInvariant class can be added to γ Invariants do not have a transition.
DDP distinguishes between two kinds of invariants. Each is either a FixedEffect or a RandomEffect. This distinction plays the same role as in panel models. Fixed effects typically correspond to constant observed variables, such as gender. So if the model is to be solved separately for men and women, with different primitives, then a binary FixedEffect would be added to γ.
On the other hand, a RandomEffect is designed to account for unobserved variation in the underlying problem. An example would be a model that allows agents to have different (unobserved and permanent) skill levels. A single skill variable taking on discrete values could be added to γ In estimation DDP will sum over the distribution of skills.
If a TimeInvariant class has a Transit()
function defined it is never called because the DP problem presumes the value will never change. A fixed effect also has no distribution, but a random effect does, Distribution().
The distribution is used to integrate out the random effect after solving for behavior conditional on each value. The distribution can depend on the value of the fixed effects. For example, the distribution of skills can depend on gender. This is called once for each value of the fixed effects and this updates the pdf(), a vector of probabilities or weights place on the values of the random effect.
Correlated random effects are created by adding a RandomEffectBlock to γ
Each combination of random and fixed effects implies a value of γ, and for each γ a Group node is created. The set of all group nodes is the group space, denoted Γ.
As with allowing state variables to be declared exogenous, moving time invariant variables from θ to γ saves time and especially storage while solving the model. An invariant does not need to be tracked during iteration over Θ. So group variable Transit()
methods are not called for each iteration over t and t″.
Storage is re-used while solving for different values of γ. DDP reuses the state space Θ for each value of γ. Choice probabilities P⋆(α|ϵ,η,θ,γ) are stored separably for each random value of γ. EV(θ) integrates out the random effects, conditional on the value of the fixed effects in γ. Both utility and a transitions can depend on the value of fixed effects. However, only utility can depend on random effects.
An auxiliary value x is typically a function of the state and action vectors that would be observed in the data or is of interest in itself. It is based on a class derived from AuxiliaryValue and is added to the list χ.
Elements of χ are user-defined auxiliary variables, sometimes referred to as payoff-relevant
variables in the DP literature. Auxiliary variables are functions of current states and actions (and not functions of past or future outcomes), so χ adds no additional information to the full outcome. That is, χ is redundant within Y*, whereas the other actions and states each contain information. Auxiliary variables are involved in partial observability of the realized DP.
Auxiliary values are added to the outcome (appended to χ) using AuxiliaryOutcomes(). The value of the variable is set in the method Realize()
, which is a replacement for virtual Realize().
Realize()
sets the value of v
given the current state and outcome. This value is then added to the realization. Auxiliary
variables are never referenced nor realized by niqlow while solving the DP problem, because they plays no role in the solution. They are realized only when simulating the solved model or when matching the model to external data. The user might realize values in the process of computing utility.
The user may break up Utility()
into functions that help determine it. These can be static methods of MyModel
. Then some of these functions may be observed in data and might enter econometric estimation. StaticAux is a auxiliary value class that can be used as a wrapper
for such functions as shown below.
Auxiliary values are also used to code indicators and interaction variables in data sets. For example, if a is an action to choose among 5 options the data may include four variables that are indicators for all but one of the options being observed, yk=I{a=k}. Instead of having to create a set of binary action variables to match this data the user can create a set of Indicators, each of which is an auxiliary outcome that will match whether a=k or not (as well as the probability of that occurring in a prediction data set).The Indicators() method can be used to generate all four indicators at once. Interactions and MultiInteractions between discrete actions and states can be created as can interactions between a auxiliary variable and a discrete quantity.
A model may have a utility of the form: U(α,θ)=u(x⋆;α,θ) where x⋆=argmaxxc(x;α,θ). That is, the agent makes a continuous choice over x conditional on the state and the discrete action α. This conditional continuous choice (
CondContChoice) then determines the indirect utility of the discrete actions. One special case would that u()=c(x⋆;α,θ): that is, the utility is simply the indirect utility over the continuous choice. (ASnother possibility is that x⋆ directly affects the transition as in P(θ′;x⋆,α,θ). Currently this is not supported. The user may be able to use the features below to help implement this.
A continuous choice is related to AuxiliaryValues
, in the sense that you may match x⋆ to observed data. If the continuous choice has a closed form then it can be represented as a static function and included in observations using the StaticAux wrapper. If the optimal choice requires sequential optimization then the CondContChoice objectitve can be used to carry this out efficiently. This allows for a single objective to represent each conditional choice and to minimize the amount of additional storage required.
struct Cost : CondContChoice { vfunc(); } Cost::Cost(x) { CondContChoice("mC",x); } Cost::vfunc() { v = -exp( x ); }struct MyModel : Bellman { decl xvals;
} MyModel::MyModel() {
} mCost = new CondContChoice("mC",x); x = new Positive("x",2.0); mCost = new CondContChoice("mC",x); mCost -> Algor( new BFGS(mCost) );
Volume
member.SILENT
.
Variables
log file.
Functions | ||
![]() |
Remove all transitions for which the transition probability is EXACTLY 0. | |
Enumerations | ||
![]() |
. |
Public fields | ||
![]() |
const | |
![]() |
||
Public methods | ||
![]() |
Create a binary endogenous absorbing state. | |
![]() |
s' = min( s+ s.a, s.N-). Ρ(s'=z | α,ε, θ,ψ) = I{z = min(s + s.a,s.N-) }.
Public fields | ||
![]() |
const | Variable to track |
Public methods | ||
![]() |
virtual |
decl tothrs = new ActionAccumulator("TotHrs",work,10); //track work up to 10 hours EndogenousStates(tothrs);
Public fields | ||
![]() |
||
![]() |
||
Public methods | ||
![]() |
Create a variable that counts how many times an action has taken on certain values. | |
![]() |
virtual | . |
s′=s+I{CV(x)∈tau}I{s<N−1}.
decl exper = new ActionCounter("Yrs Experience",10,work); //track up to 10 years working EndogenousStates(exper);
Public fields | ||
![]() |
||
Public methods | ||
![]() |
Create a variable that counts how many times an action has taken on certain values. |
Public fields | ||
![]() |
||
Public methods | ||
![]() |
Create a ActionCounter that checks reachability of an array of state counters. | |
![]() |
Trims the state space if the clock is exhibits aging and there is a list of state counters for a Target, assuming all are initialized as 0. |
q' = I{a ∈ r}.
Public fields | ||
![]() |
||
Public methods | ||
![]() |
Create a binary variable that indicates if the previous value of action variable was in a set. | |
![]() |
virtual | . |
q
denote this state variable.
b
be the base state variable to be agumented (not added to model itself)a
be the action variable that is the triggert
be the value of a
that pulls the trigger.r
be the value to go to when triggered.P(q′=z)=(1−I{a∈t})Prob(b′=z)+I{a∈t}}r
Public methods | ||
![]() |
Create a new augmented state variable for which an action triggers the change. | |
![]() |
virtual |
MyModel : DPparent { static decl a; ⋮ } ⋮ MyModel::Initialize() { DPparent::Initialize(new MyModel()); ⋮ a = new ActionVariable("choice",2); Actions(a); ⋮ CreateSpaces(); }
Public fields | ||
![]() |
vector of value labels. | |
Public methods | ||
![]() |
Create a new action variable. | |
![]() |
virtual | Return A[][pos]. |
![]() |
virtual | Return C[][pos]. |
![]() |
Return aC[][pos], current external value. |
Time increments randomly, with probability equal to inverse of bracket length.
If the bracket has size 1, then it is equivalent to Aging at that period. If the bracket is bigger than 1, the value function at that period is solved as a fix point.
RandomAging
SetClock(RandomAging,constant(5,1,10));which is equivalent to this:
SetClock(new AgeBrackets(constant(5,1,10)));Have normal aging for the first 10 years, then 3 brackets of five years, then 2 of 10 years:
SetClock(RandomAging,constant(1,1,10)~constant(5,1,3)~constant(10,1,2));
Public fields | ||
![]() |
const | Vector of period lengths for each age |
![]() |
Transition matrix based on Brackets | |
Public methods | ||
![]() |
Create an aging clock with brackets. | |
![]() |
virtual | Return flag for very last period possible. |
![]() |
virtual |
NormalAging
SetClock(NormalAging,10);which is equivalent to this:
SetClock(new Aging(10));
Public methods | ||
![]() |
Normal Finite Horizon Aging. | |
![]() |
virtual |
actual
vector should either be set by the user after creation of the state
variable (if the actual asset levels are fixed), or the user should define a derived class and
provide an Update()
method that will keep actual
current based on
dynamic changes in parameters.
actual[v]
.
A* = min( max( rA + S , actual[0] ), actual[N-1] )Define Ab ≤ A* ≥ At as the actual values that bracket A* (they are equal to each other if A* equals one of the discrete actual values). Define m = (Ab+At)/2. Let
b
and t
be the indices of the bracketing values.
Prob(a′ = b) = (A-Ab)/m Prob(a′ = t ) = 1 - Prob(a′=b).
Public fields | ||
![]() |
const | AV()-compatible object, interest rate on current holding. |
Public methods | ||
![]() |
Create a new asset state variable. |
An Augmented state variable has a transition that is based on another state variable's transtion.
For example, the Triggered state variables have a transition that is the same as the base variable sent when defining the augmented state variable unless a triggering condition is met. In that case the value next period is a special one that is not what the base transition would be.
The virtual IsReachable() for Augmented state variables is to return the base IsReachable() value.
Public methods | ||
![]() |
Base creator augmented state variables | |
![]() |
virtual | Default Indicator for intrinsic reachable values of a state variable. |
![]() |
virtual | Normalize the actual values of the base variable and then copy them to these actuals. |
![]() |
Synchronize base state variable value with current value. | |
![]() |
virtual | Default Transit (transition for a state variable). |
![]() |
virtual | Default Augmented Update. |
q
is autonomous when:q'
is independent of all other transitions. q
is not necessarily independent of another state variable s
because both
transitions can depend on the current state and actions.
For example, in a search model, suppose the worker's ability x and the match quality m are both unobserved, but the wage, w = xm, is observed. Then an auxiliary variable can be created for wage, added to the outcome and read in along with other data.
Public fields | ||
![]() |
observed in external data . | |
Public methods | ||
![]() |
Create a new element of χ, the space of auxiliary outcomes. | |
![]() |
virtual | Default contribution to likelihood. |
![]() |
virtual | Default realized auxililary variable, sets v=1.0 . |
![]() |
virtual |
MyModel : DPparent { static decl a; ⋮ } … MyModel::Initialize() { DPparent::Initialize(new MyModel()); ⋮ a = new BinaryChoice(); Actions(a); ⋮ CreateSpaces(); }
Public methods | ||
![]() |
Create a binary action variable. |
Public fields | ||
![]() |
const | Binary ActionVariable. |
![]() |
const | 2x1 proportions. |
Public methods | ||
![]() |
Create a 3-valued state variable that records a binary outcome of a binary choice. |
t≤ Tbar
d = new BinaryChoice("d"); d5 = new ChoiceAtTbar("d5",d,5); EndogenousStates(dvals);Track the first four choices of d:
dvals = new array[4]; decl i; for(i=0;i<3;++i) dvals[i] = new ChoiceAtTbar("d"+sprint(i),d,i); EndogenousStates(dvals);
Public fields | ||
![]() |
const | |
Public methods | ||
![]() |
Record the value of an action variable at a given time. | |
![]() |
virtual | Prune non-zero values before Tbar |
Public fields | ||
![]() |
const | Store Ergodic Distribution. |
![]() |
static | Norm parameter for stationary convergence. |
![]() |
const | t, current TimeVariable |
![]() |
const | t', next period TimeVariable used during V iteration to track possible clock values next period. |
![]() |
static | Volume for clocks (static so don't need to access internal var. |
Public methods | ||
![]() |
Base clock block. | |
![]() |
virtual | Flag for last period. |
![]() |
virtual | |
![]() |
virtual | |
![]() |
virtual | The base calculation to be carried out after the value of all states at time t have been computed. |
Public fields | ||
![]() |
StateBlock that I belong to | |
![]() |
Index into block array/vector | |
Public methods | ||
![]() |
Create a new Coevolving random variable. | |
![]() |
virtual | Return actual[v]. |
Public fields | ||
![]() |
EffectBlock that I belong to | |
![]() |
Index into block array/vector | |
Public methods | ||
![]() |
Create a new CorrelatedEffect. |
Public fields | ||
![]() |
const | AV()-compatiable reset to 0 flag |
![]() |
const | Variable to track |
![]() |
const | Values to track |
Public methods | ||
![]() |
Create a new counter variable. | |
![]() |
virtual | Trims the state space assuming counter is initialized to 0. |
![]() |
virtual |
decl qtr = new Cycle("Q",4); EndogenousStates(qtr);
Public methods | ||
![]() |
Create a deterministic cycle variable. |
Public fields | ||
![]() |
matrix with transition. | |
Public methods | ||
![]() |
Create a state variable with an arbitrary deterministic transition. |
Models with subperiods typically index a period t
and within t
a fixed
and finite number of subperiods, s=0,…,SubT-1
.
T = MajT × SubT
and
ordinary time simply advances from t=0
to t=T‾
like normal aging.t
advances, subt
cycles through 0…SubT‾
.majt=0
and stays there until subt
returns to 0
at which point it increments to 1
and so on until it gets to MajT‾
.
MajT = 2, SubT =3 t 0 1 2 3 4 5 ------------- majt 0 0 0 1 1 1 subt 0 1 2 0 1 2
MajT=0
: If the user sets MajT=0
it indicates an infinite horizon problem
(i.e. really MajT=∞
). subt
continues to cycle deterministically. The
difference with the ordinary case is that normal time cycles as well: t=0 1 …SubT‾ 0 1 … SubT‾ …
.
The problem is then ergodic and the value of states is a solution to a fixed point problem. Convergence
only needs to be checked for subt=0
. Once that period has converged the subsequent subperiods
will also automatically be at the fixed point after one last iteration.
MajT = 0, SubT =2 t 0 1 0 1 0 1 0 1 0 … --------------------------- majt 0 0 0 0 0 0 0 0 0 … subt 0 1 0 1 0 1 0 1 0 …
HasInit
: A model may have an initial undivided period that allows for, say, permanent
heterogeneity to be realized.
MajT = 2, SubT =3, HasInit = TRUE t 0 1 2 3 4 5 6 --------------- majt 0 1 1 1 2 2 2 subt 0 0 1 2 0 1 2MajT = 0, SubT =2, HasInit = TRUE t 0 1 2 1 2 1 2 1 2 … --------------------------- majt 0 1 1 1 1 1 1 1 1 … subt 0 0 1 0 1 0 1 0 0 …
HasFinal
: A model may have a final undivided period to allow for terminal conditions
such as bequest motives. This option cannot be combined with MajT=0
.
MajT = 2, SubT =3, HasFinal = TRUE t 0 1 2 3 4 5 6 --------------- majt 0 0 0 1 1 1 2 subt 0 1 2 0 1 2 0
St=0
it then cycles back to
s
the state variable
must be endogenous (added to θ). So a state variable that would normally be exogenous has to
be treated as endogenous unless it changes value every subperiod.SetClock(SubPeriods,0,2);Allow for an initial period before the stationary phase begins:
SetClock(SubPeriods,0,2,TRUE);Create a finite horizon model in which a person makes binary decisions in 3 separate subperiods over a lifetime of 10 periods
SetClock(SubPeriods,10,3); d = {new Binary("d0"),new Binary("d1"),new Binary("d2")}; Actions(d); ⋮ FeasibleActions(A) { decl i,v; notfeas = zeros(rows(A),1); foreach(v in d[i]) notfeas .+= (I::subt.!=i)*A[][d[i].pos]; // d[i]=1 not feasible if subt &neq; i return 1-notfeas; }
Public fields | ||
![]() |
zeros(SubT-1,0)|δ. | |
![]() |
const | Final subdivided period (TRUE/FALSE). |
![]() |
const | Initial subdivided period (TRUE/FALSE). |
![]() |
const | Number of major periods, 0=Infinite horizon. |
![]() |
const | Number of sub periods. |
![]() |
||
Public methods | ||
![]() |
Create a clock divided into subperiods. | |
![]() |
virtual | Not stationary and last period. |
![]() |
virtual | |
![]() |
virtual | |
![]() |
virtual | Transition for sub-divided time. |
![]() |
virtual | |
![]() |
virtual | Check newly computed values; an alternative to computing norm of value function differences. |
This variable requires a target s.X and the lagged value of the target, denoted s.Lx.
Public fields | ||
![]() |
||
![]() |
const | |
![]() |
||
![]() |
const | |
![]() |
const | |
![]() |
const | |
Public methods | ||
![]() |
Create a variable that counts the consecutive periods a state or action has had the same value. |
k
and t
, the type of
episode and duration. Duration is not tracked for k=0. New episodes occur with Onset
probability and end with
End
probability.
Public fields | ||
![]() |
const | episode end probabilities |
![]() |
const | all spells end at t=T-1. |
![]() |
const | Coevolving, current type of episode |
![]() |
const | episode onset probabilities |
![]() |
Stores end probability at t=T-1 for non-Finite episodes | |
![]() |
const | Coevolving, duration of episode, NOT tracked for k.v=0 |
Public methods | ||
![]() |
K mutually exclusive episodes. |
This class is used internally so that no subvector is empty. It takes on the value 0 each period. Since it takes on only one value it does not affect the size of state spaces, only the length of the state vector.
The user might want to add a fixed variable to the model as placeholder for a state variable that
is not coded yet. For example, if in the complete model q
is a complex new state
variable that will require coding of a new Transit()
function, then you made start
with q
as a Fixed so formulas involving it can be completed.
Public methods | ||
![]() |
Create a constant entry in the state vector. |
Public fields | ||
![]() |
const | ActionVariable. |
Public methods | ||
![]() |
Create a new FIXED asset state variable. |
Members of γf are derived from this class.
Solution methods loop over fixed effect values, re-using storage space.
Public methods | ||
![]() |
A state variable that is non-random an invariant for and individual DP problem. |
Public methods | ||
![]() |
Create a list of FixedEffect state variables. |
Public methods | ||
![]() |
An augmented state variable that 'forgets' its value when an action leads to a (semi-) permanent condition. | |
![]() |
virtual | Trims the state space of values that can't be reached because they are forgotten. |
Public methods | ||
![]() |
Create an augmented state variable that 'forgets' its value when at t==T* (effective T*+1). | |
![]() |
Trim states because t > T* . |
Public methods | ||
![]() |
Augment a state variable so it freezes at its current value as long as a trigger is TRUE. | |
![]() |
virtual |
decl poff = new Probability("op",0.6); decl hasoffer = new IIDBinary"off",poff);
Public methods | ||
![]() |
Create binary state variable for which Prob(s=1) = Pi. |
decl health = new IIDJump("h",<0.1;0.8;0.1>);
Public methods | ||
![]() |
Create a new IID Jump process. |
Public fields | ||
![]() |
const | |
![]() |
const | |
![]() |
const | |
![]() |
const | |
![]() |
const | |
Public methods | ||
![]() |
Create an indicator (or indicator interacted with another outcome). | |
![]() |
virtual | Default realized auxililary variable, sets v=1.0 . |
Public fields | ||
![]() |
const | parameter for jump probability π |
Public methods | ||
![]() |
Create a variable that jumps with probability |
value that enters the endogenous vector θ depending on reservation values to keep it. A kept random discrete version of ζ enters the state as this variable.
Its value is constant as long as a state variable indicates it is being held.
Public fields | ||
![]() |
const | |
![]() |
const | |
Public methods | ||
![]() |
virtual | The default cdf of a kept continuous variable: Φ(z*). |
![]() |
virtual | Pold 0 0 Pold |
![]() |
virtual | |
![]() |
Create a discretized verison of the continuous ζ variable that enters the state when accepted (or kept). | |
![]() |
virtual | The default quantile of a kept continuous variable: Φ-1(u). |
s′=CV(Target).
Public fields | ||
![]() |
const | Order of lag (for pruning). |
![]() |
const | Variable to track. |
Public methods | ||
![]() |
virtual | Presumes an initial value of 0 for prunable clocks. |
![]() |
Takes on a previous value of another state or action. | |
![]() |
virtual |
0
, so for finite horizon clocks, t=0,q>0
is marked unreachable. Set Prune=FALSE
to not prune these unreachable states automatically.Public fields | ||
![]() |
||
![]() |
||
Public methods | ||
![]() |
Create a variable that tracks the previous value of action variable. |
Note the Target variable must be in η or θ. It canot be added to ϵ.
Public methods | ||
![]() |
Create a variable that tracks the previous value of another state variable. |
Public fields | ||
![]() |
const | |
![]() |
const | AV()-compatible static function or ActionVariable. |
Public methods | ||
![]() |
Create a new LIQUID asset state variable. |
Public fields | ||
![]() |
const | |
Public methods | ||
![]() |
Create a discretized lognormal offer. | |
![]() |
virtual | Updates the actual values. |
Like Mortality but at t = T*-1 ≡ T-2
a stationary environment occurs
with constant probability of mortality.
UncertainLongevity
t=5
) is stationary and transitions to the last period with
probability 1/10 each period.
SetClock(UncertainLongevity,7,0.1);which is equivalent to:
SetClock(new Longevity(7,0.1));Over periods allow mortality rate to increase linearly in age up to a stationary mortality rate of 50% in the final stage of life:
MyModel::Mprob() { return 0.5*I::t/(N::T-2) ; } ⋮ SetClock(RandomMortality,9,MyModel::Mprob);
Prob(t' = T*) = π(t) Prob(t' = t+1) = 1 - π(t) if t < T*-1 Prob(t' = t) = 1-π(t) if t = T*-1.
Public fields | ||
![]() |
const | |
Public methods | ||
![]() |
virtual | With Longevity the last period (death) is definitively the last. |
![]() |
Random death and uncertain maximum lifetime. | |
![]() |
virtual | |
![]() |
virtual | Mixed value updating. |
q
is an autonomous random variable
q
is discrete and the transition to q'
can only depend on q
,
the transition is a square matrix.
The transition must be square. The number of values it takes is determined by the dimension of the column or vector.
If it is a matrix, then the rows are next period states and the columns are currents. P(s′=z|s)=Πzs To be handled properly the state variable must be placed in the endogenous state vector θ.
decl tmat =< 0.90 ~ 0.09 ~0.05; 0.02 ~ 0.80 ~0.3; 0.08 ~ 0.01 ~0.11> x = new Markov("x",tmat); EndogenousStates(x);
See TransitionMatrix
Public fields | ||
![]() |
||
![]() |
const | |
Public methods | ||
![]() |
Create a discrete Markov process. |
This clock has the agent age each period but allows for a state-dependent probability that they die before the start of the next period.
Death is a transition to the maximum value of the counter, T-1. This allows for endogenous bequest motives which are computed before iterating backwards on other ages.
The probability of death should be a CV-compatible quantity. This means that the mortality risk can be state dependent.
SetClock(RandomMortality,7,0.1);which is equivalent to:
SetClock(new Mortality(7,0.1));Over 9 periods, allow mortality rate to be an increasing function of age and health status. Health is a binary state variable in the model, where 1 is good health and 0 is poor health and acts like aging 3 periods:
MyModel::Mprob() { decl v = exp(I::t + 3*(1-CV(health))); return v/(1+v); } ⋮ SetClock(RandomMortality,9,MyModel::Mprob);
Public fields | ||
![]() |
EV at t=T-1, value of death | |
![]() |
const | CV-compatible mortality probability |
![]() |
||
![]() |
const | |
Public methods | ||
![]() |
Set clock to be deterministic aging with early random death. | |
![]() |
virtual | |
![]() |
virtual | Mixed value updating. |
Public methods | ||
![]() |
||
![]() |
Default realized auxililary variable, sets v=1.0 . |
Public fields | ||
![]() |
const | Number of points per variable |
![]() |
const | total number points in the block. |
Public methods | ||
![]() |
Create a multivariate normal state block. |
x ∼ N( μ, C'C ).
Public fields | ||
![]() |
const | AV()-compatible vector-valued object which returns
vech() for lower triangle of C, the Choleski decomposition
of the variance matrix Ω, Ω = CC'. |
![]() |
const | vector of means μ |
![]() |
const | matrix of Z vals for simulated actual values |
Public methods | ||
![]() |
Create a block for a multivariate normal distribution (IID over time). | |
![]() |
virtual | Updates the grid of Actual values. |
This variable needs to be used with care: CV()
will return a single value which is the index between 0 and N-1.
AV()
returns the 1×M vector that corresponds to the index.
Public fields | ||
![]() |
||
![]() |
||
![]() |
||
Public methods | ||
![]() |
Create a single IID variable which vectorizes a multivariate normal using psuedo-random draws. | |
![]() |
||
![]() |
Public fields | ||
![]() |
implied normal value. | |
![]() |
const | noise is additive (not multiplicative). |
![]() |
const | standard deviation of the normal nois. |
![]() |
const | Object that is measured with noise. |
Public methods | ||
![]() |
likelihood contribution of a noisy auxiliary value . | |
![]() |
Create an auxiliary value that adds normally-distributed noise to the actual value. | |
![]() |
Default realized auxililary variable, sets v=1.0 . |
Public methods | ||
![]() |
virtual | |
![]() |
virtual | Mixed value updating. |
DDP makes no distinction between random and non-random state variables except to organize different kinds of transitions into a taxonomy.
A non-random state variable has a transition which is 0 for all possible states except one, for which it equals 1. Its state next period is determined by the current state and action.
q
counts how many times another state variable s
has equalled 1 in the past. Since
the current value of s
is known, the value of q
next period is (conditionally)
deterministic.
As a transition function:
q′=q+I{s=1}.
As a transition probability:
P(q′;q,s)=I{q′=q+I{s=1}}.
Public methods | ||
![]() |
virtual | The baseline transition for non-stationary models (normal aging). |
![]() |
virtual | Check newly computed values; an alternative to computing norm of value function differences. |
Public methods | ||
![]() |
A single element of a MVNormal block. |
Public fields | ||
![]() |
const | |
Public methods | ||
![]() |
Update pdf, the distribution over the random effect. | |
![]() |
Create a permanent discretized normal random effect. |
Public fields | ||
![]() |
||
Public methods | ||
![]() |
Create a Normal N(mu,sigma) discretize jump variable. | |
![]() |
virtual |
Public fields | ||
![]() |
const | ActionVariable - indicates aceptance |
![]() |
||
![]() |
||
![]() |
const | π Double, Parameter or static function, offer probability |
Public methods | ||
![]() |
Create an offer state variable. | |
![]() |
virtual |
If unemployed an offer was generated with prob. φ. The agent can accept an offer and become employed or reject and continue searching again next period. If employed the agent can quit or can be laid off with probability λ or keep their job.
If quit or laid off last period, the offer stays the same for one period but transition to employment is impossible. This allows the previous wage to be captured and used in a block derived from this.
φ and λ can depend on (w,m) and on α
(w,m) m ∈ Status If m=Unemp: w′ = w with prob. a 0 with prob. (1-a)(1-φ) or 1 ... N-1 with prob. (1-a)φ m′ = aEmp + (1-a)Unemp; If m=Emp, w′ = w m′ = (1-a)Quit aLaidOff with prob. λ aEmp with prob. 1-λ If m=Quit or LaidOff, m′ = Unemp w′ same as m=Unemp and a=0
Public fields | ||
![]() |
const | Action to accept offer |
![]() |
const | Probability of reverting to 0 |
![]() |
const | Variable containing offer index |
![]() |
const | Job Offer Probability |
![]() |
const | searching, just laid off, currently holding |
Public methods | ||
![]() |
||
![]() |
An offer with layoff (match dissolution) risk. | |
![]() |
||
Enumerations | ||
![]() |
Status values. |
Public methods | ||
![]() |
Create a variable that tracks a one-time permanent choice. |
Public methods | ||
![]() |
Create a variable that tracks that some Target condition has occurred in the past. |
Public fields | ||
![]() |
const | indicates time=Rmax in phase |
![]() |
const | vector of times in phases |
![]() |
const | |
![]() |
const | The last treatment phase (0 if no treatment) |
![]() |
const | vector of phases |
![]() |
const | The index into the vector of the state variable that are the first period of each phase |
![]() |
const | Vector of positive integers that are the length of each of the phases. |
Public methods | ||
![]() |
A sequence of finite-termed phases of treatment. | |
![]() |
virtual | The default transition for treatment. |
![]() |
virtual | Mixed value updating. |
DDP makes no distinction between random and non-random state variables except to organize different kinds of transitions into a taxonomy.
A random state variable has a transition which is not always 0/1 .
Its value next period is determined stochastically by the current state and action.
Elements of γr derived from this class. A random effect plays a similar to its counterpart in an econometric model. Solution methods loop over random effect values and will account for the distribution in computing predictions, simulations and econometric objectives.
RandomEffect("g",N);A binary variable with constant and unequal weights:
RandomEffect("g",2,<0.8;0.2>);A two-point random effect for which the distribution is a function of the value of a fixed effect (level of education) and a parameter vector β:
decl beta; ⋮ hd() { decl v = exp((1~AV(educ))*CV(beta)), s = sumc(v); return v/s; } ⋮ GroupVariables( h=RandomEffect("h",2,hd), educ = FixedEffect("ed",4); }A random effect with a distribution that equals a Simplex parameter block:
enum{Npts = 5}; hd = new Simplex("h",Npts); RandomEffect("h",Npts,hd);
fDist
is stored in a non-constant member, so it can be changed after the CreateSpaces()
has been called.
Public fields | ||
![]() |
holds the object that computes and returns the distribution across values. | |
Public methods | ||
![]() |
virtual | Update pdf, the distribution over the random effect. |
![]() |
Create a state variable that is random but invariant for an individual DP problem. |
Public methods | ||
![]() |
virtual | |
![]() |
Create a list of CorrelatedEffect state variables. |
Public methods | ||
![]() |
||
![]() |
A binary state that starts and stops according to a random process. | |
![]() |
The difference with ValueTriggered is that this transition is random as of the current state but ValueTriggered the transition is deterministic.
q
denote this state variable.
b
be the base state variable to be agumented τ
be the actual value of the trigger probability.r
be the value to go to when triggered.Prob( q′= z | q,b ) = τI{z=r} + (1-τ)Prob(b′=z)
a = ActionVariable("work",2); lprob = new Probability("layoff prob",0.1); m = RandomTrigger( new LaggedAction("emp",a),lprob,0);
Public methods | ||
![]() |
Augment a base transition so that a special value occurs with some probability. | |
![]() |
virtual |
P(s′=s−1)=I{s>0}Π0P(s′=s)=Π1+I{s=0}Π0+I{s=N−1}Π2P(s′=s+1)=I{s<N−1}Π2
Π can be a vector, a Simplex parameter block
or a static function that returns a 3×1 vector.
Fertility : FiniteHorizon { static decl q, i, M, x; static Mortality(); } Fertility::Initalize() { q = new Coefficients("q",); Actions(i = new Action("i",2)); EndogenousState( M = new RandomUpDown("M",20,Fertility::Mortality) ); } Fertility::Mortality() { decl p = probn(x*q); // x has to be updated using current age and other x values. decl ivals = CV(i); return 0 ~ (1-p*ivals) ~ p*ivals; }
Public fields | ||
![]() |
||
![]() |
const | |
Public methods | ||
![]() |
virtual | Prune assuming initial value = 0. |
![]() |
Create a state variable that increments or decrements with (possibly) state-dependent probabilities. | |
Enumerations | ||
![]() |
Public methods | ||
![]() |
virtual | Default realized auxililary variable, sets v=1.0 . |
![]() |
Realized utility, U(). |
Public methods | ||
![]() |
Public fields | ||
![]() |
||
Public methods | ||
![]() |
Returns TRUE if any rows of InObservedX equal the current actual value of the block. | |
![]() |
Create a vector of fixed effects (FixedEffectBlock). |
Zurcher : Ergodic { static decl q, i, x; ⋮ } Zurcher::Initialize() { ⋮ q = new Simplex("q",3); AddVariable(i = new Action("i",2)); EndogenousState( x = new Renewal("x",90,i,q) ); ⋮ }
Public fields | ||
![]() |
const | length of Pi |
![]() |
const | block or vector of increment probabilities |
![]() |
const | state or action that resets the process to 0 |
Public methods | ||
![]() |
Create a Rust renewal process state variable. |
A special case of ActionTriggered in which 1 is the trigger and the special reset value is 0.
If the trigger takes on more than two values only the value of 1 is the trigger.
Public methods | ||
![]() |
Augment a base transition to reset to 0 if the trigger is 1. |
s' = .
Public fields | ||
![]() |
const | |
![]() |
||
![]() |
||
![]() |
const | |
![]() |
const | |
![]() |
||
![]() |
const | |
Public methods | ||
![]() |
Accept or reject an offer then retain. |
Public fields | ||
![]() |
||
![]() |
||
![]() |
||
![]() |
||
![]() |
const | |
![]() |
||
![]() |
||
Public methods | ||
![]() |
Rouwenhorst discretizization of a correlated normal process. | |
![]() |
virtual | Dynamic updating of the transition probabilities. |
Public methods | ||
![]() |
Create an equally likely discrete Exogenous variable. | |
![]() |
Transition . |
s' = min( s+ s.x, s.N-). Ρ(s'=z | α,ε, θ,ψ) = I{z = min(s + s.x,s.N-) }.
decl totoff = new StateAccumulator("Total Offers",noffers,10); //track total offers received up to 10 EndogenousStates(totoff);
Public methods | ||
![]() |
Create a variable that tracks the cumulative value of another state. |
t≤ Tbar
Public fields | ||
![]() |
const | value of I::t to record at. |
Public methods | ||
![]() |
virtual | Prune non-zero values before Tbar |
![]() |
Record the value of a state variable at a given time (starting at the next period). |
Public fields | ||
![]() |
matrix of all actual values. | |
![]() |
matrix of all current values | |
![]() |
vector <0:N-1>, used in AV(). | |
Public methods | ||
![]() |
Add state variable(s) to a block. | |
![]() |
virtual | Sets and returns the vector of actual values of the block as a row vector. |
![]() |
Create a list of Coevolving state variables. | |
![]() |
virtual | Default Transit (transition for a state variable). |
The values of x
in τ (ToTrack)
s′=s+I{CV(x)∈tau}I{s<N−1}.
decl wks = new ActionState("wksunemp",work,10,<0>); //track up to 10 years EndogenousStates(wks);
Public methods | ||
![]() |
Create a variable that counts how many times another state has taken on certain values. |
Public fields | ||
![]() |
||
Public methods | ||
![]() |
Trims the state space if the clock is exhibits aging and there is a list of state counters for a Target, assuming all are initialized as 0. | |
![]() |
Create a StateCounter that checks reachability of an array of state counters. |
t
denote the state variable being tracked. Let r
denote
the value or vector of values to track.
q' = I{t ∈ r}.
Public methods | ||
![]() |
Create a binary variable that indicates if the previous value of state variable was in a set. |
Public fields | ||
![]() |
const | Trim unreachable states if finite horizon clock is detected. |
![]() |
A vector of values that end decision making Equal to < > if state is not terminating. | |
Public methods | ||
![]() |
virtual | Check dimensions of actual . |
![]() |
static | Check if variable is a block of state variables. |
![]() |
static | Check if variable is a valid member of a block. |
![]() |
virtual | Default Indicator for intrinsic reachable values of a state variable. |
![]() |
Designate one or more values of the state variable as terminal. | |
![]() |
virtual | Return actual[v]. |
![]() |
The base state variable creator. | |
![]() |
Track the state variable in predictions. | |
![]() |
virtual | Default Transit (transition for a state variable). |
![]() |
virtual | Returns the transition for a state variable that is unchanged next period. |
Public fields | ||
![]() |
const | static function. |
Public methods | ||
![]() |
virtual | Default realized auxililary variable, sets v=1.0 . |
![]() |
Create an wrapper for a static function AV()-compatible object. |
StaticProgram
SetClock(StaticProgram);which is equivalent to this:
SetClock(new StaticP());which is also equivalent to this:
SetClock(NormalAging,1);
Public methods | ||
![]() |
A static problem: Aging and T=1. |
InfiniteHorizon
, Ergodic
SetClock(InfiniteHorizon);Sent directly:SetClock(Ergodic);
SetClock(new Stationary(FALSE));SetClock(new Stationary(TRUE));
Public methods | ||
![]() |
virtual | No period is the last period in a stationary environment. |
![]() |
A stationary clock block. |
Public fields | ||
![]() |
EffectBlock that I belong to | |
![]() |
Index into block array/vector | |
Public methods | ||
![]() |
Create a new Sub. |
Public methods | ||
![]() |
virtual | Use majt for time period when checking reachability of the base state variable. |
![]() |
Make a state variable only transit in one subperiod of a Divided clock. |
Public fields | ||
![]() |
||
![]() |
const | |
![]() |
||
![]() |
const | support. |
![]() |
const | |
![]() |
||
![]() |
||
Public methods | ||
![]() |
Tauchen discretizization of a correlated normal process. |
Public fields | ||
![]() |
const | |
![]() |
const | |
Public methods | ||
![]() |
virtual | Update pdf, the distribution over the random effect. |
![]() |
Create a permanent discretized normal random effect. |
Public methods | ||
![]() |
Do Nothing and prohibit derived Updates. |
Public methods | ||
![]() |
. |
Public fields | ||
![]() |
const | Variable to track |
![]() |
const | Values to track |
Public methods | ||
![]() |
Default initial value for Tracker variables is 0. | |
![]() |
Indicator that a state or action took on particular value(s) last period. |
Public methods | ||
![]() |
The base for triggered augmentations. | |
![]() |
virtual | Default Augmented Update. |
Public methods | ||
![]() |
virtual | |
![]() |
virtual | |
![]() |
q
denote this state variable.
b
be the base state variable to be agumented (not added to model itself)s
be the trigger objectt
be the value of s
that pulls the trigger.r
be the value to go to when triggered.q′ = I{a≠t} b′ + (1-I{s==t}}r
Public methods | ||
![]() |
virtual | |
![]() |
Augment a base transition so the value of some other object trigger a special value. |
Public fields | ||
![]() |
const | |
Public methods | ||
![]() |
||
![]() |
Create a exponentially distributed discretize jump variable. |
Public methods | ||
![]() |
Create a standard normal N(0,1) discretize jump variable. |
Enumerations | ||
TransOutput | . elements of array returned by Transit | Qind, Qprob, TransOutput |
L | label |
N | integer, maximum number of times to count |
Action | ActionVariable to accumulate values for. |
L | label |
Act | ActionVariable to track. |
N | integer, maximum number of times to count |
ToTrack | vector, values of Act to count, default=<1> DoAll: all non-zero values |
Reset | AV()-compatible binary value that resets the count if TRUE (default=0). |
Prune | TRUE (default), prune unreachable states if finite horizon detected. |
decl exper = new ActionCounter("Yrs Experience",10,work); EndogenousStates(exper);
L | label |
N | integer, maximum number of times to count |
Act | ActionVariable to track. |
ToTrack | integer , values of State to count. |
L | Label |
Target | ActionVariable to track |
ToTrack | integer or vector of values to track |
Prune | Prune in finite horizon (starts at 0) |
b | the base StateVariable whose transition is augmented. |
t | the ActionVariable that triggers a different transition |
tv | the integer (actual) or vector of integer values of tv that triggers the transition [default=1]. |
rval | the integer (current) value of this state variable when the trigger occurs [default=0] -1, then the reset value this.N = b.N+1 and rval = b.N.
Prob( q' = v ) = I{t∈tv}I{v==rval} + (1-I{t∈tv})Prob(b=v) |
L | string a label or name for the variable. default = "a" |
NorVLabels | positive integer, number of values the variable can take on. default N=1 is a constant, which can be included as a placeholder for extensions of a model. OR N-array of strings, holding labels for each choice (used in printing)
|
Brackets | vector of period lengths
StaticProgram |
SetClock(StaticProgram);which is equivalent to this:
SetClock(new StaticP());which is also equivalent to this:
SetClock(NormalAging,1);
NormalAging
SetClock(NormalAging,10);which is equivalent to this:
SetClock(new Aging(10));
L | label |
N | number of values |
r | AV()-compatible object, interest rate on current holding. |
Lorb | either a StateVariable object, the base variable to augment
Or, string the label for this variable. In thise case the base is Fixed |
N | integer, if Otherwise, if > b.N number of points of the augmented variable. Otherwise, ignored and this.N = b.Nl
If Lorb is a string then b = new Fixed(Lorb) . |
MaxV | positive real, default = 1.0 Sets the actual vector to 0,…, MaxV. |
This is a virtual method, so derived classes replace this default version with the one that goes along with the kind of state variable.
Alpha::N
× L or 1 × L
matrix of action-specific transition probabilities, Ρ(q′;α,η,θ). <0> , CondProbOne }
. That is the with
probability one the value of the state variable next period is 0.
Not called by user code
L | label
|
y, | the current realized outcome, υ. |
v=1.0
.
y, | the current realized outcome, υ. |
L | label |
b | BinaryChoice action variable. |
ratios | AV-compatible object that returns a 2-d probabilities (ratio[0]+ration[1]=1)
Transition values: s' Prob Interpretation ------------------------------------------------------------------ 0 CV(b)==0 Choose not to have a child this period 1 CV(b)*ratio[0] Had child, it's a boy 2 CV(b)*ratio[1] Had child, it's a girl |
L | label |
Target | ActionVariable to record |
Tbar | clock time at which to record choice |
Prune | TRUE [default], prune unreachable states (non-zero values) before Tbar
s′={0 if t less than TbarCV((Target) if t equals Tbars if t greater than Tbar
|
Nt | integer, the number of different values t takes on
|
Ntprime | integer, the number of values t' takes on.
|
If TRUE then transitions are not computed.
Vupdate()
is called by Update() at the end of one Bellman iteration.
norm()
of the change in the value function for convergence check.Lorb | label or base StateVariable |
N | number of values it takes on. |
s
contains a state variable then
AV(s)
calls s->myAV()
actual[v]
.
L | label |
N | number of values it takes on. |
L | label |
N | number of values |
Target | state or action to track |
ToTrack | integer or vector of values to count DoAll, track all non-zero values |
Reset | AV()-compatiable reset to 0 flag |
Prune | prunt assuming counter starts at 0. |
L | label |
N | integer, number of periods in the cycle |
decl qtr = new Cycle("Quarter",4); EndogenousStates(qtr);
L | label |
nextValues | vector of next (current) values
|
s = Deterministic(L,<1;0>);
MajT | integer number of periods 0 infinite horizon |
SubT | positive integer, sub-periods |
HasInit | FALSE [default] t=0 is a sub-period (so the subperiod increments and the major period stays the same)TRUE t=0 is not subdivided (so the subperiod stays 0 in the following period and the major period increments).
|
HasFinal | FALSE [default] The final period is the final sub-period. the final period is an undivided period (so its subperiod is 0) This cannot be combined with an infinite horizon
|
L | label |
Current | ActionVariable or StateVariable that holds the current value |
Lag | StateVariable holding previous value or vector of values to count runs for |
N | integer, maximum number of periods to count |
MaxOnce | FALSE [default] spells repeat; TRUE once max is reached it remains frozen. |
Prune | TRUE [default]: prune states if finite horizon detected. |
streak = new Duration("Streak",Won,<1>,10); //track winning streaks up to 9 periods (9 means "9 or more");Suppose Result equals 0 for loss, 1 for a tie and 3 for a win. Then the right-censored unbeaten streak is
noloss = new Duration("Unbeaten",Result,<1,2>,10); //track unbeaten streaks up to 10 periods long
Choice = new ActionVariable("c",3); prechoice = new LaggedAction("lagc",Choice); contchoice = new Duration("Streak",Choice,prechoice,5); //track streaks of making same choice up to 5 periods
L | string, label |
K | integer, number of different episodes. k=0 is a special no-episode state. Spells with k>0 end with a transition k'=0 |
T | integer, maximum duration of episodes, or maximum duration to track if not Finite |
Onset | AV()-compatible K vector of onset probabilities (first element is k=0 spell continuation probability) |
End | AV()-compatible probability of current spell ending. |
Finite | TRUE, T is the actual limit of spells. At t=T-1 transition is k'=0, t'=0. FALSE, T is limit of duration tracking. Spell ends with prob. calculated at first case of t=T-1. |
L | label
|
h = new Fixed("health");
L | label |
N | number of values |
r | AV()-compatible object, interest rate on current holding. |
delta | ActionVariable |
L | label for block |
b | base StateVariable to augment |
t | the ActionVariable that triggers a different transition |
FCond | (permanent) CV-compatible value that indicates condition already exists if TRUE FALSE otherwise |
tv | the integer (actual) or vector of integer values that triggers the transition [default=1]. |
rval | the integer (current) value of this state variable when the trigger occurs [default=0] -1, then the reset value this.N = b.N+1 and rval = b.N.
|
b | base StateVariable to augment |
Tstar | time to forget |
b | base StateVariable |
t | AV()-compatible object |
L | label |
Pi | probability of 1 [default = 0.5] |
L | label |
Pi | column vector or a Simplex parameter block or a static function |
target | State or Action variable to create indicator value for |
myval | value to indicate |
iobj | integer (no indicator) or an outcome to interact. |
prefix | string for column matching or integer (not in data) |
v=1.0
.
y, | the current realized outcome, υ. |
L | label |
N | number of values |
Pi | AV()-compatible jump probability. |
zstar, | cut-off value |
L label | |
N number of points the state variable will take on. | |
keep | ActionVariable() that indicates ζ is kept. Next period this variable will contain a discrete approximiation to it. |
held |
u, | vector of |
L | label |
Target | Variable to lag |
Prune | TRUE [default], prune state space assuming initial value of 0 |
Order, | order of the lag [default=1] current value becomes next value tomorrow |
L | label |
Target | ActionVariable to track. |
Prune | TRUE [default]: prune non-zero states at t=0 if finite horizon detected. |
Order | number of lags [default = 1] |
wrked = new LaggedAction("Worked Last Year",work);
L | label |
Target | StateVariable to track. |
Prune | [optional default=TRUE] prune unreachable states automatically if finite horizon |
Order | [optional default=1] order of the lag (how many periods to prune)
|
prevoccup = new LaggedState("Prev",occup);
L | label |
N | number of values |
NetSavings | AV()-compatible static function of the form NetSavings() orActionVariable
Example. A model has a choice of consume or save with a budget constraint of the form:
c+a≤Y+(1+r)A.
Here c is consumption, a are assets to carry forward to tomorrow, Y is non-asset income, r is the current interest rate and
A are assets that were carried forward yesterday. Rewriting in terms of feasible actions:
aL≤a≤Y+(1+r)A.
Consumption would then be total income minus a. Thus, if A and a are continuous the transition is simply:
A′=a.
However, if a and A are both discretized, then two things occur. First, the current values are simply indices (0,1,...). So
the actual values need to set using a = new ActionVariable("a",5); A = new LiquidAsset("A",5,a); a-<SetActual(<-10;0;10;100;1000>); A-<SetActual(<-10;0;10;100;1000>); AV(Y) is too small.
|
L | label |
N | number of distinct offers. 0 means no offer |
Pi | double, Parameter or static function, prob. of an offer |
Accept | ActionVariable, acceptance of the offer. |
pars | 2x1 vector or array of NormalParams |
1,2,...,N-
.
T | number of age phases. T-1 = death; T-2 = stationary last period. |
MortProb | AV() compatible probability of death |
t=T-1
is computed as usual as a terminal state.t=T-2
is treated as an infinite horizon problem and iterated on.t
is deterministic or convergence was reached on the last iteration.t
and check for validityL | label |
Pi | N×1 array of AV(Alpha::C) compatible transition probabilities. |
decl tmat =< 0.90 ~ 0.09 ~0.05; 0.02 ~ 0.80 ~0.3; 0.08 ~ 0.01 ~0.11> x = new Markov("x",tmat);
T | length of horizon. T-1 = death; |
MortProb | AV()-compatible probability of early death |
t=T-1
is stored and used for the value of an early death.
t
is deterministic or convergence was reached on the last iteration.t
and check for validityv=1.0
.
y, | the current realized outcome, υ. |
L | label |
N | dimensions |
M | number of points in each dimension |
base |
vech()
for lower triangle of C, the Choleski decomposition
of the variance matrix Ω, Ω = CC'.
L | label for block |
N | integer, length of the vector (number of state variables in block) |
M | integer, number of values each variable takes on (So MN is the total number of points added to the state space) |
mu | either a Nx1 constant vector or a Parameter Block containing means
|
Sigma | either a N(N+1)/2 x 1 vector containing the lower diagonal of the Choleski matrix or a parameter block for it
NOTE: the first observation actual values always equals the mean for KW approximation. |
L | label |
Ndraws | number of draws of the vector. |
Ndim | width of the vector. |
pars | 2-array of NormalParams , mean and Cholesky distrubution (in the Nsigma spot) |
myseed | 0 [default] do not set the seed. Positive will set the seed for the draw of the IID Z matrix. See Ox's ranseed(). |
v=1.0
.
y, | the current realized outcome, υ. |
t
is deterministic or convergence was reached on the last iteration.t
and check for validityL |
N |
The user can supply a replacement in a derived class.
L | label |
N | number of points |
pars | AV()-compatible NormalParams vector/array containing mean and standard deviation
The probabilitY of each discrete value is 1/N. The distribution is captured by adjusting the actual values to be at the quantiles of the distribution. |
L | label |
Ndraws | N |
pars | 2x1 vector or array of NormalParams |
L | label |
N | integer, number of values (0 ... N-) |
Pi | a AV(Pi,Feas)-compatible offer probability. |
Accept | ActionVariable that indicates the offer is accepted. |
Enumerations | ||
Status | Status values. | Unemp, Quit, LaidOff, Emp, Nstatus |
L | string, label |
N | integer, number of distinct offers. 0 is no offer |
accept | ActionVariable, indicator for accepting offer or not quitting if already employed |
Pi | CV compatible offer probability |
Lambda | CV compatible layoff probability |
L | label |
Target | ActionVariable to track.
s′={0 if CV(Target)==0CV(Target) if s==0 and CV(Target) greater than 0s if s greater than 0
|
retired = new PermanentChoice("Ret",retire);
L | label |
Target | CV-compatible value to track. |
ToTrack | vector of values to Track. Default=<1>. |
Prune. | Begins at 0 in finite horizon clocks. |
??
Rmaxes | vector of maximum times in each phase. |
IsErgodic | TRUE, store Ρ* and use it for sampling |
t
is deterministic or convergence was reached on the last iteration.t
and check for validityThe user can supply a replacement in a derived class.
L | label |
N | number of points of support. |
fDist | integer [default], uniform distribution otherwise, a AV()-compatible object that returns the N-vector of probabilities.
RandomEffect("g",N);means: Prob(g=k) = 1/N, for k=0,...,N‾ RandomEffect("g",2,<0.8;0.2>);a static function: decl X, beta; ⋮ hd() { decl v = exp(X*beta), s = sumc(v); return v/s; } ⋮ RandomEffect("h",rows(X),hd);or a parameter block: enum{Npts = 5}; hd = new Simplex("h",Npts); RandomEffect("h",Npts,hd);
fDist is stored in a non-constant member, so it can be changed after the CreateSpaces()
has been called.
|
L | label for block |
L | label |
ProbEnd | probability (AV()-compatible) of a transition from 0 to 1 [default=0.5] |
Start | 0/1 value (ActionVariable or CV-compatible) that moves from 0 to 1 [default=1]
|
injury = new RandomSpell("i",0.2,0.4); ExogenousStates(injury);
An injury begins each period with probability 0.2. Once it starts it ends with probablity 0.4. The arguments can be functions that are called and compute probabilities that depend on states and actions.
b | base StateVariable |
Tprob | τ, probability of the special event |
rval | r, integer, value to jump to in the special event. -1, then the reset value this.N = b.N+1 and rval = b.N. Prob( q′= z | q,b ) = I{z=r}τ + (1-τ)Prob(b′=z) |
Enumerations | ||
Anonymous enum 1 | down, hold, up, NRUP |
L | label |
N | integer, number of values, N ≥ 3 |
fPi() | a AV()-compatible object that returns a rows(A) x 3 vector of probabilities.
|
Prune | TRUE [default], prune unreachable states assuming initial value of 0 FALSE do not prune |
v=1.0
.
y, | the current realized outcome, υ. |
L | string prefix or array of strings, individual labels |
vNorM | vector of integer values, number of values each effect takes on. OR, a matrix of actual values. |
UseOnlyObserved | [default=TRUE] if a matrix is sent as the second argument then TRUE means that only combinations (rows) of actual fixed effects in the matrix will be created and solved. Otherwise, all possible combinations of observed actual values will be solved for. |
x = new Regressors("X",<2,3,5>);Give each effect a label:
x = new Regressors({"Gender","Education","Occupation"},<2,3,5>);
Here is an approach to combining parameters (estimated or chosen) with fixed effects. Suppose a parameter ψ in the DP model depends on a subset of X variables, Xp and coefficients β. That is,
ψ = exp{ Xp ψ }For example, ψ is a function of an intercept and gender, but other X variables that do not affect ψ are in the model. The code segments below show how to give names to the columns of X, define the number of values each takes on and then how to make their current values determine the dynamically determined value of ψ.
enum{intercept,gender,race,sibling,test,NX} enum{Ni=1,Ng=2,Nr=2,Ns=2,Nt=3} ⋮ const decl psispec = intercept~gender; ⋮ beta = new Coefficients("B",columns(psispec)); GroupVariables(X = new Regressors("X",Ng~Nr~Ns~Nt)); X.Actual[][intercept] = 1; // replace fixed 0 with fixed 1 for the intercept. ⋮ psi = [=]() { return exp(AV(X)[psipsec]*CV(Beta)); };Note the last line uses the lambda function feature introduced in Ox 7. So psi() would return the dynamically determined value of ψ. The alternative is to define a static method which would have the same code.
L | string, state variable name |
N | integer, number of values |
reset | ActionVariable |
Pi, | vector or Simplex parameter block |
b | the base StateVariable whose transition is augmented. |
t | the ActionVariable that triggers a different transition
Reset(b,t) ≡ ActionTriggered(b,t,1,0)
|
matchvalue | |
acc, | either ActionVariable or an array of ActionVariables |
replace, | integer or vector of values that replace current value with matchvalue |
keep, | integer or vector of values that retain current value |
L | label |
N | Number of discrete points in the approximation |
pars | 3x1 vector or array of AV()-compatible parametersi: Parameter (default) 0: mean (μ=0.0) Actual values will take on N equally spaced values in a dynamically determined range The udpate code is based on these implementations: https://drive.google.com/file/d/0BzbiTUXVnwezQjBmU3p5NFVOYnM/view?resourcekey=0-IgQlYEXCqGMBYBGzd4Y4Zg https://quanteconpy.readthedocs.io/en/latest/_modules/quantecon/markov/approximation.html#rouwenhorst |
L | string, label |
N | integer, number of values the variable takes on.
Prob(q' = v) = 1 / N, for v=0,…,N‾ |
TFPshock = new SimpleJump("z",20); ExogenousStates(TFPshock);
Tomorrow TFPshock will take on values 0 to 19 each with probability 0.05
L | label |
N | integer, maximum value. |
State | StateVariable to accumulate values for. |
L | label |
Target | StateVariable to record |
Tbar | clock time at which to record choice (starting at Tbar+1) |
Prune | TRUE [default], prune unreachable states (non-zero values at and before Tbar)
s′={0 if t less than TbarTarget.v if t equals Tbars if t greater than Tbar
|
... | list of Coevolving state variables to add to the block. The default Actual matrix is built up from the actual vectors of variables added to the block. |
L | label for block |
... | list of Coevolving states to add to the block. |
This is a virtual method, so derived classes replace this default version with the one that goes along with the kind of state variable.
Alpha::N
× L or 1 × L
matrix of action-specific transition probabilities, Ρ(q′;α,η,θ). <0> , CondProbOne }
. That is the with
probability one the value of the state variable next period is 0.
L | label |
N | integer, maximum number of times to count |
State | StateVariable or AV()-compatible object to track. |
ToTrack | integer or vector, values of State to count. [default = <1>] DoAll: track all non-zero values |
Reset | AV()-compatible object that resets the count if TRUE. [Default=0] (no reset) |
Prune | TRUE [default]: prune states if finite horizon detected. |
noffers = new StateCounter("Total Offers",5,offer,DoAll);
L | label |
N | integer, maximum number of times to count |
State | StateVariable to track. |
ToTrack | integer or vector, values of State to count. |
Target | StateVariable to track actual values of. |
ToTrack | vector of (actual) values to track DoAll: track all non-zero values [defaut=<1>] |
Prune, | prune non-zero states at t==0 in finite horizon. |
actual
.
Called by Transitions() after Update() called. If the user's update function returns the wrong dimension this catches.
actual
should be N x 1 not 1 x N.
sv | StateVariable |
sv | StateVariable |
TermValues | integer vector in the range 0...N-1
Utility() must return a single terminating value of the process at a terminal value.
Utility of being newly deceased should return
the bequest value of the state.
|
s = new StateVariable("s",5); s->MakeTerminal(<3;4>); v = new StateVariable("v",1); v->MakeTerminal(1);
Now any state for which CV(s)=3
or CV(s)=4
or CV(v)=1
will be marked as terminal:
The classification of terminal values occurs during CreateSpaces()
and is stored at each
object of the MyModel
by setting Bellman::Type >= TERMINAL.
s
contains a state variable then
AV(s)
calls s->myAV()
actual[v]
.
N | positive integer the number of values the variable takes on. N=1 is a constant, which can be included as a placeholder for extensions of a model. |
L | string a label or name for the variable.
|
MyModel
would ever include a variable
of the base class.
LorC | either a label or a column number to match this variable to external data. This |
This is a virtual method, so derived classes replace this default version with the one that goes along with the kind of state variable.
Alpha::N
× L or 1 × L
matrix of action-specific transition probabilities, Ρ(q′;α,η,θ). <0> , CondProbOne }
. That is the with
probability one the value of the state variable next period is 0.
v=1.0
.
y, | the current realized outcome, υ. |
L | label string integer: get label from |
target | static function of the form target() |
StaticProgram
SetClock(StaticProgram);which is equivalent to this:
SetClock(new StaticP());which is also equivalent to this:
SetClock(NormalAging,1);
IsErgodic | TRUE, store Ρ* |
L | label |
N | number of values it takes on. |
b | base StateVariable |
s | integer the subperiod this variable is not frozen and goes throug a regular transition.
|
L | label |
N | Number of discrete points in the approximation |
M | AV()-compatiable max discrete value |
pars | 3x1 vector or array of AV()-compatible parametersi: Parameter (default) 0: mean (μ=0.0) Actual values will take on N equally spaced values in the range μ±Mσ/√(1−ρ2). The transition probabilities depends on the current value a la Tauchen. |
The user can supply a replacement in a derived class.
L | label |
N | number of points |
M | number of standard deviations to set the largest value as |
pars | 2x1 vector or array of AV()-compatible Normal distribution parameters, see NormalParams
actual values are equally spaced between -Mσ and Mσ.
The probabilities are not uniform. They are constant and depend only on N and M. |
b | the base StateVariable whose transition is augmented. |
t | the trigger for a different transition (action, state, or AV()-compatible object) |
tv | the integer (actual) value of t that triggers the transition [default=1] or vector of integer values that trigger |
rval | the integer (current) value of this state variable when the trigger occurs [default=0] -1, then the reset value this.N = b.N+1 and rval = b.N. |
Not called by user code
b | the base StateVariable whose transition is augmented (the base should not be added to the model separately). |
t | the AV()-compatible value which triggers the change. Usually this would be another state variable that is present in the model. |
tv | the integer (actual) or vector of integer values of tv that triggers the transition [default=1]. |
rval | the integer (current) value of this state variable when the trigger occurs [default=0] -1, then the reset value this.N = b.N+1 and rval = b.N. |
L | label |
N | number of points |
gamma | decay rate [default=1.0] |