
    6bim                        S r SSKJr  SSKJr  SSKJr  SSKrSSKrSSKJs  J	r
  SSKJr  SSKJr  SSKJr  SS	KJr  SS
KJr  SSKJr  SSKJr  SSKJr  SSKJr  SSKJr  SrSrS rS7S jr S r!\" S/S9\" S/S9 " S S\RD                  5      5       5       r# " S S\#5      r$\" S/S9\" S/S9 " S S\$5      5       5       r%\" S /S9\" S!/S9 " S" S#\$5      5       5       r&\RN                  " S$S%5      r(\" S&/S9\" S'/S9 " S( S$\(5      5       5       r)\" S)/S9\" S*/S9 " S+ S,\$5      5       5       r*\" S-/S9\" S./S9 " S/ S0\$5      5       5       r+\" S1/S9\" S2/S9 " S3 S4\#5      5       5       r,S5 r-S6 r.g)8a}  Legacy module implementing RNN Cells.

This module provides a number of basic commonly used RNN cells, such as LSTM
(Long Short Term Memory) or GRU (Gated Recurrent Unit), and a number of
operators that allow adding dropouts, projections, or embeddings for inputs.
Constructing multi-layer cells is supported by the class `MultiRNNCell`, or by
calling the `rnn` ops several times.
    )absolute_import)division)print_functionN)activations)backend)initializers)base_layer_utils)
input_spec)base)tf_utils)
tf_logging)keras_export)	tf_exportbiaskernelc                 <     [        X5        g! [         a     gf = f)NTF)getattrAttributeError)obj	attr_names     ^/home/james-whalen/.local/lib/python3.13/site-packages/tf_keras/src/layers/rnn/legacy_cells.py_hasattrr   3   s(       s    
c                    [        U [        R                  5      (       a  U n[        R                  " U 5      nUR                  R
                  S:X  a+  [        R                  R                  R                  US5      nOUR                  R
                  S:w  a  [        SU 35      eO}[        R                  " U 5      nUR
                  b  UR                  5       OSnUR                  5       (       a1  [        R                  " UR                  5       [        R                  S9OSn[        U[        R                  5      (       a  Un[        R                  " U5      nUR                  R
                  S:X  a+  [        R                  R                  R                  US5      nOUR                  R
                  S:w  a  [        SU 35      eO}[        R                  " U5      nUR
                  b  UR                  5       OSnUR                  5       (       a1  [        R                  " UR                  5       [        R                  S9OSnU(       aH  [        R                  " U5      R                  U5      nUR
                  b  UR                  5       nU$ SnU$ Ub  Uc  [        SU  SU 35      e[        R                   " X54S5      nU$ )	a  Concat that enables int, Tensor, or TensorShape values.

This function takes a size specification, which can be an integer, a
TensorShape, or a Tensor, and converts it into a concatenated Tensor
(if static = False) or a list of integers (if static = True).

Args:
  prefix: The prefix; usually the batch size (and/or time step size).
    (TensorShape, int, or Tensor.)
  suffix: TensorShape, int, or Tensor.
  static: If `True`, return a python list with possibly unknown dimensions.
    Otherwise return a `Tensor`.

Returns:
  shape: the concatenation of prefix and suffix.

Raises:
  ValueError: if `suffix` is not a scalar or vector (or TensorShape).
  ValueError: if prefix or suffix was `None` and asked for dynamic
    Tensors out.
r      zFPrefix tensor must be either a scalar or vector, but received tensor: NdtypezFsuffix tensor must be either a scalar or vector, but received tensor: z2Prefix or suffix can't be None. Received prefix = z and suffix = )
isinstancetfTensorget_static_valueshapendimscompatv1expand_dims
ValueErrorTensorShapeas_listis_fully_definedconstantint32concatenateconcat)prefixsuffixstaticpp_staticss_staticr!   s           r   _concatr5   <   sJ   , &"))$$&&v.77==A		((A.AWW]]a(()s,    NN6""#''"5199;4 !!## KK		2884 	

 &"))$$&&v.77==A		((A.AWW]]a(()s,    NN6""#''"5199;4 !!## KK		2884 	
 x(44X>#(;;#: L AE L 9	%%+HN6(D  		1&!$L    c                 R   ^^ UU4S jn[         R                  R                  X05      $ )zCCreate tensors of zeros based on state_size, batch_size, and dtype.c                    > [        TU 5      n[        R                  " UTS9n[        R                  " 5       (       d  [        TU SS9nUR	                  U5        U$ )z7Combine s with batch_size to get a proper tensor shape.r   T)r0   )r5   r   zerosexecuting_eagerly	set_shape)r3   csizec_static
batch_sizer   s       r   get_state_shape,_zero_state_tensors.<locals>.get_state_shape   sM    J"xx'##%%z1T:HNN8$r6   )r   nestmap_structure)
state_sizer?   r   r@   s    `` r   _zero_state_tensorsrE      s     77  ==r6   z*keras.__internal__.legacy.rnn_cell.RNNCell)r$   znn.rnn_cell.RNNCellc                      ^  \ rS rSrSrSU 4S jjrSU 4S jjrS r\S 5       r	\S 5       r
S rSS	 jrS
 rU 4S jr\S 5       rSrU =r$ )RNNCell   a#  Abstract object representing an RNN cell.

Every `RNNCell` must have the properties below and implement `call` with
the signature `(output, next_state) = call(input, state)`.  The optional
third input argument, `scope`, is allowed for backwards compatibility
purposes; but should be left off for new subclasses.

This definition of cell differs from the definition used in the literature.
In the literature, 'cell' refers to an object with a single scalar output.
This definition refers to a horizontal array of such units.

An RNN cell, in the most abstract setting, is anything that has
a state and performs some operation that takes a matrix of inputs.
This operation results in an output matrix with `self.output_size` columns.
If `self.state_size` is an integer, this operation also results in a new
state matrix with `self.state_size` columns.  If `self.state_size` is a
(possibly nested tuple of) TensorShape object(s), then it should return a
matching structure of Tensors having shape `[batch_size].concatenate(s)`
for each `s` in `self.batch_size`.
c                 :   > [         TU ]  " SXUS.UD6  SU l        g )N)	trainablenamer   T )super__init___is_tf_rnn_cell)selfrJ   rK   r   kwargs	__class__s        r   rN   RNNCell.__init__   s$    O9uOO  $r6   c                    > UbI  [         R                  R                  R                  X0R                  S9 n[
        TU ]  XUS9sSSS5        $ Sn[        XS5      nUcd  [         R                  R                  R                  [         R                  R                  R                  5       U R                  S9n[        XU5        U   [
        TU ]  X5      sSSS5        $ ! , (       d  f       g= f! , (       d  f       g= f)a  Run this RNN cell on inputs, starting from the given state.

Args:
  inputs: `2-D` tensor with shape `[batch_size, input_size]`.
  state: if `self.state_size` is an integer, this should be a
    `2-D Tensor` with shape `[batch_size, self.state_size]`. Otherwise,
    if `self.state_size` is a tuple of integers, this should be a tuple
    with shapes `[batch_size, s] for s in self.state_size`.
  scope: VariableScope for the created subgraph; None uses class name.
    Defaults to `None`.

Returns:
  A pair containing:

  - Output: A `2-D` tensor with shape `[batch_size, self.output_size]`.
  - New state: Either a single `2-D` tensor, or a tuple of tensors
    matching the arity and shapes of `state`.
N)custom_getter)scopernncell_scope)
r   r#   r$   variable_scope_rnn_get_variablerM   __call__r   get_variable_scopesetattr)rP   inputsstaterV   scope_attrnamerR   s        r   rZ   RNNCell.__call__   s    & ,,%;%; - w'U'C 
 -ND$7E}		33IILL335"&"8"8 4  e4w'6   s   CC/
C,/
C=c                   ^ U" U0 UD6m[         R                  R                  R                  5       (       a  TR                  nOT[         R                  R                  R                  5       ;   =(       dT    [        R                  " T5      =(       a7    [        T5      S   [         R                  R                  R                  5       ;   nU(       aA  [        U4S jU R                   5       5      (       a  U R                  R                  T5        T$ U(       d?  [        U4S jU R                   5       5      (       a  U R                  R                  T5        T$ )Nr   c              3   ,   >#    U  H	  nTULv   M     g 7fNrL   .0vvariables     r   	<genexpr>,RNNCell._rnn_get_variable.<locals>.<genexpr>   s      
'>!HA'>   c              3   ,   >#    U  H	  nTULv   M     g 7frc   rL   rd   s     r   rh   ri      s      #
'B!HA'Brj   )r   r#   r$   #executing_eagerly_outside_functionsrJ   trainable_variablesr	   is_split_variablelistall_trainable_weightsappend_non_trainable_weights)rP   getterargsrQ   rJ   rg   s        @r   rY   RNNCell._rnn_get_variable   s   4*6*99<<;;== **I BIILL$D$D$FF  228< LN1%)I)I)KK   
'+'>'>
 
 
 ##**84
 	 s #
'+'B'B#
  
  
 ''..x8r6   c                     [        S5      e)zsize(s) of state(s) used by this cell.

It can be represented by an Integer, a TensorShape or a tuple of
Integers or TensorShapes.
Abstract methodNotImplementedErrorrP   s    r   rD   RNNCell.state_size   s     ""344r6   c                     [        S5      e)z>Integer or TensorShape: size of outputs produced by this cell.rx   ry   r{   s    r   output_sizeRNNCell.output_size   s     ""344r6   c                     g rc   rL   )rP   _s     r   buildRNNCell.build   s     	r6   c                    UGb4  [         R                  " USS9nUb  [         R                  " U5      (       a  [         R                  " USS9nOUnUR                  R
                  S   R                  U:w  a2  [        SUR                  R
                  S   R                   SU 35      eUb+  UR                  U:w  a  [        SUR                   S	U 35      eUR                  R
                  S   R                  =(       d,    [         R                  R                  R	                  U5      S   nUR                  nUb  Uc  [        S
U SU 35      eU R                  X#5      $ )Nr]   rK   T)partialr   zTbatch size from input tensor is different from the input param. Input tensor batch: z, batch_size: zOdtype from input tensor is different from the input param. Input tensor dtype: z	, dtype: zQbatch_size and dtype cannot be None while constructing initial state: batch_size=z, dtype=)r   convert_to_tensor	is_tensorr    r!   dimsvaluer&   r   r#   r$   
zero_state)rP   r]   r?   r   static_batch_sizes        r   get_initial_stateRNNCell.get_initial_state   sl    ))&x@F%<<
++(*(;(;"D)% )3%<<$$Q'--1BB$<!<<,,Q/556 7''1l4   V\\U%: 88>~ F#W&  !!!$**Kbiill.@.@.H.K  LLE--7LI  z11r6   c                    U R                   n[        R                  " 5       nU(       a3  [        U S5      (       a"  [	        U S5      u  nnnnXa:X  a  Xr:X  a  XS:X  a  U$ [
        R                  " [        U 5      R                  S-   5         [        X1U5      n	SSS5        U(       a
  X1UW	4U l
        W	$ ! , (       d  f       N!= f)a  Return zero-filled state tensor(s).

Args:
  batch_size: int, float, or unit Tensor representing the batch size.
  dtype: the data type to use for the state.

Returns:
  If `state_size` is an int or TensorShape, then the return value is a
  `N-D` tensor of shape `[batch_size, state_size]` filled with zeros.

  If `state_size` is a nested list or tuple, then the return value is
  a nested list or tuple (of the same structure) of `2-D` tensors with
  the shapes `[batch_size, s]` for each s in `state_size`.
_last_zero_state	ZeroStateN)rD   r   r:   r   r   r   
name_scopetype__name__rE   r   )
rP   r?   r   rD   is_eagerlast_state_sizelast_batch_size
last_dtypelast_outputoutputs
             r   r   RNNCell.zero_state$  s    " __
'')'9:: 01  -'#1""T
 3 3k AB(GF C%/UF$KD!	 CBs   B00
B>c                     > [         TU ]  5       $ rc   )rM   
get_config)rP   rR   s    r   r   RNNCell.get_configJ  s    w!##r6   c                     g)NFrL   r{   s    r   !_use_input_spec_as_call_signature)RNNCell._use_input_spec_as_call_signatureM  s    
 r6   )rO   r   )TNNrc   )NNN)r   
__module____qualname____firstlineno____doc__rN   rZ   rY   propertyrD   r~   r   r   r   r   r   __static_attributes____classcell__rR   s   @r   rG   rG      sj    *$"7H& 5 5 5 5
$2L$L$  r6   rG   c                   "    \ rS rSrSrSS jrSrg)LayerRNNCelliU  a  Subclass of RNNCells that act like proper `tf.Layer` objects.

For backwards compatibility purposes, most `RNNCell` instances allow their
`call` methods to instantiate variables via `tf.compat.v1.get_variable`.
The underlying variable scope thus keeps track of any variables, and
returning cached versions.  This is atypical of `tf.layer` objects, which
separate this part of layer building into a `build` method that is only
called once.

Here we provide a subclass for `RNNCell` objects that act exactly as
`Layer` objects do.  They must provide a `build` method and their
`call` methods do not access Variables `tf.compat.v1.get_variable`.
Nc                 P    [         R                  R                  " XU/UQ7SU0UD6$ )a  Run this RNN cell on inputs, starting from the given state.

Args:
  inputs: `2-D` tensor with shape `[batch_size, input_size]`.
  state: if `self.state_size` is an integer, this should be a `2-D
    Tensor` with shape `[batch_size, self.state_size]`.  Otherwise, if
    `self.state_size` is a tuple of integers, this should be a tuple
    with shapes `[batch_size, s] for s in self.state_size`.
  scope: optional cell scope.
  *args: Additional positional arguments.
  **kwargs: Additional keyword arguments.

Returns:
  A pair containing:

  - Output: A `2-D` tensor with shape `[batch_size, self.output_size]`.
  - New state: Either a single `2-D` tensor, or a tuple of tensors
    matching the arity and shapes of `state`.
rV   )
base_layerLayerrZ   )rP   r]   r^   rV   ru   rQ   s         r   rZ   LayerRNNCell.__call__d  s:    . ((%
/3
',
7=
 	
r6   rL   rc   )r   r   r   r   r   rZ   r   rL   r6   r   r   r   U  s    
r6   r   z/keras.__internal__.legacy.rnn_cell.BasicRNNCellznn.rnn_cell.BasicRNNCellc                      ^  \ rS rSrSr    S
U 4S jjr\S 5       r\S 5       r\	R                  S 5       rS rU 4S jrS	rU =r$ )BasicRNNCelli  a  The most basic RNN cell.

Note that this cell is not optimized for performance.

Args:
  num_units: int, The number of units in the RNN cell.
  activation: Nonlinearity to use.  Default: `tanh`. It could also be string
    that is within TF-Keras activation function names.
  reuse: (optional) Python boolean describing whether to reuse variables in
    an existing scope. If not `True`, and the existing scope already has the
    given variables, an error is raised.
  name: String, the name of the layer. Layers with the same name will share
    weights, but to avoid mistakes we require reuse=True in such cases.
  dtype: Default dtype of the layer (default of `None` means use the type of
    the first input). Required when `build` is called before `call`.
  **kwargs: Dict, keyword named properties for common layer attributes, like
    `trainable` etc when constructing the cell from configs of get_config().
c                   > [         R                  " SSS9  [        TU ]  " SX4US.UD6  [	        U R
                  5        [        R                  " 5       (       a;  [        R                  R                  S5      (       a  [        R                  " SU 5        [        R                  " SS9U l        Xl        U(       a  [        R                   " U5      U l        g [        R$                  U l        g )	Nz`tf.nn.rnn_cell.BasicRNNCell` is deprecated and will be removed in a future version. This class is equivalent as `tf.keras.layers.SimpleRNNCell`, and will be replaced by that in Tensorflow 2.0.   
stacklevel_reuserK   r   GPUz9%s: Note that this cell is not optimized for performance.ndimrL   )warningswarnrM   rN   _check_supported_dtypesr   r   r:   configlist_logical_devicesloggingwarningr
   	InputSpec
_num_unitsr   get_activationtanh)rP   	num_units
activationreuserK   r   rQ   rR   s          r   rN   BasicRNNCell.__init__  s     	> 	
 	HHH

+!!bii&D&DU&K&KOOK %..A6#*z:D!wwDr6   c                     U R                   $ rc   r   r{   s    r   rD   BasicRNNCell.state_size      r6   c                     U R                   $ rc   r   r{   s    r   r~   BasicRNNCell.output_size  r   r6   c                 x   US   c  [        SU 35      e[        U R                  5        US   nU R                  [        X R
                  -   U R
                  /S9U l        U R                  [        U R
                  /[        R                  R                  R                  U R                  S9S9U l        SU l        g )N7Expected inputs.shape[-1] to be known, received shape: r!   r   r!   initializerT)r&   r   r   
add_weight_WEIGHTS_VARIABLE_NAMEr   _kernel_BIAS_VARIABLE_NAMEr   r#   r$   zeros_initializer_biasbuiltrP   inputs_shapeinput_depths      r   r   BasicRNNCell.build  s    ###/.2  	 

+"2&"0$//B ' 
 __??#		66TZZ6H % 

 
r6   c                    [        X/5        [        R                  " [        R                  " X/S5      U R                  5      n[        R
                  R                  X0R                  5      nU R                  U5      nXD4$ )zDMost basic RNN: output = new_state = act(W * input + U * state +
B).r   )	_check_rnn_cell_input_dtypesr   matmulr-   r   nnbias_addr   r   )rP   r]   r^   gate_inputsr   s        r   callBasicRNNCell.call  s]     	%f_5ii		6/1 =t||Leenn[**=!!+.~r6   c                   > U R                   [        R                  " U R                  5      U R                  S.n[
        TU ]  5       n[        [        UR                  5       5      [        UR                  5       5      -   5      $ )N)r   r   r   )
r   r   	serializer   r   rM   r   dictro   itemsrP   r   base_configrR   s      r   r   BasicRNNCell.get_config  sd    %//0@0@A[[

 g(*D**,-V\\^0DDEEr6   )r   r   r   r   r   r
   )NNNNr   r   r   r   r   rN   r   rD   r~   r   shape_type_conversionr   r   r   r   r   r   s   @r   r   r     so    , 'B     ## $*F Fr6   r   z*keras.__internal__.legacy.rnn_cell.GRUCellznn.rnn_cell.GRUCellc                      ^  \ rS rSrSr      S
U 4S jjr\S 5       r\S 5       r\	R                  S 5       rS rU 4S jrS	rU =r$ )GRUCelli  a  Gated Recurrent Unit cell.

Note that this cell is not optimized for performance. Please use
`tf.compat.v1.keras.layers.CuDNNGRU` for better performance on GPU, or
`tf.raw_ops.GRUBlockCell` for better performance on CPU.

Args:
  num_units: int, The number of units in the GRU cell.
  activation: Nonlinearity to use.  Default: `tanh`.
  reuse: (optional) Python boolean describing whether to reuse variables in
    an existing scope. If not `True`, and the existing scope already has
    the given variables, an error is raised.
  kernel_initializer: (optional) The initializer to use for the weight and
    projection matrices.
  bias_initializer: (optional) The initializer to use for the bias.
  name: String, the name of the layer. Layers with the same name will share
    weights, but to avoid mistakes we require reuse=True in such cases.
  dtype: Default dtype of the layer (default of `None` means use the type of
    the first input). Required when `build` is called before `call`.
  **kwargs: Dict, keyword named properties for common layer attributes, like
    `trainable` etc when constructing the cell from configs of get_config().
    References: Learning Phrase Representations using RNN Encoder Decoder
    for Statistical Machine Translation: [Cho et al., 2014]
    (https://aclanthology.coli.uni-saarland.de/papers/D14-1179/d14-1179)
    ([pdf](http://emnlp2014.org/papers/pdf/EMNLP2014179.pdf))
c                 D  > [         R                  " SSS9  [        T	U ]  " SX6US.UD6  [	        U R
                  5        [        R                  " 5       (       a;  [        R                  R                  S5      (       a  [        R                  " SU 5        [        R                  " SS9U l        Xl        U(       a  [        R                   " U5      U l        O[        R$                  U l        [&        R                   " U5      U l        [&        R                   " U5      U l        g )	Nz`tf.nn.rnn_cell.GRUCell` is deprecated and will be removed in a future version. This class is equivalent as `tf.keras.layers.GRUCell`, and will be replaced by that in Tensorflow 2.0.r   r   r   r   z%s: Note that this cell is not optimized for performance. Please use tf.compat.v1.keras.layers.CuDNNGRU for better performance on GPU.r   rL   )r   r   rM   rN   r   r   r   r:   r   r   r   r   r
   r   r   r   r   r   r   r   _kernel_initializer_bias_initializer)
rP   r   r   r   kernel_initializerbias_initializerrK   r   rQ   rR   s
            r   rN   GRUCell.__init__  s     	> 	
 	HHH

+!!bii&D&DU&K&KOO& 	 %..A6#*z:D!wwD#/#3#34F#G !-!1!12B!Cr6   c                     U R                   $ rc   r   r{   s    r   rD   GRUCell.state_size-  r   r6   c                     U R                   $ rc   r   r{   s    r   r~   GRUCell.output_size1  r   r6   c           	      8   US   c  [        SU 35      e[        U R                  5        US   nU R                  S[         3X R
                  -   SU R
                  -  /U R                  S9U l        U R                  S[         3SU R
                  -  /U R                  b  U R                  O1[        R                  R                  R                  SU R                  S9S9U l        U R                  S[         3X R
                  -   U R
                  /U R                  S9U l        U R                  S[         3U R
                  /U R                  b  U R                  O0[        R                  R                  R!                  U R                  S9S9U l        S	U l        g )
Nr   r   zgates/r   r         ?r   z
candidate/T)r&   r   r   r   r   r   r   _gate_kernelr   r   r   r#   r$   constant_initializer
_gate_bias_candidate_kernelr   _candidate_biasr   r   s      r   r   GRUCell.build5  s   ###/.2  	 

+"2& OO+,-0!doo2EF00 , 

 //()*t&' ))5 &&YY\\66s$**6M * 
 "&/010$//B00 "1 "

  $,-.??# ))5 &&YY\\33$**3E  /  
 
r6   c                 @   [        X/5        [        R                  " [        R                  " X/S5      U R                  5      n[        R
                  R                  X0R                  5      n[        R                  " U5      n[        R                  " USSS9u  pVXR-  n[        R                  " [        R                  " X/S5      U R                  5      n[        R
                  R                  XR                  5      nU R                  U5      n	Xb-  SU-
  U	-  -   n
X4$ )z-Gated recurrent unit (GRU) with nunits cells.r   r   r   num_or_size_splitsaxis)r   r   r   r-   r  r   r   r  sigmoidsplitr  r  r   )rP   r]   r^   r   r   rur_state	candidater<   new_hs              r   r   GRUCell.call]  s    $f_5iiIIvoq)4+<+<
 eenn[//B

;'xxeB)IIIIv'+T-C-C
	 EENN9.B.BC	Y'	QUaK'|r6   c                   > U R                   [        R                  " U R                  5      [        R                  " U R                  5      [
        R                  " U R                  5      U R                  S.n[        TU ]%  5       n[        [        UR                  5       5      [        UR                  5       5      -   5      $ )N)r   r   r   r   r   )r   r   r   r   r   r   r   r   rM   r   r   ro   r   r   s      r   r   GRUCell.get_configt  s    "."8"8((# !- 6 6t7M7M N%//0@0@A[[
 g(*D**,-V\\^0DDEEr6   )
r   r   r  r  r  r  r   r   r   r
   )NNNNNNr   r   s   @r   r   r     sw    < %DN     ##% $%N.F Fr6   r   LSTMStateTuple)r<   hz1keras.__internal__.legacy.rnn_cell.LSTMStateTupleznn.rnn_cell.LSTMStateTuplec                   ,    \ rS rSrSrSr\S 5       rSrg)r  i  zTuple used by LSTM Cells for `state_size`, `zero_state`, & output state.

Stores two elements: `(c, h)`, in that order. Where `c` is the hidden state
and `h` is the output.

Only used when `state_is_tuple=True`.
rL   c                     U u  pUR                   UR                   :w  a%  [        SUR                    SUR                    35      eUR                   $ )Nz(Inconsistent dtypes for internal state: z vs )r   	TypeError)rP   r<   r  s      r   r   LSTMStateTuple.dtype  sL    77agg:77)4y*  wwr6   N)	r   r   r   r   r   	__slots__r   r   r   rL   r6   r   r  r    s      I r6   z0keras.__internal__.legacy.rnn_cell.BasicLSTMCellznn.rnn_cell.BasicLSTMCellc                      ^  \ rS rSrSr      S
U 4S jjr\S 5       r\S 5       r\	R                  S 5       rS rU 4S jrS	rU =r$ )BasicLSTMCelli  a  DEPRECATED: Please use `tf.compat.v1.nn.rnn_cell.LSTMCell` instead.

Basic LSTM recurrent network cell.

The implementation is based on

We add forget_bias (default: 1) to the biases of the forget gate in order to
reduce the scale of forgetting in the beginning of the training.

It does not allow cell clipping, a projection layer, and does not
use peep-hole connections: it is the basic baseline.

For advanced models, please use the full `tf.compat.v1.nn.rnn_cell.LSTMCell`
that follows.

Note that this cell is not optimized for performance. Please use
`tf.compat.v1.keras.layers.CuDNNLSTM` for better performance on GPU, or
`tf.raw_ops.LSTMBlockCell` for better performance on CPU.
c                 ,  > [         R                  " SSS9  [        T	U ]  " S
XVUS.UD6  [	        U R
                  5        U(       d  [        R                  " SU 5        [        R                  " 5       (       a;  [        R                  R                  S5      (       a  [        R                  " SU 5        [        R                  " SS9U l        Xl        X l        X0l        U(       a  ["        R$                  " U5      U l        g	[        R(                  U l        g	)aY  Initialize the basic LSTM cell.

Args:
  num_units: int, The number of units in the LSTM cell.
  forget_bias: float, The bias added to forget gates (see above). Must
    set to `0.0` manually when restoring from CudnnLSTM-trained
    checkpoints.
  state_is_tuple: If True, accepted and returned states are 2-tuples of
    the `c_state` and `m_state`.  If False, they are concatenated along
    the column axis.  The latter behavior will soon be deprecated.
  activation: Activation function of the inner states.  Default: `tanh`.
    It could also be string that is within TF-Keras activation function
    names.
  reuse: (optional) Python boolean describing whether to reuse variables
    in an existing scope.  If not `True`, and the existing scope already
    has the given variables, an error is raised.
  name: String, the name of the layer. Layers with the same name will
    share weights, but to avoid mistakes we require reuse=True in such
    cases.
  dtype: Default dtype of the layer (default of `None` means use the
    type of the first input). Required when `build` is called before
    `call`.
  **kwargs: Dict, keyword named properties for common layer attributes,
    like `trainable` etc when constructing the cell from configs of
    get_config().  When restoring from CudnnLSTM-trained checkpoints,
    must use `CudnnCompatibleLSTMCell` instead.
z`tf.nn.rnn_cell.BasicLSTMCell` is deprecated and will be removed in a future version. This class is equivalent as `tf.keras.layers.LSTMCell`, and will be replaced by that in Tensorflow 2.0.r   r   r   _%s: Using a concatenated state is slower and will soon be deprecated.  Use state_is_tuple=True.r   %s: Note that this cell is not optimized for performance. Please use tf.compat.v1.keras.layers.CuDNNLSTM for better performance on GPU.r   NrL   )r   r   rM   rN   r   r   r   r   r   r:   r   r   r
   r   r   _forget_bias_state_is_tupler   r   r   r   )
rP   r   forget_biasstate_is_tupler   r   rK   r   rQ   rR   s
            r   rN   BasicLSTMCell.__init__  s    L 	> 	
 	HHH

+OO8
 !!bii&D&DU&K&KOO& 	 %..A6#'-*z:D!wwDr6   c                     U R                   (       a   [        U R                  U R                  5      $ SU R                  -  $ )Nr   )r"  r  r   r{   s    r   rD   BasicLSTMCell.state_size  s9     ## 4??DOO<	
 T__$	
r6   c                     U R                   $ rc   r   r{   s    r   r~   BasicLSTMCell.output_size  r   r6   c                    US   c  [        SU 35      e[        U R                  5        US   nU R                  nU R	                  [
        X#-   SU R                  -  /S9U l        U R	                  [        SU R                  -  /[        R                  R                  R                  U R                  S9S9U l        SU l        g )Nr   r      r   r   r   T)r&   r   r   r   r   r   r   r   r   r#   r$   r   r   r   )rP   r   r   h_depths       r   r   BasicLSTMCell.build  s    ###/.2  	 

+"2&//"(!doo*=> ' 
 __t&'		66TZZ6H % 

 
r6   c                    [        X/5        [        R                  n[        R                  " S[        R                  S9nU R
                  (       a  Uu  pVO[        R                  " USUS9u  pV[        R                  " [        R                  " X/S5      U R                  5      n[        R                  R                  XpR                  5      n[        R                  " USUS9u  pp[        R                  " U R                  U
R                  S9n[        R                  n[        R                   nU" U" XS" U" X5      5      5      U" U" U5      U R#                  U	5      5      5      nU" U R#                  U5      U" U5      5      nU R
                  (       a  [%        UU5      nUU4$ [        R                  " UU/S5      nUU4$ )a  Long short-term memory cell (LSTM).

Args:
  inputs: `2-D` tensor with shape `[batch_size, input_size]`.
  state: An `LSTMStateTuple` of state tensors, each shaped `[batch_size,
    num_units]`, if `state_is_tuple` has been set to `True`.  Otherwise,
    a `Tensor` shaped `[batch_size, 2 * num_units]`.

Returns:
  A pair containing the new hidden state, and the new state (either a
    `LSTMStateTuple` or a concatenated state, depending on
    `state_is_tuple`).
r   r   r   r  r+  )r   r   r  r*   r+   r"  r  r   r-   r   r   r   r   r!  r   addmultiplyr   r  )rP   r]   r^   r  oner<   r  r   ijfoforget_bias_tensorr/  r0  new_cr  	new_states                     r   r   BasicLSTMCell.call  sf    	%f_5**kk!288,DAq88%ACHDAii		6+q 94<<Heenn[**= XXKACP
a[[):):!''J ff;;QA :;<WQZ!1!1!!45
 ))%0'!*=&ue4I i 		5%.!4Iir6   c                 4  > U R                   U R                  U R                  [        R                  " U R
                  5      U R                  S.n[        TU ]!  5       n[        [        UR                  5       5      [        UR                  5       5      -   5      $ )N)r   r#  r$  r   r   )r   r!  r"  r   r   r   r   rM   r   r   ro   r   r   s      r   r   BasicLSTMCell.get_configM  sv    ,,"22%//0@0@A[[
 g(*D**,-V\\^0DDEEr6   )r   r   r!  r   r   r"  r   r
   )r   TNNNNr   r   s   @r   r  r    sw    . F'P 
 
   ## $*- ^	F 	Fr6   r  z+keras.__internal__.legacy.rnn_cell.LSTMCellznn.rnn_cell.LSTMCellc                      ^  \ rS rSrSr             S
U 4S jjr\S 5       r\S 5       r\	R                  S 5       rS rU 4S jrS	rU =r$ )LSTMCelliY  a  Long short-term memory unit (LSTM) recurrent network cell.

The default non-peephole implementation is based on (Gers et al., 1999).
The peephole implementation is based on (Sak et al., 2014).

The class uses optional peep-hole connections, optional cell clipping, and
an optional projection layer.

Note that this cell is not optimized for performance. Please use
`tf.compat.v1.keras.layers.CuDNNLSTM` for better performance on GPU, or
`tf.raw_ops.LSTMBlockCell` for better performance on CPU.
References:
  Long short-term memory recurrent neural network architectures for large
  scale acoustic modeling:
    [Sak et al., 2014]
    (https://www.isca-speech.org/archive/interspeech_2014/i14_0338.html)
    ([pdf]
    (https://www.isca-speech.org/archive/archive_papers/interspeech_2014/i14_0338.pdf))
  Learning to forget:
    [Gers et al., 1999]
    (http://digital-library.theiet.org/content/conferences/10.1049/cp_19991218)
    ([pdf](https://arxiv.org/pdf/1409.2329.pdf))
  Long Short-Term Memory:
    [Hochreiter et al., 1997]
    (https://www.mitpressjournals.org/doi/abs/10.1162/neco.1997.9.8.1735)
    ([pdf](http://ml.jku.at/publications/older/3504.pdf))
c                 z  > [         R                  " SSS9  [        TU ]  " SXUS.UD6  [	        U R
                  5        U
(       d  [        R                  " SU 5        Uc  Ub  [        R                  " SU 5        [        R                  " 5       (       a;  [        R                  R                  S5      (       a  [        R                  " S	U 5        [        R                  " SS
9U l        Xl        X l        X0l        ["        R$                  " U5      U l        XPl        X`l        Xpl        Xl        Xl        Xl        U(       a  [4        R$                  " U5      U l        O[        R8                  U l        U(       a"  U
(       a  [;        X5      OX-   U l        XPl        gU
(       a  [;        X5      OSU-  U l        Xl        g)a  Initialize the parameters for an LSTM cell.

Args:
  num_units: int, The number of units in the LSTM cell.
  use_peepholes: bool, set True to enable diagonal/peephole connections.
  cell_clip: (optional) A float value, if provided the cell state is
    clipped by this value prior to the cell output activation.
  initializer: (optional) The initializer to use for the weight and
    projection matrices.
  num_proj: (optional) int, The output dimensionality for the projection
    matrices.  If None, no projection is performed.
  proj_clip: (optional) A float value.  If `num_proj > 0` and
    `proj_clip` is provided, then the projected values are clipped
    elementwise to within `[-proj_clip, proj_clip]`.
  num_unit_shards: Deprecated, will be removed by Jan. 2017. Use a
    variable_scope partitioner instead.
  num_proj_shards: Deprecated, will be removed by Jan. 2017. Use a
    variable_scope partitioner instead.
  forget_bias: Biases of the forget gate are initialized by default to 1
    in order to reduce the scale of forgetting at the beginning of the
    training. Must set it manually to `0.0` when restoring from
    CudnnLSTM trained checkpoints.
  state_is_tuple: If True, accepted and returned states are 2-tuples of
    the `c_state` and `m_state`.  If False, they are concatenated along
    the column axis.  This latter behavior will soon be deprecated.
  activation: Activation function of the inner states.  Default: `tanh`.
    It could also be string that is within TF-Keras activation function
    names.
  reuse: (optional) Python boolean describing whether to reuse variables
    in an existing scope.  If not `True`, and the existing scope already
    has the given variables, an error is raised.
  name: String, the name of the layer. Layers with the same name will
    share weights, but to avoid mistakes we require reuse=True in such
    cases.
  dtype: Default dtype of the layer (default of `None` means use the
    type of the first input). Required when `build` is called before
    `call`.
  **kwargs: Dict, keyword named properties for common layer attributes,
    like `trainable` etc when constructing the cell from configs of
    get_config().  When restoring from CudnnLSTM-trained checkpoints,
    use `CudnnCompatibleLSTMCell` instead.
z`tf.nn.rnn_cell.LSTMCell` is deprecated and will be removed in a future version. This class is equivalent as `tf.keras.layers.LSTMCell`, and will be replaced by that in Tensorflow 2.0.r   r   r   r  Nz%s: The num_unit_shards and proj_unit_shards parameters are deprecated and will be removed in Jan 2017.  Use a variable scope with a partitioner instead.r   r   r   rL   ) r   r   rM   rN   r   r   r   r   r   r:   r   r   r
   r   r   _use_peepholes
_cell_clipr   r   _initializer	_num_proj
_proj_clip_num_unit_shards_num_proj_shardsr!  r"  r   r   r   r  _state_size_output_size)rP   r   use_peepholes	cell_clipr   num_proj	proj_clipnum_unit_shardsnum_proj_shardsr#  r$  r   r   rK   r   rQ   rR   s                   r   rN   LSTMCell.__init__x  s~   x 	> 	
 	HHH

+OO8
 &/*EOOC 	 !!bii&D&DU&K&KOO& 	 %..A6#+#(,,[9!# / /'-*z:D!wwD " y3) 
 !) " y4] 
 !*r6   c                     U R                   $ rc   )rF  r{   s    r   rD   LSTMCell.state_size  s    r6   c                     U R                   $ rc   )rG  r{   s    r   r~   LSTMCell.output_size  s       r6   c                    US   c  [        SU 35      e[        U R                  5        US   nU R                  c  U R                  OU R                  nU R
                  b3  [        R                  R                  R                  U R
                  5      OS nU R                  [        X#-   SU R                  -  /U R                  US9U l        U R                  c%  [        R                  R                  R                  nO1[        R                  R                  R                  U R                  S9nU R                  [        SU R                  -  /US9U l        U R"                  (       a  U R                  SU R                  /U R                  S9U l        U R                  SU R                  /U R                  S9U l        U R                  S	U R                  /U R                  S9U l        U R                  b  U R*                  b3  [        R                  R                  R                  U R*                  5      OS nU R                  S
[         3U R                  U R                  /U R                  US9U l        SU l        g )Nr   r   r+  )r!   r   partitionerr   r   w_f_diagw_i_diagw_o_diagzprojection/T)r&   r   r   rB  r   rD  r   r#   r$   fixed_size_partitionerr   r   rA  r   r   r   r   r?  	_w_f_diag	_w_i_diag	_w_o_diagrE  _proj_kernelr   )rP   r   r   r,  maybe_partitionerr   maybe_proj_partitioners          r   r   LSTMCell.build  sH   ###/.2  	 

+"2&%)^^%;$// $$0 IILL//0E0EF 	
 "(!doo*=>)))	 ' 
 ::)),,88K)),,88tzz8JK__t&'# % 


 !__' -- - DN
 "__' -- - DN
 "__' -- - DN >>% ((4 		33D4I4IJ #
 !%4567 --2	 !0 !D 
r6   c                 d   [        X/5        U R                  c  U R                  OU R                  n[        R                  nU R
                  (       a  Uu  pVOL[        R                  " USS/SU R                  /5      n[        R                  " USU R                  /SU/5      nUR                  5       R                  S5      R                  S   R                  nUc  [        SUR                  5        35      e[        R                  " [        R                  " X/S5      U R                  5      n[        R                  R!                  XR"                  5      n[        R$                  " USSS9u  ppU R&                  (       aR  U" XR(                  -   U R*                  U-  -   5      U-  U" XR,                  U-  -   5      U R/                  U
5      -  -   nO2U" XR(                  -   5      U-  U" U	5      U R/                  U
5      -  -   nU R0                  b,  [        R2                  " XR0                  * U R0                  5      nU R&                  (       a*  U" XR4                  U-  -   5      U R/                  U5      -  nOU" U5      U R/                  U5      -  nU R                  bY  [        R                  " XR6                  5      nU R8                  b,  [        R2                  " XR8                  * U R8                  5      nU R
                  (       a  [;        X5      O[        R                  " X/S5      nX4$ )aA  Run one step of LSTM.

Args:
  inputs: input Tensor, must be 2-D, `[batch, input_size]`.
  state: if `state_is_tuple` is False, this must be a state Tensor,
    `2-D, [batch, state_size]`.  If `state_is_tuple` is True, this must
    be a tuple of state Tensors, both `2-D`, with column sizes `c_state`
    and `m_state`.

Returns:
  A tuple containing:

  - A `2-D, [batch, output_dim]`, Tensor representing the output of the
    LSTM after reading `inputs` when previous state was `state`.
    Here output_dim is:
       num_proj if num_proj was set,
       num_units otherwise.
  - Tensor(s) representing the new state of LSTM after reading `inputs`
    when the previous state was `state`.  Same type and shape(s) as
    `state`.

Raises:
  ValueError: If input size cannot be inferred from inputs via
    static shape inference.
r   r   r   r   zMCould not infer input size from inputs.get_shape()[-1].Received input shape: r+  r  )r   rB  r   r   r  r"  slice	get_shape	with_rankr   r   r&   r   r-   r   r   r   r   r  r?  r!  rY  rZ  r   r@  clip_by_valuer[  r\  rC  r  )rP   r]   r^   rJ  r  c_prevm_prev
input_sizelstm_matrixr2  r3  r4  r5  r<   mr8  s                   r   r   LSTMCell.call9  s   4 	%f_5&*nn&<4??$..**$VVXXeaVb$//-BCFXXea%9B>JF%%'11!499!<BB
))/)9)9);(<>  ii		6*:A >Meenn[**=XXKAAN
a%%%(?? NNV++  A ---.7';  #;$ $A ??&  __$4dooFANNQ../$2B2B12EEA
T--a00A>>%		!../A*$$Q(8$//J ## 1 A61% 	
 |r6   c                   > U R                   U R                  U R                  [        R                  " U R
                  5      U R                  U R                  U R                  U R                  U R                  U R                  [        R                  " U R                  5      U R                  S.n[        TU ]A  5       n[#        [%        UR'                  5       5      [%        UR'                  5       5      -   5      $ )N)r   rH  rI  r   rJ  rK  rL  rM  r#  r$  r   r   )r   r?  r@  r   r   rA  rB  rC  rD  rE  r!  r"  r   r   r   rM   r   r   ro   r   r   s      r   r   LSTMCell.get_config  s    !00'11$2C2CD#44#44,,"22%//0@0@A[[
 g(*D**,-V\\^0DDEEr6   )r   r   r@  r!  rA  r   rB  rE  rD  r   rG  rC  r\  r"  rF  r?  rY  rZ  r[  r   r
   )FNNNNNNr   TNNNNr   r   s   @r   r=  r=  Y  s    > y*v     ! ! ##; $;zRhF Fr6   r=  z/keras.__internal__.legacy.rnn_cell.MultiRNNCellznn.rnn_cell.MultiRNNCellc                      ^  \ rS rSrSrSU 4S jjr\S 5       r\S 5       rU 4S jr	\S 5       r
\S 5       rS	 rS
rU =r$ )MultiRNNCelli  zRNN cell composed sequentially of multiple simple cells.

Example:

```python
num_units = [128, 64]
cells = [BasicLSTMCell(num_units=n) for n in num_units]
stacked_rnn_cell = MultiRNNCell(cells)
```
c                   > [         R                  " S5        [        TU ]  5         U(       d  [	        S5      e[
        R                  R                  U5      (       d  [        SU S35      e[        [        S U 5       5      5      [        U5      :  a&  [         R                  " [         R                  SS5        Xl        [        U R                  5       HN  u  p4[        U[
        R                   R"                  R$                  5      (       d  M:  U R'                  USU4-  S	9  MP     X l        U(       dT  [+        S
 U R                   5       5      (       a2  [	        SU R                   Vs/ s H  oUR,                  PM     sn 35      eggs  snf )a  Create a RNN cell composed sequentially of a number of RNNCells.

Args:
  cells: list of RNNCells that will be composed in this order.
  state_is_tuple: If True, accepted and returned states are n-tuples,
    where `n = len(cells)`.  If False, the states are all concatenated
    along the column axis.  This latter behavior will soon be
    deprecated.

Raises:
  ValueError: if cells is empty (not allowed), or at least one of the
    cells returns a state tuple but the flag `state_is_tuple` is
    `False`.
z`tf.nn.rnn_cell.MultiRNNCell` is deprecated. This class is equivalent as `tf.keras.layers.StackedRNNCells`, and will be replaced by that in Tensorflow 2.0.z0Must specify at least one cell for MultiRNNCell.z-cells must be a list or tuple, but received: .c              3   8   #    U  H  n[        U5      v   M     g 7frc   )idre   cells     r   rh   (MultiRNNCell.__init__.<locals>.<genexpr>  s     .2d88   zWAt least two cells provided to MultiRNNCell are the same object and will share weights.r   zcell-%dr   c              3   t   #    U  H.  n[         R                  R                  UR                  5      v   M0     g 7frc   )r   rB   	is_nestedrD   )re   r<   s     r   rh   ru    s&     HKq277$$Q\\22Ks   68z]Some cells return tuples of states, but the flag state_is_tuple is not set. State sizes are: N)r   r   rM   rN   r&   r   rB   rx  r  lensetlog_first_nWARN_cells	enumerater   __internal__tracking	Trackable_track_trackabler"  anyrD   )rP   cellsr$  cell_numberrt  r<   rR   s         r   rN   MultiRNNCell.__init__  sU    	>	

 	OPPww  ''?waH  s.../#e*<>	 !*4;;!7K $ 8 8 B BCC%%dk^1K%L "8  .HDKKHHH (?C{{(K{!{(K'LN  I 
 )Ls   (F
c                     U R                   (       a  [        S U R                   5       5      $ [        S U R                   5       5      $ )Nc              3   8   #    U  H  oR                   v   M     g 7frc   rD   rs  s     r   rh   *MultiRNNCell.state_size.<locals>.<genexpr>  s     A[T[rv  c              3   8   #    U  H  oR                   v   M     g 7frc   r  rs  s     r   rh   r    s     ?;4;rv  )r"  tupler}  sumr{   s    r   rD   MultiRNNCell.state_size  s5    AT[[AAA?4;;???r6   c                 4    U R                   S   R                  $ )Nr   )r}  r~   r{   s    r   r~   MultiRNNCell.output_size  s    {{2***r6   c                 *  >^^ [         R                  " [        U 5      R                  S-   5         U R                  (       a)  [        UU4S jU R                   5       5      sS S S 5        $ [        TU ]!  TT5      sS S S 5        $ ! , (       d  f       g = f)Nr   c              3   F   >#    U  H  oR                  TT5      v   M     g 7frc   )r   )re   rt  r?   r   s     r   rh   *MultiRNNCell.zero_state.<locals>.<genexpr>  s       CN4OOJ66;s   !)	r   r   r   r   r"  r  r}  rM   r   )rP   r?   r   rR   s    ``r   r   MultiRNNCell.zero_state  si    T
 3 3k AB## CG;;  CB w)*e< CBBs   1B+B
Bc                     U R                   (       d  / $ / nU R                   H2  n[        U[        R                  5      (       d  M$  XR
                  -  nM4     U$ rc   )rJ   r}  r   r   r   trainable_weights)rP   weightsrt  s      r   r  MultiRNNCell.trainable_weights  sI    ~~IKKD$
 0 011111   r6   c                 @   / nU R                    H2  n[        U[        R                  5      (       d  M$  XR                  -  nM4     U R
                  (       dH  / nU R                    H2  n[        U[        R                  5      (       d  M$  X2R                  -  nM4     X1-   $ U$ rc   )r}  r   r   r   non_trainable_weightsrJ   r  )rP   r  rt  r  s       r   r  "MultiRNNCell.non_trainable_weights  s    KKD$
 0 011555   ~~ "dJ$4$455%)?)??% $ %..r6   c                    SnUn/ n[        U R                  5       H  u  pg[        R                  R                  R                  SU-  5         U R                  (       aM  [        R                  R                  U5      (       d$  [        S[        U R                  5       SU 35      eX&   nO4[        R                  " USU/SUR                  /5      nX7R                  -  nU" XH5      u  pIUR                  U	5        SSS5        M     U R                  (       a  [        U5      O[        R                  " US5      nXE4$ ! , (       d  f       GM1  = f)z9Run this multi-layer cell on inputs, starting from state.r   zcell_%dz'Expected state to be a tuple of length z, but received: r   Nr   )r~  r}  r   r#   r$   rX   r"  rB   rx  r&   ry  rD   ra  rr   r  r-   )
rP   r]   r^   cur_state_poscur_inp
new_statesr2  rt  	cur_stater8  s
             r   r   MultiRNNCell.call  s"   
 -GA,,Y];''77,,U33(E"4??34.ug7 
 !&I "=1B3H!I "__4M%)'%="!!), <; .( ## *:q) 	 ""/ <;s   B.D==
E	)r}  r"  )T)r   r   r   r   r   rN   r   rD   r~   r   r  r  r   r   r   r   s   @r   rn  rn    so    	2h @ @ + +
=    # #r6   rn  c                 z    [         R                  R                  U 5       H  n[        UR                  5        M     g)a  Check whether the input tensors are with supported dtypes.

Default RNN cells only support floats and complex as its dtypes since the
activation function (tanh and sigmoid) only allow those types. This function
will throw a proper error message if the inputs is not in a supported type.

Args:
  inputs: tensor or nested structure of tensors that are feed to RNN cell as
    input or state.

Raises:
  ValueError: if any of the input tensor are not having dtypes of float or
    complex.
N)r   rB   flattenr   r   )r]   ts     r   r   r   1  s'     WW__V$( %r6   c                     U c  g [         R                  " U 5      n U R                  (       d   U R                  (       d  [	        SU  35      eg g )NzBRNN cell only supports floating point inputs, but received dtype: )r   as_dtypeis_floating
is_complexr&   r   s    r   r   r   D  sM    }KKE!1!1##('+
 	
 "2r6   )F)/r   
__future__r   r   r   collectionsr   tensorflow.compat.v2r#   v2r   tf_keras.srcr   r   r   tf_keras.src.enginer	   r
   tf_keras.src.legacy_tf_layersr   r   tf_keras.src.utilsr   tensorflow.python.platformr   r    tensorflow.python.util.tf_exportr   r   r   r   r   r5   rE   r   rG   r   r   r   
namedtuple_LSTMStateTupler  r  r=  rn  r   r   rL   r6   r   <module>r     s.   '  %   ! ! $   % 0 * < ' = 9 6 ! EP> >?@
$%&}j } ' A}@(
7 (
V CDE
)*+cF< cF , FcFL >?@
$%&UFl UF ' AUFp (()9:F EFG
+,-_  . H, DEF
*+,wFL wF - GwFt ?@A
%&'BF| BF ( BBFJ
 CDE
)*+L#7 L# , FL#^)&
r6   