ó
áp7]c           @   s€   d  Z  d d l Z d d l m Z d d l j j Z e d d e e d d „ Z
 d „  Z d „  Z d „  Z d	 „  Z d
 „  Z d S(   sR   
Holds files for l1 regularization of LikelihoodModel, using
scipy.optimize.slsqp
iÿÿÿÿN(   t
   fmin_slsqpiè  c   '         sy  t  j | ƒ j d ƒ } t | ƒ ‰ t  j | t  j | ƒ ƒ } t  j | d ƒ j d ƒ ‰  ˆ  t  j ˆ ƒ ‰  ˆ  j ƒ  d k s† t ‚ t	 | | ƒ } | j
 d d ƒ } ‡  ‡ ‡ ‡ f d †  } ‡ f d †  } ‡  ‡ ‡ f d †  } ‡ f d	 †  } t | | d
 | d | d | d | d | d |	 d | ƒ} t  j | d ˆ  ƒ } | d } | d } t j | ˆ  ˆ | | ƒ } | d } | d } | d } t j | ˆ ˆ  ˆ | | | | ƒ \ } } |	 ra| \ } } } } } | t  j | ƒ ƒ }  | d k }! t | ƒ d | }" | }# t d ƒ }$ t d ƒ }% i |  d 6|! d 6|# d 6|$ d 6|% d 6| d 6|" d 6}& n  |	 rq| |& f S| Sd S(   s  
    Solve the l1 regularized problem using scipy.optimize.fmin_slsqp().

    Specifically:  We convert the convex but non-smooth problem

    .. math:: \min_\beta f(\beta) + \sum_k\alpha_k |\beta_k|

    via the transformation to the smooth, convex, constrained problem in twice
    as many variables (adding the "added variables" :math:`u_k`)

    .. math:: \min_{\beta,u} f(\beta) + \sum_k\alpha_k u_k,

    subject to

    .. math:: -u_k \leq \beta_k \leq u_k.

    Parameters
    ----------
    All the usual parameters from LikelhoodModel.fit
    alpha : non-negative scalar or numpy array (same size as parameters)
        The weight multiplying the l1 penalty term
    trim_mode : 'auto, 'size', or 'off'
        If not 'off', trim (set to zero) parameters that would have been zero
            if the solver reached the theoretical minimum.
        If 'auto', trim params using the Theory above.
        If 'size', trim params if they have very small absolute value
    size_trim_tol : float or 'auto' (default = 'auto')
        For use when trim_mode === 'size'
    auto_trim_tol : float
        For sue when trim_mode == 'auto'.  Use
    qc_tol : float
        Print warning and don't allow auto trim when (ii) in "Theory" (above)
        is violated by this much.
    qc_verbose : Boolean
        If true, print out a full QC report upon failure
    acc : float (default 1e-6)
        Requested accuracy as used by slsqp
    t   Ft   alpha_rescaledi    t   accg»½×Ùß|Û=c            s   t  ˆ |  ˆ ˆ  ˆ Œ S(   N(   t   _objective_func(   t   x_full(   t   alphat   argst   ft   k_params(    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyt   <lambda>F   t    c            s   t  |  ˆ  ƒ S(   N(   t
   _f_ieqcons(   R   (   R	   (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR
   G   R   c            s   t  ˆ |  ˆ ˆ  ƒ S(   N(   t   _fprime(   R   (   R   R	   t   score(    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR
   H   R   c            s   t  |  ˆ  ƒ S(   N(   t   _fprime_ieqcons(   R   (   R	   (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR
   I   R   t	   f_ieqconst   fprimet   itert   dispt   full_outputt   fprime_ieqconst   qc_tolt
   qc_verboset	   trim_modet   size_trim_tolt   auto_trim_tolt    t   nant   foptt	   convergedt
   iterationst   goptt   hoptt   trimmedt   warnflagN(   t   npt   arrayt   ravelt   lent   appendt   fabst   onest   mint   AssertionErrort   _get_disp_slsqpt
   setdefaultR    t   asarrayt   l1_solvers_commont
   qc_resultst   do_trim_paramst   strt   float('   R   R   t   start_paramsR   t   kwargsR   t   maxitert   callbackt   retallR   t   hesst   x0t
   disp_slsqpR   t   funct   f_ieqcons_wrapt   fprime_wrapt   fprime_ieqcons_wrapt   resultst   paramsR   R   t   passedR   R   R   R"   R   t   fxt   itst   imodet   smodeR   R   R#   R   R    R!   t   retvals(    (   R   R   R   R	   R   s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyt   fit_l1_slsqp
   sT    )	





c         C   s7   |  s | r- |  r d } n  | r3 d } q3 n d } | S(   Ni   i   i    (    (   R   R9   R<   (    (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR-   w   s    	c         G   s/   | |  } | | } |  | | Œ | | j  ƒ  S(   s,   
    The regularized objective function
    (   t   sum(   R   R   R	   R   R   t   x_paramst   x_added(    (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR   ‚   s    

c         C   s    | |  } t  j |  | ƒ | ƒ S(   s$   
    The regularized derivative
    (   R$   R(   (   R   R   R	   R   RK   (    (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR   Œ   s    
c         C   s,   |  |  } |  | } t  j | | | | ƒ S(   s%   
    The inequality constraints.
    (   R$   R(   (   R   R	   RK   RL   (    (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR   •   s    

c         C   se   t  j | ƒ } t  j | | f d d ƒ} t  j | | f d d ƒ} t  j | | f d d ƒ} | S(   s2   
    Derivative of the inequality constraints
    t   axisi   i    (   R$   t   eyet   concatenate(   R   R	   t   It   At   Bt   C(    (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyR   Ÿ   s
    (   t   __doc__t   numpyR$   t   scipy.optimizeR    t"   statsmodels.base.l1_solvers_commont   baseR0   t   Falset   NoneRI   R-   R   R   R   R   (    (    (    s8   lib/python2.7/site-packages/statsmodels/base/l1_slsqp.pyt   <module>   s   k		
			
