
    h$,f                         d Z ddlmZmZ ddlZddlmZ ddlm	Z	 ddl
mZmZmZ ddlmZmZmZ dd	lmZ dd
lmZ  G d deeee      Zy)z)Principal Component Analysis Base Classes    )ABCMetaabstractmethodN)linalg)issparse   )BaseEstimatorClassNamePrefixFeaturesOutMixinTransformerMixin)_add_to_diagonaldeviceget_namespace)_implicit_column_offset)check_is_fittedc                   J    e Zd ZdZd Zd Zed	d       Zd Zd Z	e
d        Zy)
_BasePCAzwBase class for PCA methods.

    Warning: This class should not be used directly.
    Use derived classes instead.
    c           
         t        | j                        \  }}| j                  }| j                  }| j                  r)||j	                  |ddt
        j                  f         z  }|| j                  z
  }|j                  || j                  kD  ||j                  dt        |                  }|j                  |z  |z  }t        || j                  |       |S )as  Compute data covariance with the generative model.

        ``cov = components_.T * S**2 * components_ + sigma2 * eye(n_features)``
        where S**2 contains the explained variances, and sigma2 contains the
        noise variances.

        Returns
        -------
        cov : array of shape=(n_features, n_features)
            Estimated covariance of data.
        N        r   )r   components_explained_variance_whitensqrtnpnewaxisnoise_variance_whereasarrayr   Tr   )selfxp_r   exp_varexp_var_diffcovs          ;lib/python3.12/site-packages/sklearn/decomposition/_base.pyget_covariancez_BasePCA.get_covariance    s     d../A&&**;;%2::0F(GGK!5!55xxd***JJs6'?J3

 }}|+{:d22B7
    c           
      l   t        | j                        \  }}| j                  j                  d   }| j                  dk(  r|j	                  |      | j
                  z  S |r|j                  j                  }nt        j                  }| j
                  dk(  r || j                               S | j                  }| j                  }| j                  r)||j                  |ddt        j                  f         z  }|| j
                  z
  }|j                  || j
                  kD  ||j                  dt!        |                  }||j"                  z  | j
                  z  }t%        |d|z  |       |j"                   ||      z  |z  }|| j
                  dz   z  }t%        |d| j
                  z  |       |S )a8  Compute data precision matrix with the generative model.

        Equals the inverse of the covariance but computed with
        the matrix inversion lemma for efficiency.

        Returns
        -------
        precision : array, shape=(n_features, n_features)
            Estimated precision of data.
           r   r   Nr   g      ?r   )r   r   shapen_components_eyer   r   invr&   r   r   r   r   r   r   r   r   r   r   )	r   r    is_array_api_compliant
n_features
linalg_invr   r"   r#   	precisions	            r%   get_precisionz_BasePCA.get_precision<   s    &343C3C%D""%%++A.
 "66*%(<(<<<!JJ3&d11344 &&**;;%2::0F(GGK!5!55xxd***JJs6'?J3

  +--/$2F2FF	C,$6;MMJy$99KG	t++Q.//	C$*>*>$>Cr'   Nc                      y)a  Placeholder for fit. Subclasses should implement this method!

        Fit the model with X.

        Parameters
        ----------
        X : array-like of shape (n_samples, n_features)
            Training data, where `n_samples` is the number of samples and
            `n_features` is the number of features.

        Returns
        -------
        self : object
            Returns the instance itself.
        N )r   Xys      r%   fitz_BasePCA.fiti   s    r'   c                    t        |      \  }}t        |        | j                  |d|j                  |j                  gd      }| j
                  1t        |      rt        || j
                        }n|| j
                  z
  }|| j                  j                  z  }| j                  r||j                  | j                        z  }|S )a  Apply dimensionality reduction to X.

        X is projected on the first principal components previously extracted
        from a training set.

        Parameters
        ----------
        X : {array-like, sparse matrix} of shape (n_samples, n_features)
            New data, where `n_samples` is the number of samples
            and `n_features` is the number of features.

        Returns
        -------
        X_new : array-like of shape (n_samples, n_components)
            Projection of X in the first principal components, where `n_samples`
            is the number of samples and `n_components` is the number of the components.
        )csrcscF)accept_sparsedtypereset)r   r   _validate_datafloat64float32mean_r   r   r   r   r   r   r   )r   r5   r    r!   X_transformeds        r%   	transformz_BasePCA.transform{   s    $ a A^BJJ

3KSX   
 ::!{+Atzz:

ND,,...;;RWWT%=%=>>Mr'   c                    t        |      \  }}| j                  rO|j                  | j                  ddt        j
                  f         | j                  z  }||z  | j                  z   S || j                  z  | j                  z   S )a  Transform data back to its original space.

        In other words, return an input `X_original` whose transform would be X.

        Parameters
        ----------
        X : array-like of shape (n_samples, n_components)
            New data, where `n_samples` is the number of samples
            and `n_components` is the number of components.

        Returns
        -------
        X_original array-like of shape (n_samples, n_features)
            Original data, where `n_samples` is the number of samples
            and `n_features` is the number of features.

        Notes
        -----
        If whitening is enabled, inverse_transform will compute the
        exact inverse operation, which includes reversing whitening.
        N)r   r   r   r   r   r   r   rA   )r   r5   r    r!   scaled_componentss        r%   inverse_transformz_BasePCA.inverse_transform   sy    , a A;;00BJJ?@4CSCSS  ((4::55t'''$**44r'   c                 4    | j                   j                  d   S )z&Number of transformed output features.r   )r   r*   )r   s    r%   _n_features_outz_BasePCA._n_features_out   s     %%a((r'   )N)__name__
__module____qualname____doc__r&   r2   r   r7   rC   rF   propertyrH   r4   r'   r%   r   r      sF    8+Z  "!F5@ ) )r'   r   )	metaclass)rL   abcr   r   numpyr   scipyr   scipy.sparser   baser   r	   r
   utils._array_apir   r   r   utils.sparsefuncsr   utils.validationr   r   r4   r'   r%   <module>rW      s?    / (   ! S S F F 7 .j)#%5}PWj)r'   