diff --git a/autoqild/dataset_readers/synthetic_data_generator.py b/autoqild/dataset_readers/synthetic_data_generator.py index 65691f7..334f48c 100644 --- a/autoqild/dataset_readers/synthetic_data_generator.py +++ b/autoqild/dataset_readers/synthetic_data_generator.py @@ -486,7 +486,7 @@ def bayes_predictor_pc_softmax_mi(self): - \( z_k \) is the logit or raw score for class \( k \). - \( K \) is the total number of classes. - + PC-Softmax Function: .. math:: diff --git a/autoqild/mi_estimators/pc_softmax_estimator.py b/autoqild/mi_estimators/pc_softmax_estimator.py index e554876..692bde5 100644 --- a/autoqild/mi_estimators/pc_softmax_estimator.py +++ b/autoqild/mi_estimators/pc_softmax_estimator.py @@ -42,13 +42,13 @@ class PCSoftmaxMIEstimator(MIEstimatorBase): Optimizer type to use for training the neural network. Must be one of: - - `RMSprop`: RMSprop optimizer. - - `sgd`: Stochastic Gradient Descent optimizer. - - `adam`: Adam optimizer. - - `AdamW`: AdamW optimizer. - - `Adagrad`: Adagrad optimizer. - - `Adamax`: Adamax optimizer. - - `Adadelta`: Adadelta optimizer. + - `RMSprop`: Root Mean Square Propagation, an adaptive learning rate method. + - `sgd`: Stochastic Gradient Descent, a simple and widely-used optimizer. + - "adam": Adaptive Moment Estimation, combining momentum and RMSProp for better convergence. + - `AdamW`: Adam with weight decay, an improved variant of Adam with better regularization. + - `Adagrad`: Adaptive Gradient Algorithm, adjusting the learning rate based on feature frequency. + - `Adamax`: Variant of Adam based on infinity norm, more robust with sparse gradients. + - `Adadelta`: An extension of Adagrad that seeks to reduce its aggressive learning rate decay. learning_rate : float, optional, default=0.001 Learning rate for the optimizer.