Description Usage Arguments Value References
The kernelized version of LMNN.
| 1 2 3 4 5 | KLMNN(num_dims = NULL, learning_rate = "adaptive", eta0 = 0.3,
  initial_metric = NULL, max_iter = 100, prec = 1e-08, tol = 1e-08,
  k = 3, mu = 0.5, learn_inc = 1.01, learn_dec = 0.5,
  eta_thres = 1e-14, kernel = "linear", gamma = NULL, degree = 3,
  coef0 = 1, kernel_params = NULL, target_selection = "kernel")
 | 
| num_dims | Desired value for dimensionality reduction. Ignored if solver is 'SDP'. If NULL, all features will be kept. Integer. | 
| learning_rate | Type of learning rate update for gradient descent. Possible values are: - 'adaptive' : the learning rate will increase if the gradient step is succesful, else it will decrease. - 'constant' : the learning rate will be constant during all the gradient steps. | 
| eta0 | The initial value for learning rate. | 
| initial_metric | If array or matrix, and solver is SDP, it must be a positive semidefinite matrix with the starting metric (d x d) for gradient descent, where d is the number of features. If None, euclidean distance will be used. If a string, the following values are allowed: - 'euclidean' : the euclidean distance. - 'scale' : a diagonal matrix that normalizes each attribute according to its range will be used. If solver is 'SGD', then the array or matrix will represent a linear map (d' x d), where d' is the dimension provided in num_dims. | 
| max_iter | Maximum number of iterations of gradient descent. Integer. | 
| prec | Precision stop criterion (gradient norm). Float. | 
| tol | Tolerance stop criterion (difference between two iterations). Float. | 
| k | Number of target neighbors to take. If this algorithm is used for nearest neighbors classification, a good choice is to take k as the number of neighbors. Integer. | 
| mu | The weight of the push error in the minimization algorithm. The objective function is composed of a push error, given by the impostors, with weight mu, and a pull error, given by the target neighbors, with weight (1-mu). It must be between 0.0 and 1.0. | 
| learn_inc | Increase factor for learning rate. Ignored if learning_rate is not 'adaptive'. Float. | 
| learn_dec | Decrease factor for learning rate. Ignored if learning_rate is not 'adaptive'. Float. | 
| eta_thres | A learning rate threshold stop criterion. Float. | 
| kernel | Kernel to use. Allowed values are: "linear" | "poly" | "rbf" | "sigmoid" | "cosine" | "precomputed". | 
| gamma | Kernel coefficient for rbf, poly and sigmoid kernels. Ignored by other kernels. Default value is 1/n_features. Float. | 
| degree | Degree for poly kernels. Ignored by other kernels. Integer. | 
| coef0 | Independent term for poly and sigmoid kernels. Ignored by other kernels. Float. | 
| kernel_params | Parameters (keyword arguments) and values for kernel passed as callable object. Ignored by other kernels. | 
| target_selection | How to find the target neighbors. Allowed values are: - 'kernel' : using the euclidean distance in the kernel space. - 'original' : using the euclidean distance in the original space. | 
The KLMNN transformer, structured as a named list.
Kilian Q Weinberger and Lawrence K Saul. “Distance metric learning for large margin nearest neighbor classification”. In: Journal of Machine Learning Research 10.Feb (2009), pages 207-244.
Lorenzo Torresani and Kuang-chih Lee. “Large margin component analysis”. In: Advances in neural information processing systems. 2007, pages 1385-1392.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.