Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 1420 packages in 0.02 seconds

subplex — by Aaron A. King, a year ago

Unconstrained Optimization using the Subplex Algorithm

The subplex algorithm for unconstrained optimization, developed by Tom Rowan.

maxLik — by Ott Toomet, 2 years ago

Maximum Likelihood Estimation and Related Tools

Functions for Maximum Likelihood (ML) estimation, non-linear optimization, and related tools. It includes a unified way to call different optimizers, and classes and methods to handle the results from the Maximum Likelihood viewpoint. It also includes a number of convenience tools for testing and developing your own models.

Rmosek — by Henrik A. Friberg, 6 years ago

The R to MOSEK Optimization Interface

This is a meta-package designed to support the installation of Rmosek (>= 6.0) and bring the optimization facilities of MOSEK (>= 6.0) to the R-language. The interface supports large-scale optimization of many kinds: Mixed-integer and continuous linear, second-order cone, exponential cone and power cone optimization, as well as continuous semidefinite optimization. Rmosek and the R-language are open-source projects. MOSEK is a proprietary product, but unrestricted trial and academic licenses are available.

janitor — by Sam Firke, a year ago

Simple Tools for Examining and Cleaning Dirty Data

The main janitor functions can: perfectly format data.frame column names; provide quick counts of variable combinations (i.e., frequency tables and crosstabs); and explore duplicate records. Other janitor functions nicely format the tabulation results. These tabulate-and-report functions approximate popular features of SPSS and Microsoft Excel. This package follows the principles of the "tidyverse" and works well with the pipe function %>%. janitor was built with beginning-to-intermediate R users in mind and is optimized for user-friendliness.

emoa — by Olaf Mersmann, a year ago

Evolutionary Multiobjective Optimization Algorithms

Collection of building blocks for the design and analysis of evolutionary multiobjective optimization algorithms.

smoof — by Jakob Bossek, 3 years ago

Single and Multi-Objective Optimization Test Functions

Provides generators for a high number of both single- and multi- objective test functions which are frequently used for the benchmarking of (numerical) optimization algorithms. Moreover, it offers a set of convenient functions to generate, plot and work with objective functions.

transport — by Dominic Schuhmacher, a year ago

Computation of Optimal Transport Plans and Wasserstein Distances

Solve optimal transport problems. Compute Wasserstein distances (a.k.a. Kantorovitch, Fortet--Mourier, Mallows, Earth Mover's, or minimal L_p distances), return the corresponding transference plans, and display them graphically. Objects that can be compared include grey-scale images, (weighted) point patterns, and mass vectors.

ParamHelpers — by Martin Binder, a year ago

Helpers for Parameters in Black-Box Optimization, Tuning and Machine Learning

Functions for parameter descriptions and operations in black-box optimization, tuning and machine learning. Parameters can be described (type, constraints, defaults, etc.), combined to parameter sets and can in general be programmed on. A useful OptPath object (archive) to log function evaluations is also provided.

Matching — by Jasjeet Singh Sekhon, a year ago

Multivariate and Propensity Score Matching with Balance Optimization

Provides functions for multivariate and propensity score matching and for finding optimal balance based on a genetic search algorithm. A variety of univariate and multivariate metrics to determine if balance has been obtained are also provided. For details, see the paper by Jasjeet Sekhon (2007, ).

lbfgs — by Antonio Coppola, 4 years ago

Limited-memory BFGS Optimization

A wrapper built around the libLBFGS optimization library by Naoaki Okazaki. The lbfgs package implements both the Limited-memory Broyden-Fletcher-Goldfarb-Shanno (L-BFGS) and the Orthant-Wise Quasi-Newton Limited-Memory (OWL-QN) optimization algorithms. The L-BFGS algorithm solves the problem of minimizing an objective, given its gradient, by iteratively computing approximations of the inverse Hessian matrix. The OWL-QN algorithm finds the optimum of an objective plus the L1-norm of the problem's parameters. The package offers a fast and memory-efficient implementation of these optimization routines, which is particularly suited for high-dimensional problems.