Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 110 packages in 0.01 seconds

Ruido — by Arthur Igor da Fonseca-Freire, 9 days ago

Soundscape Background Noise, Power, and Saturation

Accessible and flexible implementation of three ecoacoustic indices that are less commonly available in existing R frameworks: Background Noise, Soundscape Power and Soundscape Saturation. The functions were design to accommodate a variety of sampling designs. Users can tailor calculations by specifying spectrogram time bin size, amplitude thresholds and normality tests. By simplifying computation and standardizing reproducible methods, the package aims to support ecoacoustics studies. For more details about the indices read Towsey (2017) < https://eprints.qut.edu.au/110634/> and Burivalova (2017) .

GLDEX — by Steve Su, 7 months ago

Fitting Single and Mixture of Generalised Lambda Distributions

The fitting algorithms considered in this package have two major objectives. One is to provide a smoothing device to fit distributions to data using the weight and unweighted discretised approach based on the bin width of the histogram. The other is to provide a definitive fit to the data set using the maximum likelihood and quantile matching estimation. Other methods such as moment matching, starship method, L moment matching are also provided. Diagnostics on goodness of fit can be done via qqplots, KS-resample tests and comparing mean, variance, skewness and kurtosis of the data with the fitted distribution. References include the following: Karvanen and Nuutinen (2008) "Characterizing the generalized lambda distribution by L-moments" , King and MacGillivray (1999) "A starship method for fitting the generalised lambda distributions" , Su (2005) "A Discretized Approach to Flexibly Fit Generalized Lambda Distributions to Data" , Su (2007) "Nmerical Maximum Log Likelihood Estimation for Generalized Lambda Distributions" , Su (2007) "Fitting Single and Mixture of Generalized Lambda Distributions to Data via Discretized and Maximum Likelihood Methods: GLDEX in R" , Su (2009) "Confidence Intervals for Quantiles Using Generalized Lambda Distributions" , Su (2010) "Chapter 14: Fitting GLDs and Mixture of GLDs to Data using Quantile Matching Method" , Su (2010) "Chapter 15: Fitting GLD to data using GLDEX 1.0.4 in R" , Su (2015) "Flexible Parametric Quantile Regression Model" , Su (2021) "Flexible parametric accelerated failure time model".

SlideCNA — by Diane Zhang, a year ago

Calls Copy Number Alterations from Slide-Seq Data

This takes spatial single-cell-type RNA-seq data (specifically designed for Slide-seq v2) that calls copy number alterations (CNAs) using pseudo-spatial binning, clusters cellular units (e.g. beads) based on CNA profile, and visualizes spatial CNA patterns. Documentation about 'SlideCNA' is included in the the pre-print by Zhang et al. (2022, ). The package 'enrichR' (>= 3.0), conditionally used to annotate SlideCNA-determined clusters with gene ontology terms, can be installed at < https://github.com/wjawaid/enrichR> or with install_github("wjawaid/enrichR").

tempted — by Pixu Shi, 2 years ago

Temporal Tensor Decomposition, a Dimensionality Reduction Tool for Longitudinal Multivariate Data

TEMPoral TEnsor Decomposition (TEMPTED), is a dimension reduction method for multivariate longitudinal data with varying temporal sampling. It formats the data into a temporal tensor and decomposes it into a summation of low-dimensional components, each consisting of a subject loading vector, a feature loading vector, and a continuous temporal loading function. These loadings provide a low-dimensional representation of subjects or samples and can be used to identify features associated with clusters of subjects or samples. TEMPTED provides the flexibility of allowing subjects to have different temporal sampling, so time points do not need to be binned, and missing time points do not need to be imputed.

forceplate — by Raphael Hartmann, 2 months ago

Processing Force-Plate Data

Process raw force-plate data (txt-files) by segmenting them into trials and, if needed, calculating (user-defined) descriptive statistics of variables for user-defined time bins (relative to trigger onsets) for each trial. When segmenting the data a baseline correction, a filter, and a data imputation can be applied if needed. Experimental data can also be processed and combined with the segmented force-plate data. This procedure is suggested by Johannsen et al. (2023) and some of the options (e.g., choice of low-pass filter) are also suggested by Winter (2009) .

SpatialKWD — by Stefano Gualandi, 3 years ago

Spatial KWD for Large Spatial Maps

Contains efficient implementations of Discrete Optimal Transport algorithms for the computation of Kantorovich-Wasserstein distances between pairs of large spatial maps (Bassetti, Gualandi, Veneroni (2020), ). All the algorithms are based on an ad-hoc implementation of the Network Simplex algorithm. The package has four main helper functions: compareOneToOne() (to compare two spatial maps), compareOneToMany() (to compare a reference map with a list of other maps), compareAll() (to compute a matrix of distances between a list of maps), and focusArea() (to compute the KWD distance within a focus area). In non-convex maps, the helper functions first build the convex-hull of the input bins and pad the weights with zeros.

ggsurveillance — by Alexander Bartel, 3 months ago

Tools for Outbreak Investigation/Infectious Disease Surveillance

Create epicurves, epigantt charts, and diverging bar charts using 'ggplot2'. Prepare data for visualisation or other reporting for infectious disease surveillance and outbreak investigation (time series data). Includes tidy functions to solve date based transformations for common reporting tasks, like (A) seasonal date alignment for respiratory disease surveillance, (B) date-based case binning based on specified time intervals like isoweek, epiweek, month and more, (C) automated detection and marking of the new year based on the date/datetime axis of the 'ggplot2', (D) labelling of the last value of a time-series. An introduction on how to use epicurves can be found on the US CDC website (2012, < https://www.cdc.gov/training/quicklearns/epimode/index.html>).

degross — by Philippe Lambert, 5 years ago

Density Estimation from GROuped Summary Statistics

Estimation of a density from grouped (tabulated) summary statistics evaluated in each of the big bins (or classes) partitioning the support of the variable. These statistics include class frequencies and central moments of order one up to four. The log-density is modelled using a linear combination of penalised B-splines. The multinomial log-likelihood involving the frequencies adds up to a roughness penalty based on the differences in the coefficients of neighbouring B-splines and the log of a root-n approximation of the sampling density of the observed vector of central moments in each class. The so-obtained penalized log-likelihood is maximized using the EM algorithm to get an estimate of the spline parameters and, consequently, of the variable density and related quantities such as quantiles, see Lambert, P. (2021) for details.

detect — by Peter Solymos, a month ago

Analyzing Wildlife Data with Detection Error

Models for analyzing site occupancy and count data models with detection error, including single-visit based models (Lele et al. 2012 , Moreno et al. 2010 , Solymos et al. 2012 , Denes et al. 2016 ), conditional distance sampling and time-removal models (QPAD) (Solymos et al. 2013 , Solymos et al. 2018 ), and single bin QPAD (SQPAD) models (Lele & Solymos 2025 ). Package development was supported by the Alberta Biodiversity Monitoring Institute and the Boreal Avian Modelling Project.

APCtools — by Alexander Bauer, 8 months ago

Routines for Descriptive and Model-Based APC Analysis

Age-Period-Cohort (APC) analyses are used to differentiate relevant drivers for long-term developments. The 'APCtools' package offers visualization techniques and general routines to simplify the workflow of an APC analysis. Sophisticated functions are available both for descriptive and regression model-based analyses. For the former, we use density (or ridgeline) matrices and (hexagonally binned) heatmaps as innovative visualization techniques building on the concept of Lexis diagrams. Model-based analyses build on the separation of the temporal dimensions based on generalized additive models, where a tensor product interaction surface (usually between age and period) is utilized to represent the third dimension (usually cohort) on its diagonal. Such tensor product surfaces can also be estimated while accounting for further covariates in the regression model. See Weigert et al. (2021) for methodological details.