Found 288 packages in 0.01 seconds
Export Data Frames to Excel 'xlsx' Format
Zero-dependency data frame to xlsx exporter based on 'libxlsxwriter' < https://libxlsxwriter.github.io>. Fast and no Java or Excel required.
World Vector Map Data from Natural Earth Used in 'rnaturalearth'
Vector map data from < https://www.naturalearthdata.com/>. Access functions are provided in the accompanying package 'rnaturalearth'.
Read and Write ODS Files
Read ODS (OpenDocument Spreadsheet) into R as data frame. Also support writing data frame into ODS file.
Split, Combine and Compress PDF Files
Content-preserving transformations transformations of PDF files such as split, combine, and compress. This package interfaces directly to the 'qpdf' C++ library < https://qpdf.sourceforge.io/> and does not require any command line utilities. Note that 'qpdf' does not read actual content from PDF files: to extract text and data you need the 'pdftools' package.
Convert Data from and to 'GeoJSON' or 'TopoJSON'
Convert data to 'GeoJSON' or 'TopoJSON' from various R classes, including vectors, lists, data frames, shape files, and spatial classes. 'geojsonio' does not aim to replace packages like 'sp', 'rgdal', 'rgeos', but rather aims to be a high level client to simplify conversions of data from and to 'GeoJSON' and 'TopoJSON'.
'Entrez' in R
Provides an R interface to the NCBI's 'EUtils' API, allowing users to search databases like 'GenBank' < https://www.ncbi.nlm.nih.gov/genbank/> and 'PubMed' < https://pubmed.ncbi.nlm.nih.gov/>, process the results of those searches and pull data into their R sessions.
Text Extraction, Rendering and Converting of PDF Documents
Utilities based on 'libpoppler' < https://poppler.freedesktop.org> for extracting text, fonts, attachments and metadata from a PDF file. Also supports high quality rendering of PDF documents into PNG, JPEG, TIFF format, or into raw bitmap vectors for further processing in R.
Fast, Consistent Tokenization of Natural Language Text
Convert natural language text into tokens. Includes tokenizers for shingled n-grams, skip n-grams, words, word stems, sentences, paragraphs, characters, shingled characters, lines, Penn Treebank, regular expressions, as well as functions for counting characters, words, and sentences, and a function for splitting longer texts into separate documents, each with the same number of words. The tokenizers have a consistent interface, and the package is built on the 'stringi' and 'Rcpp' packages for fast yet correct tokenization in 'UTF-8'.
Validate 'JSON' Schema
Uses the node library 'is-my-json-valid' or 'ajv' to validate 'JSON' against a 'JSON' schema. Drafts 04, 06 and 07 of 'JSON' schema are supported.
Bibtex Parser
Utility to parse a bibtex file.