Tools for Spell Checking in R
Spell checking common document formats including latex, markdown, manual pages, and description files. Includes utilities to automate checking of documentation and vignettes as a unit test during R CMD check. Both British and American English are supported out of the box and other languages can be added. In addition, packages may define a wordlist to allow custom terminology without having to abuse punctuation.
Scientific use casesAutomated Cleaning of Occurrence Records from Biological Collections
Automated flagging of common spatial and temporal errors in biological and paleontological collection data, for the use in conservation, ecology and paleontology. Includes automated tests to easily flag (and exclude) records assigned to country or province centroid, the open ocean, the headquarters of the Global Biodiversity Information Facility, urban areas or the location of biodiversity institutions (museums, zoos, botanical gardens, universities). Furthermore identifies per species outlier coordinates, zero coordinates, identical latitude/longitude and invalid coordinates. Also implements an algorithm to identify data sets with a significant proportion of rounded coordinates. Especially suited for large data sets. The reference for the methodology is: Zizka et al. (2019) doi:10.1111/2041-210X.13152.
Scientific use casesClean Biological Occurrence Records
Clean biological occurrence records. Includes functionality for cleaning based on various aspects of spatial coordinates, unlikely values due to political centroids, coordinates based on where collections of specimens are held, and more.
Scientific use casesManipulation of Matched Phylogenies and Data using data.table
An implementation that combines trait data and a phylogenetic tree (or trees) into a single object of class treedata.table. The resulting object can be easily manipulated to simultaneously change the trait- and tree-level sampling. Currently implemented functions allow users to use a data.table syntax when performing operations on the trait dataset within the treedata.table object.
View DocumentationRead Spectrometric Data and Metadata
Parse various reflectance/transmittance/absorbance spectra file formats to extract spectral data and metadata, as described in Gruson, White & Maia (2019) doi:10.21105/joss.01857. Among other formats, it can import files from Avantes https://www.avantes.com/, CRAIC http://www.microspectra.com/, and OceanInsight (formerly OceanOptics) https://www.oceaninsight.com/ brands.
View DocumentationEcological Metadata as Linked Data
This is a utility for transforming Ecological Metadata Language (EML) files into JSON-LD and back into EML. Doing so creates a list-based representation of EML in R, so that EML data can easily be manipulated using standard R tools. This makes this package an effective backend for other R-based tools working with EML. By abstracting away the complexity of XML Schema, developers can build around native R list objects and not have to worry about satisfying many of the additional constraints of set by the schema (such as element ordering, which is handled automatically). Additionally, the JSON-LD representation enables the use of developer-friendly JSON parsing and serialization that may facilitate the use of EML in contexts outside of R, as well as the informatics-friendly serializations such as RDF and SPARQL queries.
View DocumentationBase Classes and Functions for Phylogenetic Tree Input and Output
treeio is an R package to make it easier to import and store phylogenetic tree with associated data; and to link external data from different sources to phylogeny. It also supports exporting phylogenetic tree with heterogeneous associated data to a single tree file and can be served as a platform for merging tree with associated data and converting file formats.
Scientific use casesExport Data Frames to Excel xlsx Format
Zero-dependency data frame to xlsx exporter based on libxlsxwriter. Fast and no Java or Excel required.
Scientific use casesStore and Retrieve Data.frames in a Git Repository
Make versioning of data.frame easy and efficient using git repositories.
View DocumentationParse Darwin Core Files
Parse and create Darwin Core (http://rs.tdwg.org/dwc/) Simple and Archives. Functionality includes reading and parsing all the files in a Darwin Core Archive, including the datasets and metadata; read and parse simple Darwin Core files; and validation of Darwin Core Archives.
Scientific use casesLightweight Qualitative Coding
A free, lightweight, open source option for analyzing text-based qualitative data. Enables analysis of interview transcripts, observation notes, memos, and other sources. Supports the work of social scientists, historians, humanists, and other researchers who use qualitative methods. Addresses the unique challenges faced in analyzing qualitative data analysis. Provides opportunities for researchers who otherwise might not develop software to build software development skills.
View DocumentationPolyhedra Database
A polyhedra database scraped from various sources as R6 objects and rgl visualizing capabilities.
View DocumentationParse a BibTeX File to a Data Frame
Parse a BibTeX file to a data.frame to make it accessible for further analysis and visualization.
Scientific use casesJSON for Linking Data
JSON-LD is a light-weight syntax for expressing linked data. It is primarily intended for web-based programming environments, interoperable web services and for storing linked data in JSON-based databases. This package provides bindings to the JavaScript library for converting, expanding and compacting JSON-LD documents.
View DocumentationExtensible Style-Sheet Language Transformations
An extension for the xml2 package to transform XML documents by applying an xslt style-sheet.
View DocumentationClient for jq, a JSON Processor
Client for jq, a JSON processor (https://stedolan.github.io/jq/), written in C. jq allows the following with JSON data: index into, parse, do calculations, cut up and filter, change key names and values, perform conditionals and comparisons, and more.
View DocumentationPositron Emission Tomography Time-Activity Curve Analysis
To facilitate the analysis of positron emission tomography (PET) time activity curve (TAC) data, and to encourage open science and replicability, this package supports data loading and analysis of multiple TAC file formats. Functions are available to analyze loaded TAC data for individual participants or in batches. Major functionality includes weighted TAC merging by region of interest (ROI), calculating models including standardized uptake value ratio (SUVR) and distribution volume ratio (DVR, Logan et al. 1996 doi:10.1097/00004647-199609000-00008), basic plotting functions and calculation of cut-off values (Aizenstein et al. 2008 doi:10.1001/archneur.65.11.1509). Please see the walkthrough vignette for a detailed overview of tacmagic functions.
Scientific use casesTools to Manipulate and Query Semantic Data
The Resource Description Framework, or RDF is a widely used data representation model that forms the cornerstone of the Semantic Web. RDF represents data as a graph rather than the familiar data table or rectangle of relational databases. The rdflib package provides a friendly and concise user interface for performing common tasks on RDF data, such as reading, writing and converting between the various serializations of RDF data, including rdfxml, turtle, nquads, ntriples, and json-ld; creating new RDF graphs, and performing graph queries using SPARQL. This package wraps the low level redland R package which provides direct bindings to the redland C library. Additionally, the package supports the newer and more developer friendly JSON-LD format through the jsonld package. The package interface takes inspiration from the Python rdflib library.
Scientific use casesCreate and Query a Local Copy of GenBank in R
Download large sections of GenBank https://www.ncbi.nlm.nih.gov/genbank/ and generate a local SQL-based database. A user can then query this database using restez functions or through rentrez https://CRAN.R-project.org/package=rentrez wrappers.
Scientific use casesConduct Co-Localization Analysis of Fluorescence Microscopy Images
Automate the co-localization analysis of fluorescence microscopy images. Selecting regions of interest, extract pixel intensities from the image channels and calculate different co-localization statistics. The methods implemented in this package are based on Dunn et al. (2011) doi:10.1152/ajpcell.00462.2010.
Scientific use casesInterface to Phylocom
Interface to Phylocom (http://phylodiversity.net/phylocom/), a library for analysis of phylogenetic community structure and character evolution. Includes low level methods for interacting with the three executables, as well as higher level interfaces for methods like aot, ecovolve, bladj, phylomatic, and more.
View DocumentationDendrograms for Evolutionary Analysis
Contains functions for developing phylogenetic trees as deeply-nested lists (“dendrogram” objects). Enables bi-directional conversion between dendrogram and “phylo” objects (see Paradis et al (2004) doi:10.1093/bioinformatics/btg412), and features several tools for command-line tree manipulation and import/export via Newick parenthetic text.
Scientific use casesThe Critical Care Clinical Data Processing Tools
An electronic health care record (EHR) data cleaning and processing platform. It focus on heterogeneous high resolution longitudinal data. It works with Critical Care Health Informatics Collaborative (CCHIC) dataset. It is created to address various data reliability and accessibility problems of EHRs as such.
View DocumentationSupports the Analysis of RTI MicroPEM Output Files
Supports the input and reproducible analysis of RTI MicroPEM output files.
Scientific use casesGenerate Starting Trees For Combined Molecular, Morphological and Stratigraphic Data
Combine a list of taxa with a phylogeny to generate a starting tree for use in total evidence dating analyses.
View DocumentationSimple Text Wrappers
Simple functions for common repeatable tasks in NLP and text mining.
View DocumentationAnalysis of Work Loops and Other Data from Muscle Physiology Experiments
Functions for the import, transformation, and analysis of data from muscle physiology experiments. The work loop technique is used to evaluate the mechanical work and power output of muscle. Josephson (1985) https://jeb.biologists.org/content/114/1/493 modernized the technique for application in comparative biomechanics. Although our initial motivation was to provide functions to analyze work loop experiment data, as we developed the package we incorporated the ability to analyze data from experiments that are often complementary to work loops. There are currently three supported experiment types: work loops, simple twitches, and tetanus trials. Data can be imported directly from .ddf files or via an object constructor function. Through either method, data can then be cleaned or transformed via methods typically used in studies of muscle physiology. Data can then be analyzed to determine the timing and magnitude of force development and relaxation (for isometric trials) or the magnitude of work, net power, and instantaneous power among other things (for work loops). Although we do not provide plotting functions, all resultant objects are designed to be friendly to visualization via either base-R plotting or tidyverse functions. This package has been peer-reviewed by rOpenSci (v. 1.1.0).
View DocumentationRecodes Sex/Gender Descriptions Into A Standard Set
gendercodeR allows for simple recoding of freetext gender responses.
View DocumentationTidy up nested list hairballs
This is a package to transform large, multi-nested lists into a more
user-friendly format. The initial focus is on making
processing of return values from jsonlite::fromJSON()
queries more seamless,
but ideally this package should be useful for deeply-nested lists from an array
of sources.
Convert Between Units
Provides conversion functionality between a broad range of scientific, historical, and industrial unit types.
View Documentation