Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
245 changes: 245 additions & 0 deletions doc/sphinx-guides/source/admin/features.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,245 @@
# Features

An overview of Dataverse features can be found at <https://dataverse.org/software-features>. This is a more comprehensive list.

```{contents} Contents:
:local:
:depth: 3
```


## Access and download

### Login via Shibboleth

Single Sign On (SSO) using your institution's credentials.
{doc}`More information.</installation/shibboleth>`

### Login via ORCID, Google, GitHub, or Microsoft

Log in using popular OAuth2 providers.
{doc}`More information.</installation/oauth2>`

### Login via OpenID Connect (OIDC)

Log in using your institution's identity provider or a third party.
{doc}`More information.</installation/oidc>`

### Versioning

History of changes to datasets and files are preserved.
{doc}`More information.</user/dataset-management>`

### File previews

A preview is available for text, tabular, image, audio, video, and geospatial files.
{ref}`More information.<file-previews>`

### Preview and analysis of tabular files

Data Explorer allows for searching, charting and cross tabulation analysis
{ref}`More information.<inventory-of-external-tools>`

### Guestbook

Optionally collect data about who is downloading the files from your datasets.
{ref}`More information.<dataset-guestbooks>`

### File download in R and TSV format

Proprietary tabular formats are converted into RData and TSV.
{doc}`More information.</user/tabulardataingest/index>`

### Faceted search

Facets are data driven and customizable per collection.
{doc}`More information.</user/find-use-data>`

## Administration

### Usage statistics and metrics

Download counters, support for Make Data Count.
{doc}`More information.</admin/make-data-count>`

### Private URL

Create a URL for reviewers to view an unpublished (and optionally anonymized) dataset.
{ref}`More information.<previewUrl>`

### Notifications

In app and email notifications for access requests, requests for review, etc.
{ref}`More information.<account-notifications>`

### User management

Dashboard for common user-related tasks.
{doc}`More information.</admin/dashboard>`

### Curation status labels

Let curators mark datasets with a status label customized to your needs.
{ref}`More information.<:AllowedCurationLabels>`

## Customization

### Internationalization

The Dataverse software has been translated into multiple languages.
{ref}`More information.<i18n>`

### Customization of collections

Each personal or organizational collection can be customized and branded.
{ref}`More information.<theme>`

### Widgets

Embed listings of data in external websites.
{ref}`More information.<dataverse-widgets>`

### Branding

Your installation can be branded with a custom homepage, header, footer, CSS, etc.
{ref}`More information.<Branding Your Installation>`

## FAIR data publication

### Support for FAIR Data Principles

Findable, Accessible, Interoperable, Reusable.
[More information.](https://scholar.harvard.edu/mercecrosas/presentations/fair-guiding-principles-implementation-dataverse)
### Publishing workflow support

Datasets start as drafts and can be submitted for review before publication.
{ref}`More information.<dataverse-permissions>`

## File management

### Restricted files

Control who can download files and choose whether or not to enable a "Request Access" button.
{ref}`More information.<restricted-files>`

### Embargo

Make content inaccessible until an embargo end date.
{ref}`More information.<embargoes>`

### File hierarchy

Users are able to control dataset file hierarchy and directory structure.
{doc}`More information.</user/dataset-management>`

### Fixity checks for files

MD5, SHA-1, SHA-256, SHA-512, UNF.
{ref}`More information.<:FileFixityChecksumAlgorithm>`

### Backend storage on S3 or Swift

Choose between filesystem or object storage, configurable per collection and per dataset.
{doc}`More information.</developers/big-data-support>`

### Direct upload and download for S3

After a permission check, files can pass freely and directly between a client computer and S3.
{doc}`More information.</developers/big-data-support>`

### Pull header metadata from Astronomy (FITS) files

Dataset metadata prepopulated from FITS file metadata.
{ref}`More information.<fits>`

### Auxiliary files for data files

Each data file can have any number of auxiliary files for documentation or other purposes (experimental).
{doc}`More information.</developers/aux-file-support>`

## Integrations

### DataCite integration

DOIs are reserved, and when datasets are published, their metadata is published to DataCite.
{doc}`More information.</admin/discoverability>`

### External tools

Enable additional features not built in to the Dataverse software.
{doc}`More information.</installation/external-tools>`

### Dropbox integration

Upload files stored on Dropbox.
{doc}`More information.</admin/integrations>`

### GitHub integration

A GitHub Action is available to upload files from GitHub to a dataset.
{doc}`More information.</admin/integrations>`

### Integration with Jupyter notebooks

Datasets can be opened in Binder to run code in Jupyter notebooks, RStudio, and other computation environments.
{ref}`More information.<binder>`

## Interoperability

### OAI-PMH (Harvesting)

Gather and expose metadata from and to other systems using standardized metadata formats: Dublin Core, Data Document Initiative (DDI), OpenAIRE, etc.
{doc}`More information.</admin/harvestclients>`

### APIs for interoperability and custom integrations

Search API, Data Deposit (SWORD) API, Data Access API, Metrics API, Migration API, etc.
{doc}`More information.</api/index>`

### API client libraries

Interact with Dataverse APIs from Python, R, Javascript, Java, and Ruby
{doc}`More information.</api/client-libraries>`

### Schema.org JSON-LD

Used by Google Dataset Search and other services for discoverability.
{ref}`More information.<metadata-export-formats>`

### External vocabulary

Let users pick from external vocabularies (provided via API/SKOSMOS) when filling in metadata.
{ref}`More information.<using-external-vocabulary-services>`

### Export data in BagIt format

For preservation, bags can be sent to the local filesystem, Duraclound, and Google Cloud.
{ref}`More information.<BagIt Export>`

## Reusability

### Data citation for datasets and files

EndNote XML, RIS, or BibTeX format at the dataset or file level.
{doc}`More information.</user/find-use-data>`

### Custom licenses

CC0 by default but add as many standard licenses as you like or create your own.
{ref}`More information.<license-config>`

### Custom terms of use

Custom terms of use can be used in place of a license or disabled by an administrator.
{ref}`More information.<license-terms>`

### Post-publication automation (workflows)

Allow publication of a dataset to kick off external processes and integrations.
{doc}`More information.</developers/workflows>`

### Provenance

Upload standard W3C provenance files or enter free text instead.
{ref}`More information.<provenance>`

1 change: 1 addition & 0 deletions doc/sphinx-guides/source/admin/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@ This guide documents the functionality only available to superusers (such as "da
.. toctree::
:maxdepth: 2

features
dashboard
external-tools
discoverability
Expand Down
64 changes: 64 additions & 0 deletions scripts/issues/11998/tsv2md.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,64 @@
#!/usr/bin/env python
#
# Download features.tsv like this:
# curl -L "https://docs.google.com/spreadsheets/d/1EIFGAfDfZAboFa3_ShRfgoT6xSDpKohDH2_iCyO5MtA/export?gid=729532473&format=tsv" > features.tsv
#
# The gid above is a specific tab in this spreadsheet:
# https://docs.google.com/spreadsheets/d/1EIFGAfDfZAboFa3_ShRfgoT6xSDpKohDH2_iCyO5MtA/edit?usp=sharing
#
# Here's the README for the spreadsheet:
# https://docs.google.com/document/d/1wqLVoEpnD93Y_wQtA2cQEkAuC0QstC6XVs9XlA7yvbM/edit?usp=sharing
import sys
from optparse import OptionParser
import csv
from itertools import groupby

parser = OptionParser()
options, args = parser.parse_args()

if args:
tsv_file = open(args[0])
else:
tsv_file = sys.stdin

print("""# Features

An overview of Dataverse features can be found at <https://dataverse.org/software-features>. This is a more comprehensive list.

```{contents} Contents:
:local:
:depth: 3
```

""")

reader = csv.DictReader(tsv_file, delimiter="\t")
rows = [row for row in reader]
missing = []
# Sort rows by category
rows.sort(key=lambda x: x["Categories"])

# Group by category
for category, group in groupby(rows, key=lambda x: x["Categories"]):
# print('BEGIN')
print("## %s" % category)
print()
for row in group:
title = row["Title"]
description = row["Description"]
url = row["URL"]
dtype = row["DocLinkType"]
target = row["DocLinkTarget"]
print("### %s" % title)
print()
print("%s" % description)
if target == 'url':
print("[More information.](%s)" % (url))
elif target != '':
print("{%s}`More information.<%s>`" % (dtype, target))
print()
else:
missing.append(url)
tsv_file.close()
for item in missing:
print(item)