Reset all


Content Types


AID systems



Data access

Data access restrictions

Database access

Database access restrictions

Database licenses

Data licenses

Data upload

Data upload restrictions

Enhanced publication

Institution responsibility type

Institution type


Metadata standards

PID systems

Provider types

Quality management

Repository languages



Repository types


  • * at the end of a keyword allows wildcard searches
  • " quotes can be used for searching phrases
  • + represents an AND search (default)
  • | represents an OR search
  • - represents a NOT operation
  • ( and ) implies priority
  • ~N after a word specifies the desired edit distance (fuzziness)
  • ~N after a phrase specifies the desired slop amount
Found 115 result(s)
The modENCODE Project, Model Organism ENCyclopedia Of DNA Elements, was initiated by the funding of applications received in response to Requests for Applications (RFAs) HG-06-006, entitled Identification of All Functional Elements in Selected Model Organism Genomes and HG-06-007, entitled A Data Coordination Center for the Model Organism ENCODE Project (modENCODE). The modENCODE Project is being run as an open consortium and welcomes any investigator willing to abide by the criteria for participation that have been established for the project. Both computational and experimental approaches are being applied by modENCODE investigators to study the genomes of D. melanogaster and C. elegans. An added benefit of studying functional elements in model organisms is the ability to biologically validate the elements discovered using methods that cannot be applied in humans. The comprehensive dataset that is expected to result from the modENCODE Project will provide important insights into the biology of D. melanogaster and C. elegans as well as other organisms, including humans.
MicrosporidiaDB belongs to the EuPathDB family of databases and is an integrated genomic and functional genomic database for the phylum Microsporidia. In its first iteration (released in early 2010), MicrosporidiaDB contains the genomes of two Encephalitozoon species (see below). MicrosporidiaDB integrates whole genome sequence and annotation and will rapidly expand to include experimental data and environmental isolate sequences provided by community researchers. The database includes supplemental bioinformatics analyses and a web interface for data-mining.
Apollo (previously DSpace@Cambridge) is the University of Cambridge’s institutional repository, preserving and providing access to content created by members of the University. The repository stores a range of content and provides different levels of access, but its primary focus is on providing open access to the University’s research publications.
EartH2Observe brings together the findings from European FP projects DEWFORA, GLOWASIS, WATCH, GEOWOW and others. It will integrate available global earth observations (EO), in-situ datasets and models and will construct a global water resources re-analysis dataset of significant length (several decades). The resulting data will allow for improved insights on the full extent of available water and existing pressures on global water resources in all parts of the water cycle. The project will support efficient and globally consistent water management and decision making by providing comprehensive multi-scale (regional, continental and global) water resources observations. It will test new EO data sources, extend existing processing algorithms and combine data from multiple satellite missions in order to improve the overall resolution and reliability of EO data included in the re-analysis dataset. The resulting datasets will be made available through an open Water Cycle Integrator data portal : the European contribution to the GEOSS/WCI approach. The datasets will be downscaled for application in case-studies at regional and local levels, and optimized based on identified European and local needs supporting water management and decision making . Actual data access:
The National River Flow Archive is the primary archive of daily and peak river flows for the United Kingdom. The archive incorporates daily, monthly and flood peak data from over 1500 gauging stations. The NRFA holds a wide range of hydrological information to assist in the understanding and interpretation of measured river flows. In addition to time series of gauged river flow, the data centre maintains hydrometric information relating to the gauging stations and the catchments they command and data quantifying other parts of the hydrological cycle.
The National Digital Archive of Datasets (NDAD) provides access to archived datasets and documents from United Kingdom government departments which can be searched or browsed by subjects such as armed forces service or wills and death duties. Statistics and information gathered through census data as well as public records are used to compile the available datasets. All datasets are available to download and contain a record summary as well as custodial history, background on the source of the data and whether or not data may be added to the dataset in the future.
DDBJ; DNA Data Bank of Japan is the sole nucleotide sequence data bank in Asia, which is officially certified to collect nucleotide sequences from researchers and to issue the internationally recognized accession number to data submitters.Since we exchange the collected data with EMBL-Bank/EBI; European Bioinformatics Institute and GenBank/NCBI; National Center for Biotechnology Information on a daily basis, the three data banks share virtually the same data at any given time. The virtually unified database is called "INSD; International Nucleotide Sequence Database DDBJ collects sequence data mainly from Japanese researchers, but of course accepts data and issue the accession number to researchers in any other countries.
mzCloud is an extensively curated database of high-resolution tandem mass spectra that are arranged into spectral trees. MS/MS and multi-stage MSn spectra were acquired at various collision energies, precursor m/z, and isolation widths using Collision-induced dissociation (CID) and Higher-energy collisional dissociation (HCD). Each raw mass spectrum was filtered and recalibrated giving rise to additional filtered and recalibrated spectral trees that are fully searchable. Besides the experimental and processed data, each database record contains the compound name with synonyms, the chemical structure, computationally and manually annotated fragments (peaks), identified adducts and multiply charged ions, molecular formulas, predicted precursor structures, detailed experimental information, peak accuracies, mass resolution, InChi, InChiKey, and other identifiers. mzCloud is a fully searchable library that allows spectra searches, tree searches, structure and substructure searches, monoisotopic mass searches, peak (m/z) searches, precursor searches, and name searches. mzCloud is free and available for public use online.
Codex Sinaiticus is one of the most important books in the world. Handwritten well over 1600 years ago, the manuscript contains the Christian Bible in Greek, including the oldest complete copy of the New Testament. The Codex Sinaiticus Project is an international collaboration to reunite the entire manuscript in digital form and make it accessible to a global audience for the first time. Drawing on the expertise of leading scholars, conservators and curators, the Project gives everyone the opportunity to connect directly with this famous manuscript.
The European Nucleotide Archive (ENA) captures and presents information relating to experimental workflows that are based around nucleotide sequencing. A typical workflow includes the isolation and preparation of material for sequencing, a run of a sequencing machine in which sequencing data are produced and a subsequent bioinformatic analysis pipeline. ENA records this information in a data model that covers input information (sample, experimental setup, machine configuration), output machine data (sequence traces, reads and quality scores) and interpreted information (assembly, mapping, functional annotation). Data arrive at ENA from a variety of sources. These include submissions of raw data, assembled sequences and annotation from small-scale sequencing efforts, data provision from the major European sequencing centres and routine and comprehensive exchange with our partners in the International Nucleotide Sequence Database Collaboration (INSDC). Provision of nucleotide sequence data to ENA or its INSDC partners has become a central and mandatory step in the dissemination of research findings to the scientific community. ENA works with publishers of scientific literature and funding bodies to ensure compliance with these principles and to provide optimal submission systems and data access tools that work seamlessly with the published literature.
>>>!!!<<< Crystaleye has now been excitingly integrated into the Crystallography Open Database at>>>!!!<<< Crystallography Open Database now is including data and software from CrystalEye, developed by Nick Day at the department of Chemistry, the University of Cambridge under supervision of Peter Murray-Rust. The aim of the CrystalEye project is to aggregate crystallography from web resources, and to provide methods to easily browse, search, and to keep up to date with the latest published information.At present we are aggregating the crystallography from the supplementary data to articles at publishers websites.
Content type(s)
>>>!!!<<< 2019-05-15: the repository is offline >>>!!!<<< MIRAGE is developing a warehouse of medical images to facilitate effective online retrieval tools in the institutional web site to complement the existing online e-leaning and teaching system OASISplus, also known as Blackboard Vista , that is currently in operation at Middlesex University (MU); Follow-up project MIRAGE 2011:
Content type(s)
UK RED is a database documenting the history of reading in Britain from 1450 to 1945. Reading experiences of British subjects, both at home and abroad presented in UK RED are drawn from published and unpublished sources as diverse as diaries, commonplace books, memoirs, sociological surveys, and criminal court and prison records.
-----<<<<< The repository is no longer available. This record is out-dated. >>>>>----- The Clean Energy Project Database (CEPDB) is a massive reference database for organic semiconductors with a particular emphasis on photovoltaic applications. It was created to store and provide access to data from computational as well as experimental studies, on both known and virtual compounds. It is a free and open resource designed to support researchers in the field of organic electronics in their scientific pursuits. The CEPDB was established as part of the Harvard Clean Energy Project (CEP), a virtual high-throughput screening initiative to identify promising new candidates for the next generation of carbon-based solar cell materials.
!!! the repository is no longer available, archived site:!!! DataBox is a digital archive for scientific primary data for use by researchers at The University of Copenhagen. DataBox is available to researchers, departments and institutes at the University and research groups with an affiliation to the University of Copenhagen. DataBox serves as an additional backup system, which archives data in a structured form for both short and medium term preservation. It can also serve as a way of sharing data. Each researcher/group can create his/her own space in DataBox and can store and process the data, and if he/she chooses to share his/her data. Version history of files is retained by the system.
The Intermediate Filament Database will function as a continuously updated review of the intermediate filament field and it is hoped that users will contribute to the development and expansion of the database on a regular basis. Contributions may include novel variants, new patients with previously discovered sequence and allelic variants. Suggestions on ways to improve the database are also welcome.
the Data Hub is a community-run catalogue of useful sets of data on the Internet. You can collect links here to data from around the web for yourself and others to use, or search for data that others have collected. Depending on the type of data (and its conditions of use), the Data Hub may also be able to store a copy of the data or host it in a database, and provide some basic visualisation tools.
The European Social Survey (the ESS) is a biennial multi-country survey covering over 30 nations. The first round was fielded in 2002/2003, the fifth in 2010/2011. The questionnaire includes two main sections, each consisting of approximately 120 items; a 'core' module which remains relatively constant from round to round, plus two or more 'rotating' modules, repeated at intervals. The core module aims to monitor change and continuity in a wide range of social variables, including media use; social and public trust; political interest and participation; socio-political orientations; governance and efficacy; moral; political and social values; social exclusion, national, ethnic and religious allegiances; well-being; health and security; human values; demographics and socio-economics
The FAIRDOMHub is built upon the SEEK software suite, which is an open source web platform for sharing scientific research assets, processes and outcomes. FAIRDOM (Web Site) will establish a support and service network for European Systems Biology. It will serve projects in standardizing, managing and disseminating data and models in a FAIR manner: Findable, Accessible, Interoperable and Reusable. FAIRDOM is an initiative to develop a community, and establish an internationally sustained Data and Model Management service to the European Systems Biology community. FAIRDOM is a joint action of ERA-Net EraSysAPP and European Research Infrastructure ISBE.
This database is a global archive and describes plant traits from throughout the globe. TRY is a network of vegetation scientists headed by DIVERSITAS, IGBP, iDiv, the Max Planck Institute for Biogeochemistry and an international Advisory Board. About half of the data are geo-referenced, providing a global coverage of more than 8000 measurement sites.
The University of Oxford Text Archive develops, collects, catalogues and preserves electronic literary and linguistic resources for use in Higher Education, in research, teaching and learning. We also give advice on the creation and use of these resources, and are involved in the development of standards and infrastructure for electronic language resources.
Content type(s)
A genome database for the genus Piroplasma. PiroplasmaDB is a member of pathogen-databases that are housed under the NIAID-funded EuPathDB Bioinformatics Resource Center (BRC) umbrella.
The DCS allows you to search a catalogue of metadata (information describing data) to discover and gain access to NERC's data holdings and information products. The metadata are prepared to a common NERC Metadata Standard and are provided to the catalogue by the NERC Data Centres.