SGCI webinars bring together community members across the globe.

Content with tag data .

Webinar: DataONE—Data Partner for Science Communities

April 10, 2019

DataONE—Data Partner for Science Communities

Presented by

  • Robert J. Sandusky, Ph.D., Associate Dean for Information Technology & Associate Professor, University Library, University of Illinois at Chicago
  • Amber E. Budden, Ph.D., DataONE Co-PI and Director for Community Engagement & Outreach, Arctic Data Center Co-PI, Senior Research Fellow at the National Center for Ecological Analysis & Synthesis
  • Karl Benedict, Ph.D., Director at Research Data Services, Associate Professor at College of University Libraries & Learning Sciences, University of New Mexico

DataONE federates more than 45 independent scientific data repositories from six continents providing access and discovery, preservation, data provenance, metadata quality services, data citations, data science training, and more. This webinar will provide an overview of existing capabilities and services currently in development to support a wide range of scientific communities and data formats.

DataONE was funded in 2009 by the National Science Foundation to provide solutions for long-term access and preservation of data generated by other projects and conduct a broad education and outreach campaign to transform scientific practice. DataONE is a global leader in research data infrastructure development, research, and education.

Webinar Slides

Resources mentioned during the webinar:
DataONE API 

Watch on YouTube


Webinar: National Data Service (NDS) Labs Workbench

December 13, 2017

National Data Service (NDS) Labs Workbench: A Scalable Platform for Research Data Access, Education, and Training

Presented by Craig Willis, Technical Coordinator at NDS and Senior Research Programmer at National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign

The growing size and complexity of high-value scientific datasets are pushing the boundaries of traditional models of data access and discovery. Many large datasets are only accessible through the systems on which they were created or require specialized software or computational resources for re-use. In response to this growing need, the National Data Service (NDS) consortium is developing the Labs Workbench platform, a scalable, web-based system intended to support turn-key deployment of encapsulated data management and analysis tools to support exploratory analysis and development on cloud resources that are physically "near" the data and associated high-performance computing (HPC) systems.  The Labs Workbench may complement existing science gateways by enabling exploratory analysis of data and the ability for users to deploy and share their own tools. The Labs Workbench platform has also been used to support a variety training and workshop environments. In this webinar, I will present the Labs Workbench platform and discuss several key use cases. I will also discuss findings from the recent Workshop on Container Based Analysis Environments for Research Data Access and Computing which further highlighted compatibilities between science gateways and interactive analysis platforms such as Labs Workbench.

NDS Labs Workbench Webinar Slides

Watch on YouTube


Webinar: Data and Software Carpentry: Using Training to Build a Worldwide Research Community

May 10, 2017

Data and Software Carpentry: Using Training to Build a Worldwide Research Community
Presented by Tracy Teal, co-founder and the Executive Director of Data Carpentry, and Adjunct Assistant Professor with BEACON, Michigan State University

Although petabytes of data are now available, most scientific disciplines are failing to translate this sea of data into scientific advances. The missing step between data collection and research progress is a lack of training for scientists in crucial skills for effectively and reproducibly managing and analyzing large amounts of data. Already faced with a deluge of data, researchers themselves are demanding this training. Short, intensive, hands-on Software and Data Carpentry workshops give researchers the opportunity to engage in deliberate practice as they learn these skills. This model has been shown to be effective, with the vast majority (more than 90%), of learners saying that participating in the workshop was worth their time and led to improvements in their data management and data analysis skills. Data Carpentry events have trained over 20,000 learners since 2014 on 6 continents with over 800 volunteer instructors. The strategies of growing this community could be applied toward growing communities of gateway users, particularly by offering training and demonstrating the value of the skills and tools that will enhance their work.

View the slides (Slideshare)

Watch on YouTube

 

Questions asked during the webinar (and some answers)

If you have further questions, you are welcome to contact Tracy at tkteal AT datacarpentry DOT org.

Q: What is the relationship between SGCI and her organization? There seems to be some overlap...in training, for example? 
A: Currently there is no formal relationship between SGCI and Data Carpentry, but we definitely want to look into that option further!

Q: I wonder how additional topics and instructors get added to the set of offerings.

Q: When discussing "active learning", she used an acronym - IBU? IVU? What's that?
A: I, We, You [First the instructor shows it, then we do it together, and then you do it yourself.]

Q: Do you request attendees install software before a workshop (or during)? In the Python ecosystem, do you recommend a particular distribution?
A: Anaconda

Example of a lesson: http://swcarpentry.github.io/lesson-example/

Q: So what about Jupyter? Do you use it?

Q: Where does the instructor training take place? and is there also a cost for this?

Q: Who pays for the volunteer instructor's travel?

Q: What is the relationship between Data Carpentry and Software Carpentry.


Webinar: Building a Modern Research Data Portal with Globus - Introduction to the Globus Platform

March 8, 2017

Building a Modern Research Data Portal with Globus - Introduction to the Globus Platform
Presented by Steve Tuecke and Greg Nawrocki, University of Chicago - Globus.org

Abstract: Science DMZ (a portion of the network optimized for high-performance scientific applications) architectures provide frictionless end-to-end network paths; and Globus APIs allow programmers to create powerful research data portals that leverage these paths for data distribution, staging, synchronization, and other useful purposes. In this tutorial, we use real-world examples to show how these new technologies can be applied to realize immediately useful capabilities.

Attendees will develop an understanding of key identity management concepts as they are applied to data management across the research lifecycle, and will be exposed to tools and techniques for implementing these concepts in their own systems.

We will explain how the Globus APIs provide intuitive access to authentication, authorization, sharing, transfer, and synchronization services. Companion iPython/Jupyter notebooks will provide application skeletons that workshop participants can adapt to realize their own research data portals, science gateways, and other web applications that support research data workflows.

Download a PDF of the slides

Watch the YouTube recording

Answers to questions asked during the webinar

The slides have many links to various online resources. If you don't see what you are looking for, feel free to contact Greg greg@globus.org directly.

Q: The globus sample portal is written in which language?
A: Python.

Q: For share endpoint, one can't see another share endpoint right?
A: Someone can see the data in an endpoint only if it's been explicitly shared with them. The endpoints themselves are all publicly visible.

Q: Does the Transfer/Share API include download from a share/endpoint to local machine that is not an endpoint?
A: All transfers are to and from endpoints. Globus Connect Personal is a very easy way to set up an endpoint on a local machine: https://www.globus.org/globus-connect-personal

Q: What if someone doesn't want to set up a personal endpoint? We just have resistance from people who don't want to setup a personal endpoint for infrequent downloads.
A: Native “in browser” http transfers are on the roadmap. Transfers themselves are easy; getting them to work within the constraints of our security model requires care. We should have some more concrete timelines for delivery at GlobusWorld in April.

Q: When the user "logins" to the gateway, the gateway redirects to Globus and the user signs in, then Globus redirects back to gateway, is https required for this process or is http ok?
A: Https is required, standard OAuth.

Useful links:
http://docs.globus.org/api/transfer
http://docs.globus.org/api/auth