https://docs.google.com/document/d/1UND3RohNBRIwmLbZdTTfom2KOAo-dohsBOteaqOttqU
https://dataoneorg.github.io/provathon-2017/
CB: Challenge: The example that motivates the need to have provenance metadata. It’s clear why writing clear metadata is empowering. Lacking the same for provenance. MJ: Why provenance and use cases.
JL: Benefit in meeting earlier with Archeology since there are shared questions / stories with respect to the ‘why’
CJ: Nice to have both the prose documentation and the script provenance
Josh: Would love to avoid parallel processing and expand rmd
JA: When does the story start? What’s the most useful starting point?
MJ: Linking to publications. Are publications derived from data or do they contain the data?
CB: Is provenance the future of citations?
Why provenance
Use cases: OHI, NOAA
What is useful?
Who uses it?
Merge ProvONE and RMD/Prose methods
References to “other people’s data”
Citations as provenance
Tool compatibility
To reproduce the work, CB and others want an explanation of the process
Need an entry point to understand the where to begin
Reasons:
Get a high level sense of the project
Combination of understanding what has happened in the process. Understanding low level details. If you want to understand someone’s script you can look at it, you can run it and you can look at the provenance. Helps you think about what someone does.
Be able to feel confident about the step of the process that is most useful. Spend a lot of time looking for data for research. Being able to see the steps and locate relevant intermediate files would be useful. For OHI can support communication by reducing requests - being transparent will allow others understand / use the products.
Like to be able to run a ‘make clean’ and see elements, be able to programmatically run an update across all. Auto-execution, smart runs.
Good provenance means good credit for people that write software / data that is impactful
High level motivation. Prose is important - contains the motivation, the ‘why’.
Accountability - chain of custody
Documentation tool for future self
Provenance as a discovery tool across projects
Full network, visual representation is another way to communicate what has been done. Another way to understand the research process. Visual is a benefit.
More efficient representation of the low level details in comparison with prose (recognising there is benefit to prose)
Expose when workflow data dependencies have newer versions
Discussion on data that are used that can’t be redistributed. One solution is to archive the file in a repository, but don’t provide public/read access.
Even data that are publicly available often don’t have persistent identifiers associated with them. This is the case for OHI processing, where driver data are pulled from websites (like SST data), but the data for any given run gets the “most recent” bytes served at the link.
For each dataset used, you need to do your due diligence to see if the agency actually does archive the data in a persistent repository. Web interfaces may provide subsets of archived data that are not persisted.
Researcher desire to cite a paper rather than the data
YW comment annotations could be knittr chunks
YW could detect Rmd headers as nested code blocks (or be generally Markdown-aware to be language-agnostic)
Discuss Indigenous Data Science NEH institute idea
Brainstorm (re)modeling archaeology
Planning for ABQ2019 #SAA2019
Update on publications committee✅
Ben Marwick is preparing a memo for the SAA Publications Committee:
Article level metrics should be emphasised, less than journal impact factor
Citations and References lists should feed into citation databases, as open data (for mining by all)
DORA?: Ethical guidelines for publishing (editors, peer-reviewers), we need to see these in archaeology publications
TOP: “Transparency and Openness for Promotion”, citation standards, data access, etc. Ben suggests we adopt these recommendations / ratings.
Badging for papers that achieve certain levels of reproducibility (the Internet Archaeology journal does this now).
Demo of rrtools package by Ben Marwick
Discussion of site location sensitivity, protection, and de-identification
Hands-on Wholetale if bug fixed
Discuss use cases and provide feedback/critique of yesworkflow, recordr, et al.
Intersection of reproducibility and data prov in current archaeological practice
SAA Board
Gordon
Ricky Lightfoot
Emily
SAA Publications Committee
Here’s the chair of the SAA pub com: Terry Majewski tmajewski@sricrm.com, to make suggestions about future directions for the SAA journals (please help me get her to read my memo! BM)
Proposing that the SAA get on board with DORA, COPE, TOPS, here’s the memo: https://docs.google.com/document/d/1X37JM-O9uyEsB4JPOIt2bUgCQGQRWkstewEui2IEXNo/edit
Link to SAA style guide (2017): http://www.saa.org/Portals/0/SAA%20Style%20Guide_Final_6-5-17.pdf
Future SAA session on peer reviewing to raise awareness about lifting standards
Appeal to NSF program director (John Yellen), “the main professional organisation says we must do …”
Get the SAA Board to pass a resolution that “it would be a good thing if NSF required measurement and closer scrutiny of data management plans” to pass the message onto John Yellen. Who to draft this resolution? Co-sign by:
Digital Data Interest Group,
Open Science Interest Group & Quantitative Methods and Statistical Computing in Archaeology Interest Group,
Public Archaeology Interest Group,
Teaching Archaeology Interest Group,
Committee on Ethics (Arlen F. Chase, arlen.chase@unlv.edu)
Committee on Curriculum (Larkin N. Hood, lnapua@gmail.com)
Public Education Committee (Elizabeth C. Reetz, elizabeth-reetz@uiowa.edu)
TF on Implementing Data Access & Archiving Recommendations (Willeke Wendrich, wendrich@humnet.ucla.edu)
Archaeological specific Data Management Plan (DMP) template for Dissertation improvement grants from NSF.
E.g. Arctic Data Center Program Officer utilizes data standards for grants.
Following in Wholetale, you have to manually open the new rproj after use_compendium()
Ben highlighted the value of setting expectations for community conduct, the code of conduct document to signal these expectations (see: https://github.com/benmarwick/rrtools/blob/master/CONDUCT.md)
Preprint of Ben’s new paper: Packaging data analytical work reproducibly using R (and friends) https://peerj.com/preprints/3192
Rendering LaTeX in rmarkdown:
Related for Jupyter:
Rmarkdown code chunk options (re. Chunk dependencies)
Note: devtools::session_info()
is slightly different from (and better than) the commonly used sessionInfo()
Reproducibility round robin evaluation exercise:
Get a bunch of authors from the https://github.com/benmarwick/ctv-archaeology list of papers to attempt to reproduce each other’s papers
Get funding: rOpenSci, NSF
Present results at SAA, what works, what doesn’t, what’s the
How to get to archaeologists that are not us, to help them learn about this stuff:
Summer schools
SAA webinar (Matt did one)
SAA workshops (Matt and Ben did one), during or after the meetings
Indigenous heritage data science (via NEH)
NSF has separate money for training activities
Archaeology data science field schools (cf. IFR)
WT dashboard not working :-(
Discussion on data versioning (github often cannot be used to deposit data)
Exploring WT dashboard
Ben Marwick’ tale has been showing up many times, despite sustained efforts at deleting!
… and after deleting, the tales are back..
saving new file to home created “no such file or directory”
Successful build/install of Stan; Bayesian MCMC sampling engine and language. https://github.com/stan-dev/rstan
Typically a tough toolchain to make work on first try; but built just fine in Wholetale on first shot.
Now it would be nice to preserve this container so that I don’t have to build it again :)
Kyle B reporting
..
community needs support for versioned data (not under the control of the scientists!?)
drawing data from a db, don’t want to reference all of it; just a subset
good opportunities for using WT in teaching
concurrent use of WT:
how to “take out” a tale (e.g., publish to dockerhub)
another issue: if an image build takes a long time; how to avoid redoing it every time?
Ecology group reporting:
Why provenance? (what for?)
…
complementing Rmd with YW annotations
OHI example
..
YW annotation only took 10 min! ;)
where does the R code go? (not CRAN, not github… DataONE!?)
Prioritization
Community Outreach
Sponsor incubation actual research outcomes (e.g. workshops)
Highlight examples in the community
Archeology group brainstormed - committee work (ESA, ..). Discussed motions to propose to bairds of societies, discussed bottom up approach to behavioral change, workshops for training new people (not sure what that looks like), try to reproduce archeological research and anticipate this to be eye opening. Junior scholars - create materials for this cohort?
What has been successful thus far in the community? And what hasn’t been successful? Showcase examples.
People adopt things if they are easy to adopt. Support ‘leaders’ to get provenance metadata to facilitate the spread of knowledge.
Use other forms of communication - blogs, social media etc for high exposure.
Ted talk on DataONE
Drop the word provenance from materials and talk about reproducible research? Citable data? Workflows? Provenance is different to reproducible. It brings out communication. The distinction is useful. However, is the distinction important to people that don’t even understand reproducible?
Workshop report out
WholeTale working groups
Exposing data via WholeTale blogs
Recycle award - sponsored prize for someone that reuses data. Through WholeTale?
Follow through on getting uses cases (in a collaborative fashion?) completed
Submission review process for reproducible packages. Community developed criteria.
Github badges
Increase visibility through development / improvement of R packages.
Integrate prov (yesworkflow?) support into Rstudio and rmd
Rtools, data repo, provenance interoperability
Feedback on tools through various mechanisms
Linking out to other objects. Introduce ‘scoring’ system for availability of data. (could extend the scoring to an author level metric as a check when entering prov data)
Develop a recommendation (with examples as rationale) for DOI use. What the DOI should resolve to.
Quick follow up survey.
Carl nominated Ben to establish a submission review model. Carl and Matt to support.
Josh will develop out a use case and communicate via slack
Peter to explore yesworklow as a stand alone package that RecordR can use
Ben to explore interoperability of tools. Kyle will support.
Matthew willing to test and guinea pig things as needed.
Bertram will lead the report.
DataONE to explore how to link out to DOIs, recognising variation in ‘class’ of objects.