owned this note
owned this note
Published
Linked with GitHub
---
tags: pyopensci, python
---
# pyOpenSci Meeting Notes - 5 December 2019
## Attendees
* Leah Wasser (Earth Lab, University of Colorado - Boulder)
* David Nicholson (Emory University)
* Jenny Palomino (Earth Lab, University of Colorado - Boulder)
* Martin Skarzynski (NIH)
* Max Joseph (Earth Lab, CU Boulder)
* Aaryn Olsson (University of Arizona Adjunct, Planet Labs)
* Luiz Irber (DIB Lab, UC Davis)
* Chris Holdgraf (UC Berkeley, Project Jupyter)
* Sasha Kielbowicz
* Sander van Rijn (Leiden University, Software submitter)
## Agenda
1. New Software Submission: https://github.com/pyOpenSci/software-review/issues/15 Can someone with some understanding of multi-fidelity analytical benchmark functions help us understand if we can review this package? he'd like to submit to JOSS as well. It would be nice if we can review.
DESCRIPTION BELOW:
> Analytical benchmark functions are a like well-established datasets in the field of machine-learning: by comparing your algorithm on the same ones, you can accurately compare your results. They are essential for reproducibility, as typos in the formulas will obviously give very different results, but are not an 'external' utility like the examples given under Package Categories in the guidebook.
>
Several of our collaborators are still not sure what this package this does.
This package may be out of scope for us right now.
We should consider -- adding additional info to the submission review. maybe about what a readme contains. etc...
We should point the author to the current guidelines that contain recommendations for structuring a README: https://www.pyopensci.org/dev_guide/packaging/packaging_guide
1.a. Notes of short discussion with submitting author (Sander) who joined a bit later:
* README/documentation was indeed still a work in progress.
* Was indeed uncertain about whether it was in scope for pyopensci.
* Given the limited pool of currently available authors, package will *not* be submitted to pyopensci for now, but just to JOSS instead. May resubmit when pyopensci has grown, if the package is definitively considered within scope.
### How do we track people / reviewers, etc?
Do we want to keep track of people somehow that are willing to review ?
Google Sheet / Form --
Give people the option to opt in -- or make it anonymous ??
### 3. AGU 2019
What is the scope of pyopensci / ropensci
* David: I did my science, and i wrote some code. can someone else use that code...
* Max: Research compendium -- how to package analysis associated with a paper... this is an ropensci in scope item that pyopensci could support.
* Sander: reproducibility compendium -- was introduced as part of a reprodcibility hackathon
* LINK from Sander: https://reprohacknl.github.io/ReproHack/
* Twitter account of the event: https://twitter.com/ReproHackNL
* Presentation about reproducibility & research compendium: https://annakrystalli.me/talks/reproducibility-value-in-practice.html
Here's some info on research compendia: https://research-compendium.science/
Here's some rOpenSci content relevant to reproducibility and research compendia: https://ropensci.github.io/reproducibility-guide/sections/introduction/
Link from Aaryn: https://github.com/benmarwick/rrtools
* Chris: if we accepted these compendiums, it would be a lot of work to review as the paper would need to be reviewed.
* Aaryn: Clarify Dev Guide Aims and Scope wrt compendia
### what data should we collect?
* Why does doumentation matter
* could we consider asking questions that are much more basic
* where do you see issues??
* Have you developed a python package?
* if so - why?
* did it include docuentation
* tests,
* CI
* etc
* if not - why?
* Possibly more basic questions to start a group discussion at AGU session
* How do you share your analysis code with others on your team, others, the broader scientifc community? Challenges there?
* How do you find code examples for your work now? Challenges there?
Post AGU -- IRB --
We could potentially team up with ropensci on the education side of things.
martin: there are mnay options associated with python packages
How to best manage an old school (prepared slides only) room??
Static link to a Dashboard page for attendees to explore, possibly with walkthrough, though attendees could be offline as well given the volume of folks accessing network
## A few notes from convos with Leo and Lindsey and others
Action items - What could be done to make software “value” more visible -- the academic challenge of getting “Credit”
Even the open source software -- abstracts lose attention because people often get funding for research not software
Publication --- when you publish you don’t get the same credit as you do for papers
What are the challenges associated with creating reproducible code /tools
Have
What is missing from your toolkit to make workflows more reproducible?
What sorts of support do you need to...
List the components of a good open source software package
Which do you understand the least
Discovering/ finding tools…
3. "Organizing Code" blog post
- David would like to discuss if that's ok :D or we could wait til next time
- (https://pyopensci.discourse.group/t/software-dev-best-practices-python-102-blog-post/133)
- example projects (PyOpenSci obvs)
- diff b/t open sci v open source dev