added two blog posts from conferences

This commit is contained in:
Vicky Steeves 2017-07-13 16:38:51 -05:00
parent 0c452c2ff0
commit c5b9417404
3 changed files with 86 additions and 0 deletions

BIN
images/nbgrader.jpg Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 380 KiB

40
posts/2017-jul14.html Normal file
View File

@ -0,0 +1,40 @@
<!--
.. title: 5 Things I learned at SciPy 2017
.. slug: scipy-report
.. date: 2017-07-14 14:08:25 UTC-04:00
.. tags: conference report
.. category: Professional Life
.. link: https://gitlab.com/VickySteeves/personal-website/blob/master/posts/2017-jul14.html
.. description:
.. type: text
-->
<!DOCTYPE html>
<html>
<body>
<p><a href="https://data-services.hosting.nyu.edu/five-things-we-learned-at-scipy-2017/">See original posting here.</a></p>
<p>I just got back from <a href="https://scipy2017.scipy.org">SciPy2017</a> (I had a talk on ReproZip accepted - <a href="https://vickysteeves.gitlab.io/2017-SciPy/#/">slides</a>)and I learned about some amazing open source tools for research! This year, SciPy 2017 was in Austin, Texas from July 10-16, 2017. It was the 16th annual Scientific Computing with Python Conference, and focused on great new tools and methods for research with Python.</p>
<p>These are my top 5 favourite takeaways from SciPy 2017!</p>
<!-- TEASER_END -->
<ol>
<li><strong><a href="https://github.com/ScienceStacks/SciSheets">SciSheets</a>:</strong> Anyone who knows me knows that I really can't stand Excel. It encodes your data weirdly, and is such a black box it causes more errors in research than it ever helps analysis. This is why I was pumped to see a session on building a better spreadsheet - one that combines programming with the simplicity of spreadsheets. SciSheets is a web application that allows users to run Python expressions or scripts in a spreadhseet, but also export spreadsheets to a standalone Python program! You can find a demo video <a href="https://www.youtube.com/watch?v=N9QQK5n72ZM">here</a>!</li>
<li><strong><a href="https://github.com/jupyter/nbgrader">nbgrader</a>:</strong> This is a phenomenal application for assignment management and grading in Jupyter notebooks. The nbgrader extension for Jupyter notebooks guides the instructor through assignment and grading tasks using the familiar Jupyter notebook interface. It's made up of a few Jupyter Notebook extensions. The formgrader extension allows instructors to use functionality from nbgrader to generate student versions of assignments (including releasing to students), collecting assignments, and auto and manual grading submissions. Students just work in the notebook and submit! You can read more at the <a href="https://github.com/jupyter/nbgrader">GitHub repo</a>.</li>
<div align="center"><img src="/images/nbgrader.jpg" height="50%" width="50%"></img></div>
<p class="caption">The nbgrader workflows from the SciPy presentation.</p>
<li><strong><a href="https://github.com/dataflownb/dfkernel">Dataflow</a></strong>: This extension to Jupyter Notebooks answers the question, "how can a notebook be structured so rewriting isn't necessary?" and "how can cells in a notebook be linked more robustly?" Their solution was to make <strong>cell IDs persistent</strong>, similarly to UUIDs<strong>.</strong> This allows users to powerfully reference previous outputs. You see the slides from SciPy <a href="http://www.cis.umassd.edu/~dkoop/presentations/dataflownb-scipy-2017.pdf">here</a>.</li>
<li><a href="http://joss.theoj.org/"><strong>The Journal of Open Source Software</strong></a>: Ok, I didn't <strong>just</strong> learn about JOSS (I have <a href="http://joss.theoj.org/papers/b578b171263c73f64dfb9d040ca80fe0">a paper</a> there!) but it's still one of my favourite things. It's an open source journal for software. Developers just have to write a short essay (2 paragraph markdown file with some references and an image) and have their code available for review on GitHub. The reviews look at the source code, and test it out before acceptance. From their website: "The Journal of Open Source Software (JOSS) is an academic journal with a formal peer review process that is designed to <em>improve the quality of the software submitted</em>." It's a great way for developers in academia to get their work reviewed, and get credit for their excellent software.</li>
<li><strong><a href="https://github.com/elegant-scipy/elegant-scipy">Elegant SciPy</a></strong> book: Written by Juan Nunez-Iglesias (@jni), Harriet Dashnow (@hdashnow), and Stéfan van der Walt (@stefanv), and published by O'Reilly Media, this fully free and open book focuses on the foundations of scientific python. You can download the book from the GitHub repository as Markdown or an executable Jupyter Notebook. Great work done on opening the book in a machine readable and executable format!!</li>
</ol>
</body>
</html>

46
posts/2017-may26.html Normal file
View File

@ -0,0 +1,46 @@
<!--
.. title: 5 Things I learned at IASSIST
.. slug: iassist-report
.. date: 2017-05-26 14:08:25 UTC-04:00
.. tags: conference report
.. category: Professional Life
.. link: https://gitlab.com/VickySteeves/personal-website/blob/master/posts/2017-may26.html
.. description:
.. type: text
-->
<!DOCTYPE html>
<html>
<body>
<p><a href="https://data-services.hosting.nyu.edu/five-things-we-learned-at-iassist-2017/">See original posting here.</a></p>
<p>I just got back from <a href="http://www.iassist2017.org/">IASSIST 2017</a> and I have to say...I was very impressed! This year, IASSIST (The International Association for Social Science Information Services &amp; Technology) 2017 was in Lawrence, Kansas from May 23-26, 2017. True to it's name, this conference brought people from all around the world:</p>
<blockquote class="twitter-tweet" data-lang="en">
<p dir="ltr" lang="en">A map of <a href="https://twitter.com/hashtag/iassist17?src=hash">#iassist17</a> attendees! <a href="https://t.co/V6fV5Ey5iv">pic.twitter.com/V6fV5Ey5iv</a></p>
— Vicky Steeves (@VickySteeves) <a href="https://twitter.com/VickySteeves/status/867383320371441665">May 24, 2017</a></blockquote>
<script async src="//platform.twitter.com/widgets.js" charset="utf-8"></script>
<p>These are my top 5 favourite takeaways from IASSIST 2017:</p>
<!-- TEASER_END -->
<ol>
<li>An interesting project that was recently published in PLoS One, <em><a href="http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0178261">Research data management in academic institutions: A scoping review</a></em>, which was presented as a poster during the conference. This was essentially a systematic review that was designed to describe the volume, topics, and methodologies of existing scholarly work on research data management in academia. They looked at 301 articles out of the original 13,002 titles. They made the data (the text, methods, etc.) available on Zenodo: <a href="https://zenodo.org/record/557043">Dataset for: Research data management in academic institutions: a scoping review</a>!</li>
<li><a href="https://rstudio.github.io/packrat/">Packrat</a>: a dependency manager in R that looks to solve the problem of "dependency hell" -- that software depends on other packages to run, and these change all the time with no warning, and these changes can break existing code. Packrat works by making a project specific package library, rather than using R's native package manager (which updates libraries as they are released). This means the R code can be packaged up with its dependencies. However, it doesn't pack the version of R, which can pose problems.</li>
<li>Sam Spencer of the <a href="http://aristotlemetadata.com/">Aristotle metadata registry</a> gave a great talk about work done in the open metadata space, giving a strong usecase: government data hosted on <a href="https://data.gov.au/">data.gov.au</a>. He shocked the crowd by keeping metadata in CSV format. He asks for 10 basic fields of metadata from users in CSV form -- and there it stays! He mentioned he was scared to admit this to this crowd, but it's yielded good things for him, including data linkages without explicitly doing linked data. He spoke specifically about using this for geo-metadata; you can check out how it's worked out on this <a href="http://nationalmap.gov.au/">map</a>.</li>
<li>One of the more interesting talks I went to was about digital preservation of 3D data! The speaker laid out 5 methods of creation: freeform (like CAD), measurement, observation, "mix," and algorithm/scanning or photogrammetry. 3D data is difficult to preserve mainly because of a lack of standards, particularly metadata standards. The speaker presented a case study that used Dublin Core as a basis for metadata for the Awash National Park Baboon Research Project's 3D data.</li>
<li>The Digital Curation Network gave an update on their initial planning grant. The DCN allows universities to staff share on data curation, which often is too much for one data curator. The first grant allowed six universities to test how local curation practices translates into a network practice. The next phase includes implementation of the network, during which time other institutions can join. The network also came out with centralized steps for curation:
<ol>
<li><span style="font-weight: 400;">Check data files and read documentation</span></li>
<li><span style="font-weight: 400;">Understand/try to understand the data</span></li>
<li><span style="font-weight: 400;">Request missing information or changes</span></li>
<li>A<span style="font-weight: 400;">ugment the submission with metadata</span></li>
<li><span style="font-weight: 400;">Transform file format for reuse and long-term preservation</span></li>
<li><span style="font-weight: 400;">Evaluate and rate the overall submission using FAIR </span></li>
</ol>
</li>
</ol>
</body>
</html>