-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Make mbon_stats response easier to work with #79
Comments
Do you mean after the table is created? Or to create the table? If the former I believe the GBIF and OBIS are in different columns and no looping over rows is necessary. If I'm mistaken, maybe we need to create two tables instead of one in the function.
That is a good question! Hard to answer without knowing what people will be doing with that table. I like to keep the functions doing a bare minimum and leave the data wrangling part to the end user, b/c that will change more often. In order to implement that we need to identify this minimum table that would empower users to get what they want more easily. (That may lead to the create of multiple functions, easier to maintain, like |
We also have institution identifiers for some of the RAs. It might be nice to do something additional for these, or to be able to query for them as well.
|
NERACOOS and SECOORA have OceanExpert IDs but no OBIS institute pages. Hopefully @sformel-usgs can help sort that out. |
Answer from Pieter:
|
@sformel-usgs thanks for digging into this! Is the tool something open source that we can help fix? I don't know what I'm looking for when I browse to that link and log in with OceanExpert. |
I'm not sure either. I use that tool to update the US node info, but I'm not sure what the back end looks like. I'm guessing from my own permission that we would have to get additional permissions to manage specific organizations and datasets. I can bring this up when I'm in Belgium since I'll be in the building with the OBIS and OceanExpert people. |
@ocefpaf, @laurabrenskelle and I added an mbon_stats function, during the code sprint, to go out and grab statistics about dataset usage from two APIs. The response is a big data frame with some nested dictionaries. I think we're collecting all the data we need, but now it's a matter of being able to parse and use the response.
I'm curious if you could take a look at the function and see what we can do to make it easier to work with the data. As it stands we have to do row wise iteration to split out the GBIF and obis download information and that seems overburdening.
A generic question I have is how much do you have the function do vs how much do you do data wrangling in the use case notebook?
Any advice is appreciated!
The text was updated successfully, but these errors were encountered: