strong recommendation
#LoveData23 postprocessing
magnificent talk by Sahra Lamdan
"For Love or Money?"
This talk will provide much more insights than you might suspect at the moment, ๐ค
think of Alice falling into the hole.
๐คจ
Do make time and listen. ๐
https://youtube.com/watch?v=Qxd8DmktT6U&feature=shares
RT @DataCite
Yesterday's webinar was great initial conversation about the Open Global Data Citation Corpus, with exciting presentations & statements from key community stakeholders.๐ฅณ
See the recording & slides (in the description) here:๐ https://youtu.be/NRTZXaJuxT8
#LoveData23 #DataCitation
And then... share it! Because if you do something, anything, with your city data (especially Boston's!) I would love to see it and share it. If you need any help/suggestions, let me know - I would love to help ๐
Let's show our civic data some love!
Now... do something with that data! Download it or access it through an API. Summarize, filter, or otherwise organize it. Visualize it! Figure out if the existing documentation for that dataset is sufficient. In other words, show that civic data some love ๐ฅฐ
(Alternatively, figure out what closest city does have an open data portal)
Then, find a dataset that you could use to answer a question you have. Maybe you want to figure out what data is available about the building or neighborhood you live in.
For this final day of #LoveDataWeek I have a challenge for you:
Go find your city's open data portal. If your city doesn't have one, see if you can find where your city does publish some/any data. Often cities will at least publish some GIS data. What data is there?
successfully applied #OpenRefine to reshape a table buried in large rdf specification html into markdown to keep for offline reference
OpenRefine is ๐
credit to #LoveData23 for the refresher
It's International Love Data Week!
JSSAM loves data on statistical and methodological issues for surveys, censuses, administrative record systems, and other related data, incl. sample design, inference, nonresponse, measurement error, mode effects, paradata & responsive survey design.
#LoveData23
Check out this great blog by my colleague from earlier this #LoveDataWeek
Highlighting women in science and open research practices
#LoveData23 #InternationalWomenandGirlsinScience #OpenData #OpenAccess #OpenResearch
#openresearch #openaccess #opendata #internationalwomenandgirlsinscience #lovedata23 #lovedataweek
This is my kind of #LoveData23. My retirement plans include spending days messing around with Spotify's API so I can find out things like this. Why do songs have so many writers credited these days?
https://tedium.co/2023/02/04/why-do-modern-pop-songs-have-so-many-credited-writers/
#LoveData23
workshop:
Computational Approaches for Text
at Sherman Center of Digital Scholarship
#OpenRefine #jupyterNotebooks #NLP
very interesting and excellent unfolded by Devon and Jay ๐
#nlp #jupyternotebooks #openrefine #lovedata23
Behold, the knitted results of Tuesday's poll about how you all feel about data! There were about as many Mastodon responses as in-person colleagues responding ("Survey"), only 2 responses from Twitter, and a handful of people who came by our table at the #LoveData23 #LoveDataStanford Fair. #Knitting #DataViz #DigitalHumanities
#digitalhumanities #dataviz #knitting #lovedatastanford #lovedata23
It's International Love Data Week!
JSSAM loves data that comes with code. Where ethically feasible, JSSAM strongly encourages authors to make all data and software code on which the conclusions of the paper rely available to readers.
It's International Love Data Week!
JSSAM loves data that comes with code. Where ethically feasible, JSSAM strongly encourages authors to make all data and software code on which the conclusions of the paper rely available to readers.
Then, you can keep on exploring! I wonder what the top keywords are...
df_keywords = df[['identifier', 'keyword']]
.explode('keyword')
df_keywords.groupby('keyword')
.agg({'identifier': 'count'})
.sort_values('identifier', ascending=False)
.head(10)
Open a new notebook at http://colab.research.google.com and try running this code:
import requests
import pandas as pd
data = requests.get('https://data.boston.gov/data.json').json()
df = pd.json_normalize(data['dataset'])
df.describe().T
Here's another cool feature of Analyze Boston (courtesy of CKAN): you can very easily get metadata about all resources on the site by visiting:
http://data.boston.gov/data.json
The metadata is in JSON format, so you can easily write a script to start examining it...
The usual greatest hits (or should that be misses?) of data and spreadsheet disasters here - Utah's education budget, JPMorgan Chase risk calculations, genomes as dates, Rogoff and Reinhart, the ONS's recent fail. Plus, a good discussion on how spreadsheets should be used and not misused. https://www.bbc.co.uk/sounds/play/p0f2cytq #LoveData23