We began this week by contacting the Career Center, Alumni Relations, the Admissions Office, the Dean of Students, Gould Library, and Institutional Research & Assessment Office. While we heard back from many of the offices, they unfortunately did not have the data we were looking for, although many of them expressed support of the project.
Since we were not able to obtain the data from the administration, we had to design ways of collecting the data on our own. We scraped the Alumni Directory by job field and exported the data to a CSV file, capturing data on locations, majors, class year, and industry. We also used Optimal Character Recognition on the Zoobooks from 1970-2020 and exported the data to a CSV file. This created our two datasets: one on alumni locations and one on student origins.
Even though we were able to gather the data, it was extremely messy and we had to do extensive data cleaning using OpenRefine, Python, and manually through Excel. For each of the datasets, we had to delete rows that had not been processed correctly or were missing information; separate the geographic locations by city, state, zip code, and country; and add spaces to some of the cities that were multiple words long. We are currently geocoding the data in Google Sheets so we can map the locations by their coordinates in ArcGIS.
In regards to problems that we have faced, the speed at which we gather, clean, and analyze the data that we’re interested in has been limited by a lack of replies from the relevant offices, and as a result we have relied more heavily on scraped data, such as that from the Alumni Directory. These data, however, have their own set of challenges, as we are left with an incomplete sample of alumni data, and are ultimately unable to complete as comprehensive an analysis as we’d like. Another issue that we have encountered relates to the mapping of the data points that we’ve extracted. In particular, it is not possible to geocode zip codes to translate them into coordinates that can be mapped cohesively in our utilized mapping softwares. A proposed solution to this is to split our current data into two different CSV files, those for which coordinates were able to be extracted, and those for which zip codes must be utilized. We would then import these files separately into ArcGIS, and then merge them into a single layer.
The plan is currently on track and there were only minor changes. Our updated timeline of deliverables includes: geocoding and importing CSV files into ArcGIS by February 26th; visualizing data by March 3rd; compiling story map by March 6th; and completing presentation by March 8th. Visualization will be done in ArcGIS, Palladio, and Flourish.
My blog post
Here is a link to my personal blog post!
https://blog.enriquezd.sites.carleton.edu/digital-humanities/final-project-first-weeks-progress/
Here’s my personal post: https://hh2021.amason.sites.carleton.edu/uncategorized/final-project-week-8-update/
Here’s my blog post: http://brindleg.sites.carleton.edu/hacking-the-humanities/8th-week-my-final-project-contributions/
Personal blog post: https://dgah110.rayxu.sites.carleton.edu/final-project/final-project-update/
Team Carls Around the World,
You all have been making great progress on the data acquisition and cleaning side, and from what I’ve seen in slack are also well on your way to great visualizations. This post does a good job of outlining the initial goals and places where access or data quality forced compromises, which you will definitely want to elaborate on in your final project website. Addressing the missing data and acknowledging how they might have altered your interpretations and conclusions will be a key aspect of the final analysis.
Your presentation website has already got a clean theme and some containers to hold the content, and I look forward to seeing it start to fill in.
Personal blog post (from last week) https://hh2021.amason.sites.carleton.edu/uncategorized/final-project-personal-update/