Category Archives: readings

Turning “Digital Natives” into Digital Historians

One of themes that spans across the readings for this week about teaching digital history, is the notion that “digital natives” (the generation of students who have grown up with computers) are expected by their professors to be comfortable and confident in their use of technology and media. However, despite knowing how to use Facebook, YouTube and other social media, these students are lacking in their digital literacy skills. Understanding, interpreting, evaluating and creating media is not the same as consuming media, and are skills that need to be honed by future digital historians.

Through several case studies, we can see how professors are attempting to teach students these vital skills, with varying degrees of success. At first glance, T. Mills Kelly’s experiment of having his class create an historical hoax struck me as a bad idea. However, throughout the course, the students were able to learn media literacy skills by being the hoax creators themselves. The fact that the hoax was meant to be revealed at the end the semester helped sway me into believing that this tactic was clearly more educational rather than mischievous. However, the end result was much more convincing. The students had actually learned how easy it is for hoaxes to fool people, how important it is to check sources, and actual maritime history about the Lower Chesapeake during the post-Civil War era by concocting a fake researcher who has discovered the “Last American Pirate.” While reading the chapter, I was actually glad to see that the students’ hoax had worked so well as to trick the academic world of historians, as I am sure that reinforced to the students the benefits of information and media literacy.

Another hands-on experiment that didn’t go as well as the professor had hoped, was Allison Marsh’s attempt to teach Omeka to her museum studies students. In trying to prepare her students for the real world that is increasingly more reliant on digital technology, Marsh found that her students weren’t as tech-savvy as she had thought. Omeka isn’t extremely hard to use, and I’m finding it very useful in my halfway completed final project, but for students who haven’t had to create metadata or lack basic graphic design knowledge, I can see how it could be a burden for some students. I think Marsh is correct in telling her students that they no longer have a choice in working with technology and digital media, as there is no going back to purely analog museum exhibits. From my own personal experience, if I visit a museum’s webpage and they don’t offer any type of virtual exhibit, my interest immediately wanes. I’m curious to see how Marsh’s 10 year project unfolds, and I hope future classes are still required to build some form of digital exhibit, if only to stay relevant within the public history field.

While it can appear baffling that digital natives are not prone to using technology for educational purposes as much as we want them to, I think there is room to grow. If more professors take on new approaches to using technology and digital media in the classroom by applying more active learning techniques such as Kelly, Marsh, Rabinowitz and Trepanier have done, the more capable students will become in utilizing technology and media in their own research. Some of these students may go on to work as digital historians and possibly go on to teach their own students, hopefully passing on the lessons they have learned from their professors who have been able to bridge the gap between using and consuming media.

Notes from Discussion on Copyright/Open Access/Fair Use

I think Monday’s discussion on Copyright/Open Access/Fair Use went rather well considering that the topic is very nuanced, especially when dealing with digital history. The discussion on the embargoing of dissertations wasn’t as vociferous as I had anticipated, but maybe my expectations were based upon the strong criticism that followed the AHA’s statement on the subject on various blogs and comment sections. Our class discussion was much more milder, although several good points were made.

It appears that the AHA Statement was a sounding board for historians to express their opinions on issues that are much larger than the actual statement. The issue of embargoing dissertations isn’t the problem – the problem is that the field in general has been slow to adapt to the digital world. This is evidenced by their lack of ability to establish set criteria for promotion and tenure for digital historians, which is also true for public historians. The AHA seems happy to stand by the book as the end-all be-all for Phd students after they complete their dissertation, but the AHA and universities can’t seem to break free of the printed final copy. This poses problems for digital historians who would be more likely to create something digitally than in traditional print format.

We did talk about Open Access models and how historians and other academic writers have been writing for free forever, and Open Access doesn’t really change that. In the author pays model of OA, academics are even asked to pay to publish. Compared to the sciences, the number of Open Access journals in the humanities is drastically low. A change in the overall system needs to happen in order for OA journals in the humanities to grow.

Overall, the discussion touched on many valid points that were culled from the week’s readings. I think some of our questions could have been rephrased or re-worded, as we didn’t get through all of them. We missed out on talking about the consequences of not being familiar with fair use in the art world, especially with art historians. The topic of copyright and fair use is complicated, but the more we know about it, the more confident we can be about what is acceptable to use in our digital projects.

To Embargo, or Not to Embargo (Your History Dissertation)

Not actually being a Phd student and having to worry about what I’m going to do after writing a dissertation, I really had no idea that the issue of dissertation embargoes, especially in the history department, was such a divisive one. When the AHA issued their Statement on History Dissertation Embargoes last year, a chorus of critics in favor of Open Access quickly lashed out on much bigger problems than the statement intended to address. Some of these other issues that were mentioned in various blogs and comment sections, like promotion and tenure policies in the field and the business practices of book publishers, seemed to miss the point that the AHA was just trying to propose an action that would benefit junior scholars by allowing them more control of their finished work.

By allowing embargo periods of up to six years, junior scholars have the option of withholding their dissertations from being published online. On one hand, this option gives dissertation writers a choice in how they want their work to be accessed, in case they are in the process of negotiating a book deal with a publisher who would prefer that the dissertation should not be freely accessible. On the other hand, it limits the sharing of ideas and new research to fellow historians and other potential readers. I can see both sides of the argument, but personally I think I fall on the side of immediate access to dissertations, sans embargo, but I do think it should be the author’s choice. As someone who works in a library and often runs into journal embargoes (usually six months) while helping researchers access information, I know how frustrating it can be to be denied access to a particular article at the point of need.

Out of curiosity, I wanted to know what Mason’s stance on the embargo question was. I found the University Thesis & Dissertation Services (UTDS) FAQ page, which actually covered the embargo issue. As of last year, all dissertations must be submitted electronically to be stored in MARS (Mason Archival Repository Service), while individual academic departments can still require print copies. UTDS does provide an option to withhold finished works from public view for a specified time period. However, even if a student doesn’t want an embargo placed on their finished dissertation, they still must submit a form. The options for embargo periods are six months, a year, or 5 years, and if selected, the student must provide a reason for doing so, as well as collect signatures from committee chairs and the dean of their department.

It’s great that Mason is allowing students to chose whether or not they want their work to be made immediately available by allowing the embargo as an option as the AHA suggests in its statement, but I wonder if requiring signatures and an explanation makes it easier for students to just acquiesce to allowing public access as soon as it is uploaded to MARS. I would hope that more students would be more discerning, though, as they have put years of work into writing and completing a dissertation. Many factors contribute to such a big decision, and I am not sure what route I would choose, especially since I’m still just trying to finish a master’s degree and not sure what my potential dissertation would even be about at this stage. However, I feel more knowledgeable about the embargo debate from history student’s perspective now and not just from a librarian’s point of view.

Digital Scholarship: Moving Forward while Looking Back

The readings about digital scholarship in the humanities this week have me thinking about  the future. As Hitchcock suggest that the book is being replaced by digital texts, I wonder what scholarship, print or digital, will look like in 25 – 50 years from now. Will books still be printed? Will something else come along that replaces digital texts? I cannot see into the future, but barring some type of catastrophic event, I cannot envision the technology we have today regressing. I can only picture it advancing, as it has been doing over the past 50 years. I feel as though right now we are in the middle of a transformation and in the near future, academics will no longer question the legitimacy of digital scholarship, as it will become fully integrated into our everyday lives.

As younger generations who are used to utilizing the web for research start moving to publishing online, as has already been happening, “traditional” historians will likely either jump on board, or get left behind. The argument that digital scholarship is less credible has been stated many times, and also disproven. With Will Thomas‘ and Dougherty, et al.’s writings about the peer reviewing practices on digital scholarship, it is obvious that credible and reviewed work can be digital.

By embracing the open culture of the web, more readers will have access to historical writing, which I think is a good thing. Academics are trapped between moving forward with the technology available to them to present their writings, teachings, and other forms of scholarships, and being held accountable to out-dated standards that privilege the monograph books. I agree with Galarza, Heppler, and Seefeldt’s Call to Redefine  Scholarship in the Digital Turn, as they encourage the American Historical Association to take a serious look at the current state of digital history, in order to keep up with modern tools available to historians. Their proposal has an emphasis on reevaluating the current standards for tenure and promotion by history departments, which speaks to the underlying issue that digital scholarship is still not taken as seriously as traditional methods.

My hope is that in 50 years or less, the discussions we have today on digital scholarship will be unnecessary, and we will look back on this time much like we look back on other innovations in history. By creating credible and well-researched digital history projects, as well as advocating for a change in the academy, digital scholarship will hopefully be as accepted just as much, if not more, than the print book in the years to come.

Using Games to Teach History

This week’s readings about history’s place in gaming brought up issues such as form over content, games as instructional aids, and whether or not games can be a worthy outlet for historical scholarship.  I have long believed that games can be a useful tool in the teaching and learning of a subject, but have been cautious in believing that historical games are historically accurate. Most games are created with the sole intent of entertaining and making profits, which can leave the educational component lacking. As a kid, I remember playing computer games in school, such as Oregon Trail. At the time, I didn’t consider that game to be educational since I loved playing it so much, but looking at it from my perspective now, I can see how it used an episode of American history to teach school kids about the strife of traveling West by covered wagon in 1848 (and also what dysentery was!). So even though I didn’t realized I was learning something, I actually was.

Looking at The Lost Museum or Pox and the City, I can relate to my childhood experience with Oregon Trail. By creating an interactive and 3-D mystery game, I can envision school children using this website to discover P.T. Barnum’s American Museum and its subsequent fire in 1865. However, whenever I find myself visiting the Lost Museum, I seem to navigate to the Search the Archive page, where I can find primary sources that relate to the museum itself, rather than playing the actual game.

With Pox and the City, the player is placed inside the world of Edinburgh circa 1800 amid an outbreak of smallpox. Taking a real-life event, such as Edward Jenner’s experimentation with inoculating smallpox victims with cowpox material, and turning that into a game that builds on narrative interaction with other players, helps the player learn about the social history of that real-life event. This game reminded me of another childhood favorite computer game that I played obsessively in 1995 called Dust: A Tale of the Wired West. Dust, while loosely based on an historical time period (American West, 1880s), made no claims of being an historical game, especially since it took place in the fictional town of Diamondback, New Mexico. However, thematic elements such as silver mining, Chinese immigrants, and Native Americans captured my interest and led me to do my own research on these topics outside of the game. The role of interactive narrative is also similar to Pox in the City, in that certain dialogue responses would yield different outcomes.

Based on my own experience and the articles this week about Pox in the City that highlight its historical accuracy and user interaction, the ability to educate through gaming seems undeniable. I also believe that historical games can be their own form of historical scholarship, as Trevor Owens ponders in his Play the Past entry. By changing the format of scholarship, historians are able to break out of the traditional mold of papers, articles and books, and engage audiences more interactively.

Crowdsourcing History, the Good and the Bad

This week, we are taking a look at crowdsourcing and how it can help and also hinder how history is used on the web. The idea of crowdsourcing in history isn’t  new, as many genealogists and volunteers have been helping archives for quite some time. However, the transition to the Internet has opened up more avenues for crowdsourcing than was previously available. As the walls between professionals and amateurs become less noticeable, it’s important to think about how crowdsourcing can open up new methods of collecting and distributing information without historians having to play the role of gatekeeper.

Prior to this week’s readings, my frame of reference for Wikipedia was from a library science standpoint, which encourages anyone who wants to cite Wikipedia to try to investigate the references from the article. However, a few authors brought up the issue of the gender-divide when looking at Wikipedia, which is something I had not considered previously. While Wikipedia is known for its factual writing that prohibits any original research or bias, that anyone in the world can contribute to, I had not realized that most articles were written by men. Both Rosenzweig and Madsen-Brooks each acknowledge that the average Wikipedia editor is  English speaking and male, which has shaped the types of articles that get attention by editors. This raises questions about what voices are being heard on Wikipedia and how diverse these voices are. Is it really crowdsourcing if only one segment of the population is contributing? There are other issues surrounding Wikipedia, such as the editing wars and incorrect information, that cloud its effectiveness in being a global online crowdsourced encyclopedia, but the issue about gender was the one that struck me as the most intriguing.

Other crowdsourcing projects, like the September 11 Digital Archive and the Hurricane Digital Memory Bank serve to collect personal testimonies about an event from those who experienced it. These projects reminded me of traditional oral history projects, but captured online. The issue of creating a functioning website that allows for easy submission by the user appears to be the biggest hurdle after getting the funding necessary to take on such a project. Projects like these are great in collecting the stories, images, and information that might not survive due to the impermanent nature of the Internet. The staff necessary to make these projects cannot be overlooked as well, as they are not so much playing the gatekeeper in these online archives, but are more like the construction crew. In a project like Transcribe Bentham, the majority of funding was spent on staff, as the project relied on the public to transcribe the manuscripts. In this case, although volunteers were used in the transcription process, the project’s moderators had to ensure quality, as they were going to be ultimately published in the Collected Works edition. While anyone could assist with the project, the staff still retained editing control in the final version, which makes this project an interested experiment. I have been seeing more of these kinds of crowdsourcing transcription projects (probably because I like transcribing) and have even signed up to be a Smithsonian Digital Volunteer to try to do my part.  Although, I should add that I have not had time yet to take on any projects yet as a volunteer.

The issue of crowdsourcing naturally goes against the professional historian grain, especially when we start to think of ourselves as gatekeepers. However, there are benefits to be had, especially in projects that require time and effort that dedicated staff just cannot feasible do in a short period of time. Collecting and capturing original source material is one of the better uses of crowdsourcing history, as it provides another way for archives and digital projects to gather information that might have otherwise been lost.

 

Public History on the Web

The confluence of public history and technology has enabled the average person to explore historical topics without ever leaving their home. Over the past two decades, the growth of websites dedicated to public history, including museums, historical sites, archives and libraries has been countered with a decline in actual in-person visits. The idea of the museum as a place is something that I have questioned before. Does a visit to the physical museum trump a visit to a virtual one? Or is it the other way around? I see benefits in both types of visits, especially if done well.

When Smith asked if serious history could be done on the web back in 1998, the web looked much different than it does today. Besides benefitting from cleaner and more sophisticated web design, today’s historical websites offer more interactivity than previously available, as evidenced by the evolution of the Great Chicago Fire website produced by the Chicago History Museum and Northwestern University. Although it maintains its core features of images, texts, and essays, the site today is visually different and more user-friendly than the original version.

The ability to create an exhibit online, such as the Great Chicago Fire, allows historians to bypass the traditional route of relying on a physical space. While this makes accessing more original documents and images that one probably wouldn’t encounter in a traditional museum exhibit, the experience of visiting a museum in person is lost. Although virtual tours are common, as described in Anne Lindsay’s #VirtualTourist article, there is something about the museum experience that cannot be replaced fully online. But does this matter?

In a way, the online museum is much more “public” than traditional public history institutions. By increasing access to those who cannot visit in person, allowing educators to bring museums into the classroom,  allowing users to help organize the National Library of Australia online, and putting a city’s history into the palm of one’s hand through a mobile application, the democratization of history can easily been seen. I think serious history can be done on the web, it just looks different than what traditionally has been done in the past. By combining the physical museum experience with the digital one, there is a greater potential of creating more dynamic and interactive examples of public history. In another twenty years public history on the web will probably be even more seamless, especially as we become more accustomed to accessing everything, including history, online.

Mapping History

From this week’s readings, it appears that there has been some disconnect between historians and geographers in the past. However, there has been an effort made by digital historians recently who have gravitated towards geographic information systems (GIS) to link the two fields together for scholarly purposes. Tim Hitchcock suggests that this division was partially due to geographers’ seeking more secure academic funding by aligning themselves with STEM departments rather than with history/humanities and as a result, there has been little dialogue between the two fields. However, the availability of user-friendly GIS programs and mass digitization of historic texts and lots of data, have opened the doors to collaboration between history and geography in a digital format. History can now be geo-referenced, as historical maps can be digitized and layered, as well as analyzed, as the examples covered illustrate.

The Visualizing Emancipation project is a prime example of how mapping historical data can reveal patterns through imagery. I could spend all day playing with Visualizing Emancipation. By allowing the user to select an Emancipation Event, such as the “Capture of African Americans by Union troops” or “Conscription/Recruitment” by either the Union or Confederate Armies and filtering by source type, such as book, official records, newspapers and/or personal papers, researchers are able to explore a previously primarily text-based set of documents in a visual geographic manner. This pairing is helpful to those, like myself, who like to match a document to its complete history, including its geography. With adjustable map options, users can choose how they want to visualized the data presented to them and click on specific events to learn about its details. One of the tool’s most handy features is its ability to link directly to the actual source material that is geo-referenced, as it allows users to interpret the data for themselves. The map visualization helps us to take on a new perspective on such a large topic on different scales, allowing us to expand and contract our view of emancipation in a geographic way.

On a smaller scale, geographically, Digital Harlem is somewhat similar to Visualizing Emancipation, in that historical records have been geo-referenced to reveal patterns of daily life in Harlem from 1915 – 1930. By selecting a type of event, the name of a person or place, users can create custom maps that plot each event, with their descriptions as well. Multiple layers can be built on top of a Google map and/or an historical map, helping us to visualize the geographic display of archival sources, such as crime records and newspapers.

With GIS, historians can use maps to explore historical themes, such as with the ORBIS project from Stanford. Both take historical data and apply it to a map, which ORBIS allows researchers to not only map the Roman Empire, but also calculate travel routes with respect to seasonal changes, modes of transportation, and expenses. It differs from the above examples in that it is not a tool to plot specific events, but it models various outcomes by using criteria set by the user based on historical data.

All of the tools covered in this week’s reading each combine history and geography to create customizable and interactive maps that can help us gain a new perspective on the historical record. This pairing has benefited from the digitization of documents and advances in/ increasing availability of GIS software and tools, and I can only see an increase in scholarship that utilizes such tools.

Visualizing Networks

This week’s readings on networks started off a bit confusing to me, but by the time I ended up at Weingart’s Networks Demystified series, I felt like I had learned the ins and outs of networks, more or less. I had never given much thought to the visualization of networks, nor how historians, humanists, or social scientists have been using them before, which may have explained my bewilderment with some of the week’s articles. I’ve come to understand that networks are basically connections between things, usually people. However, there can be many factors that play a part in these networks that we as historians should try to take into account.

One example, cited by several of this week’s authors, is John Snow’s cholera map showing how the 1854 outbreak began in London. John Theibault writes that Snow’s map presented a narrative, as well as analysis of the epidemic, and leaves it at that. Meanwhile, Johanna Drucker takes Snow’s map a bit further, putting into question just who all those dots were socially and demographically, as well as providing us first with a street map with plotted dots, and an updated version of the map that replaces the dots with actual humans. The human figures on the map help illustrate that each dot from Snow’s map represents a single individual, reminding us that there is more information than meets the eye in all data.

What has helped me understand the purpose of visualizing networks were Klein’s article on archival silence and data visualization in regard to Thomas Jefferson’s communication with James Hemings, who was Jefferson’s slave and chef, and the Mapping the Republic of Letters project, in particular the case study of Benjamin Franklin. Both utilize correspondence data to show patterns of communication. In Jefferson’s case, although he did not directly communicate with Hemings, the digital version of the Papers of Thomas Jefferson contains an editorial note about Hemings, as he was mentioned in his letters to other people. From this, the author was able to chart the frequency Hemings was mentioned and also in which correspondence he was referred to. This visual aid helps to show us how Jefferson communicated about Hemings, which would not be known if only relying on letters written directly to him, which were none.

More general patterns can be seen in Franklin’s letters, such as which country he was receiving letters from most during a particular time frame, what kinds of people he was corresponding with (professionals, artisans, etc), and his top correspondents. This approach helps answer questions about Franklin’s correspondence that might take large chunk of time to extrapolate, which is one of the benefits of visualizing networks.

Text Mining and Visualizing History

This week’s readings on the topic of text mining has helped me to understand a little bit more clearly just why historians might want to consider such an aspect of digital history. Text mining and topic modeling can both help reveal new patterns and themes about events, people, and documents that might otherwise have been overlooked. When information is presented in a visual map, whether it’s a chart or graph, or even word cloud, that information takes on a new perspective that researchers can choose to investigate more fully, although it is important to remember the context that surrounds the original data.

I’ve come to realize that I am very much a visual learner, which makes text mining and topic modeling quite interesting to me. By seeing the data maps that Cameron Blevins used in his article by using a program that measured the instances of geographic locations mentioned in two Houston newspapers during the 1830s/1840s and then later in the 1890s. The idea of “imagined geography” was new to me when I read the article and accompanying website, and I think it is aptly named. At the time when the newspapers’ articles, features, railroad schedules, etc were being written, I hardly doubt that anyone was thinking of all of the locations that were being referenced, nor their sociological/historical impacts.

Having read Ted Underwood’s tips for starting your own text mining project before diving into Blevin’s or Kaufman’s Quantifying Kissinger project, I had a basic understanding of the large amount of text needed for such undertakings. Underwood’s FAQ style post puts the idea of text mining into easy to understand concepts, which I found most helpful. When I saw the Wordle example, I though, “hey, I’ve done that before!” which put a personal connection to this week’s readings. Then I tried out Google’s ngram viewer, which is a cool tool for visualizing the usage of a word or phrases over time, but clearly the context is lacking, especially since we can’t see which texts are being searched.

Moving on from last week’s discussion about “buckets of words,” this week’s readings tell how we can take those buckets, do our keyword searching, and find out how those words stack up across time, for whatever it’s worth. While I think text mining definitely has a cool factor because of its employment of data maps, which I really find helpful, I need to remember that due to poor OCR and lack of context, text mining is just another tool to use in the grand historian toolbox.