Categories
Final Project

Adams Final Project Blog

Kyle Adams

Humanities 270 – Faull

May 2, 2018

Final Blog Post: Final Project Process

            As you have heard over the course of the semester, my final project involves an investigation of responses to the 2011 Penn State Child Abuse Scandal. What I have grown particularly interested in is using my newly acquired digital humanities competencies to drill down into the factors that played a role in conditioning discourse surrounding Sandusky and Penn State University’s crimes. For the final project, I chose to expand upon some of the original work I did this semester (Assignment 2) by utilizing new analytical tools (IBM Watson Natural Language Understanding tool for sentiment calculations) and looking at novel features that impacted responses to the scandal (for example, looking at response platform by adding fifty tweets to my corpus). My ultimate goal in working with this corpus and visualizing the impact time, geography, author gender, and response platform can have on discourse was outlined well in one of our early readings this semester, “…DH scholars who have successfully used visualization in their analyses … argue for the new hermeneutic of digital visualization, the way in which their visualizations produce both new knowledge and also invite ambiguity, the traditional province of humanistic critical thinking” (Faull 2). Upon completing my final project using Palladio and the WordPress platform, I believe I have accomplished each of these goals, as my visualizations not only clearly demonstrate that author gender, time, and response platform played some role in conditioning response sentiment and message, but also open the door for humanistic inquiry focused on the social dynamics and cultural mechanisms that allowed them to do so (show that there was an impact and welcome investigation into why these specific factors played a role).

            Throughout the production process for my final project, I ran into several challenges in working with the Stanford Design + Humanities Lab’s Palladio platform. Over the course of the semester, I have found that Palladio is a fantastic tool for research like mine, as it gives individuals the power to visualize many dimensions of a corpus. In my case, this prevented me from overwhelming the interpretive capacities of audiences by attempting to visualize the multiple dimensions I was interested in on a single plain. However, the versatility of Palladio did not entirely prohibit me from suffering from a common problem identified by Elijah Meeks: “Regardless, the first step is awareness of what a tool or method is dong and how it will inflect your research. I’m concerned that humanities scholars show a willingness to defer to tools…” (Meeks 2). In my case, this problem came about when I had to reformat my metadata in order to get the narrative I was trying to visualize to “fit” the software. Initially, my metadata was structured in such a way that each element of the corpus (article or tweet) had multiple “Key Themes” assigned to it in a singular row. However, because this did not work with Palladio’s “Graph” feature (as I could only represent one theme per article/tweet, therefore leaving many subsidiary themes absent from my network), I was required to duplicate entries in my metadata to accommodate the fact that I wanted to visualize more than just one “Key Theme” per article/tweet (so I went from 105 elements to over 300). Unfortunately, my “deferring” to the tool I was using caused problems when I attempted to visualize other dimensions (for example, “Gallery” or “Map”) as my duplicate entries led to redundant material in my map visualizations and catalog of corpus elements. The only solution I was able to come up with for this problem (inability to reconcile my desire to portray multiple themes and have a logical gallery and map view) was to separate my work into two separate Palladio “artifacts.” One of these would be used to generate visualizations in the “Graph” tab (networks, timelines, and tables) and the other would be used for “Gallery” and “Map” views.

Early iteration of corpus gallery which shows redundant entries due to duplication of rows in metadata sheet

In my work with the “Map” and “Graph” features of Palladio, I encountered another problem coming in the form of automatic summation of sentiment values. As can be seen in the screenshots accompanying this blog post, node size was biased by the fact that multiple articles from the corpus came from the same location or shared common entries in other columns of my metadata. Thus, my visualizations involving node size as a reflection of sentiment were biased, and would have to be tampered with in some way. In the case of the “Map” view, the only solution I was able to come up with was to set up my visualization so that hovering over artificially large nodes (representing places like New York City) would show the sentiment values for the multiple articles that had their scores added together. In the case of sentiment summation as it related to theme or gender, I chose to go about solving this problem by incorporating a new set of simple stacked bar charts into my analysis. In doing so, I made it so that I would only have to worry about the dimensions of gender and platform as they pertained to theme (did not have to size nodes).

Early visualization showing nodes sized according to net sentiment value of constituent elements from the corpus (bias emerges due to the underrepresentation of female responders and different number of times certain themes are measured)
One “Map” visualization used in the final project that demonstrates hovering over artificially large nodes to reveal constituent sentiment values
Sample stacked bar chart representing targeted IBM sentiment values for articles in the corpus

 

 

 

 

 

 

 

 

 

 

 

 

 

Once I had generated all of the necessary visualizations for the development of an argument that certain factors (like gender, time, and response platform) have had an impact on responses to the Sandusky scandal, I had to develop a logical fashion for presenting my visualizations and arguments as part of a narrative. In their article on narrative visualization, Segel and Heer explain the importance of narrative to digital humanities scholars by quoting Jonathan Harris, “‘I think people have begun to forget how powerful human stories are, exchanging their sense of empathy for a fetishistic fascination with data, networks, patterns, and total information … Really, the data is just part of the story. The human stuff is the main stuff, and the data should enrich it’” (Segel and Heer 1140). With this in mind, I realized that a multi-layered WordPress site consisting of several progressive magazine-style pages would be the most appropriate way to present my data. A site of this format allowed me to sequentially paint a picture of the real people who chose to respond to the Penn State scandal. Breaking the different dimensions of analysis into different pages helped me to contain my arguments and observations as palatable narrative bits that build to a broader understanding of the story of rhetorical conditioning in responses to the Sandusky scandal.

A final challenge in working with a project of this nature is that there is no available database that truly encompasses all of the tweets and articles written in response to the Sandusky scandal. Therefore, rather than work with a pre-compiled and “objectively complete” database, I had to gather the articles and tweets that would come to make up my corpus with no assistance. This style of corpus construction (during which I considered things like geographic location of the response) facilitated what I believe is the biggest flaw in my project (although it may have been an unavoidable pitfall): the subjective inflection of my corpus. In thinking about this, I was reminded of one of Johanna Drucker’s primary critiques of digital humanities work, “All data is capta, made, constructed, and produced, never given … So the first act of creating data, especially out of humanistic documents, in which ambiguity, complexity, and contradiction abound, is an act of interpretative reduction, even violence” (Drucker 249). In the case of my project, this data being “capta” led to two weaknesses in my final product — a disproportionate amount of responses from men and an uneven geographic distribution of articles (bias towards New York City and Washington, D.C.). Throughout my work, I was unable to find a realistic solution to this problem, so I elected to mention these flaws in my work under the “Conclusions and Further Work” menu on my WordPress site.

 

Chart showing overall sentiment values for responses written by men (the high number of bars compared to the female visualization shows potential gender bias of corpus)
Chart showing overall sentiment values for responses written by women (the low amount of bars compared to the male visualization shows potential gender bias of corpus)

 

 

 

 

 

 

 

 

 

In conclusion, I believe my final product (WordPress site) was successful from a digital humanities standpoint. I hold this belief primarily because my visualizations help to generate a novel way in which to view the discourse surrounding the Sandusky scandal. In this way, my project harkens back to the assertions of Tanya Clement:”Sometimes the view facilitated by digital tools generates the same data human beings (or humanists) could generate by hand … At other times, these vantage points are remarkably different from that which has been afforded within print culture and provide us with a new perspective on texts that continue to compel and surprise us by being so provocative and complex — so human” (Clement 12). All in all, the conclusions that can be drawn from the admittedly author-driven narrative produced by my WordPress site helps to turn simple metadata into evidence of very humanistic behavioral patterns and motivation. Therefore, my final project has manifested itself as an example of the value of digital humanistic scholarship.

Categories
Assignment 5 Uncategorized

Adams Assignment Five

     In quoting Ben Schneiderman, Isabel Meirelles opens her chapter on network design structures by articulating the positive attributes of these types of visualizations, “‘Social network analysis complements methods that focus more narrowly on individuals, adding a critical dimension that captures the connective tissue of societies and other complex interdependencies’” (Meirelles 47). Throughout my experience learning Gephi by using the Native American baptismal database, I have found this program to be incredibly helpful in painting this picture of “interdependencies” and relationships that is difficult to see by looking at metadata alone. Unfortunately, gaining a true understanding of the relational characteristics Gephi is capable of visualizing takes a certain amount of discipline in avoiding the mutual exclusion of visualization and analysis. While Gephi does not necessarily “hide” anything in terms of calculations, as users are offered an intimate look into what is being done when statistics are calculated or force-directed layouts are implemented in visualization, this opportunity to truly understand how data is being translated is easily ignored by individuals caught up in the “click-aha!” trap that is so common with digital visualization platforms (for example, when using partition tool to color and size nodes or manipulate edges). In this way, my experience with Gephi harkened back to Elijah Meeks’ work: “…I spent my time teaching folks how to use Gephi, and I tried to spend some time telling them that the network they create is the result of an interpretive act. I don’t think they cared, I think they just wanted to know how to make node sizes change dynamically in tandem with partition filters” (Meeks 2). This experience of Meeks’, which I perceive as an all-too-common one for those working in Gephi, also opens the door to some of Johanna Drucker’s skepticism, “So the first act of creating data, especially out of humanistic documents, in which ambiguity, complexity, and contradiction abound, is an act of interpretative reduction, even violence. Then, remediating these ‘data’ into a graphical form imposes a second round of interpretative activity, another translation” (Drucker 249). Simply put, by tying my short time with Gephi to the writing of Meeks and Drucker, I was able to arrive at one of my most unavoidable critiques of Gephi: that the platform, hard as it may try to avoid this, allows users to ignore the fact that their data has humanistic, nuanced, and narrative elements behind it (although the Data Laboratory tab is helpful in keeping individuals from being too far removed from their database to begin with). Unless individuals take the time to slow down and understand what is going on when different statistics are calculated or relationships are generated, the true power of Gephi is rendered almost useless.

 

For reference below: Edge color key and proportionality
For reference below: Node color key and proportionality

   

 

 

 

 

In terms of my visualization, I chose to generate a relational network consisting of Native Americans and baptizers (nodes). These nodes were connected by edges representing both baptismal and kinship relationships (for individuals labelled in the database with Unique ID’s 26-75). In total, my multimodal visualization, which utilizes a force-directed layout based on the Fruchterman-Reingold Algorithm, has 404 nodes and 438 edges (with most nodes representing Native Americans and most edges being classified as baptismal). Once I created these connections in Gephi, I elected to color both nodes and edges with nodes being colored according to an individual actor’s nation (or “baptizer”) and edges by the type of relationship represented (for example, baptismal, marital, parental, etc.). The statistics that I elected to run in order to analyze the baptismal database were Degree, Modularity, and Eigenvector Centrality.

Visualization with nodes sized by “Degree”

 

Degree:

By using the “ranking” tool in Gephi to size nodes according to degree (not in-degree or out-degree due to some edges being undirected) I was able to glean several important conclusions from the network. I immediately noticed that the nodes representing baptizers were the most impacted by sizing according to degree (more so than Native Americans). Considering baptismal relationships make up a significant portion of the connective tissue of this relational network, running a calculation for degree shows how influential individual baptizers were in spreading Christianity (baptizers with high degrees, like Camerhof, Christian Rauch, and Marin Mack were more influential in the spread of Christianity than those with lower degrees, like Grube or Utley). The degree calculation, which shows the number of connections a node has, also introduced me to Johannes, a fascinating character in the story of the spread of Christianity. The node representing Johannes was noticeably larger than those of other Native Americans when size was dependent upon degree. This is because Johannes was not only a Native American and Christian convert, but also a baptizer himself. Therefore, he effectively helped to spread Christianity through baptismal relationships, not just kinship ties (which was a characteristic that distinguished him from the other Native Americans in the database).

 

Visualization with nodes sized by “Eigenvector Centrality”

 

Eigenvector Centrality:

Unfortunately, I did not find this statistic to be terribly enlightening while I worked in Gephi. I believe this is likely because a majority of the edges I have in my database are directed, baptismal connections. For this reason, the only members of the database that have a real opportunity of having a high Eigenvector Centrality are baptizers that are connected indirectly through the limited kinship ties I was able to generate (as these would connect well-connected baptizers to one another). In my visualization, this resulted in Martin Mack and Cammerhof (along with the Native Americans whom they baptized) to have the highest measures of Eigenvector Centrality, as their “baptismal worlds” were the only two that were brought into contact with one another through the edges representing kinship ties.

 

Visualization with nodes sized by “Modularity”

 

Modularity:

After sizing nodes according to modularity, I was faced with another interesting iteration of my network diagram. As can be seen above, the modularity calculation helped to present several “small worlds” hovering around the outside of my force-directed graph. Although I was initially confused by this image, I soon came to the conclusion that these small worlds likely would not exist had I manually entered edges representing kinship relationships for all Native Americans in the database (beyond just 26-75). In my visualization, some people may have artificially high modularity for this very reason (because baptismal connections are present but kinship are absent). Essentially, I believe I have created a visualization that contains satellite baptismal communities absent of the familial ties that could effectively deflate the modularity statistic.

Following from this, the fact that modularity is relatively low amongst certain baptizers connected to Native Americans with kinship ties present also helps to show the tendency of different baptizers to work with members of singular families (as well as cross national boundaries). This is due to the fact that multiple baptizers working with members of singular families (for example, spouses being baptized by different people) helped to generate a highly interconnected Christian network of Native Americans and baptizers and effectively eliminated the presence of “small worlds” in certain areas of the network (baptizers connect different families and limit isolation in the network).

Classifying edges proved to be extremely helpful in arriving at this inference, as it demonstrated that individual baptizers likely did not share intimate connections with specific Native American families. This is shown by the colored edges themselves, as kinship relationships visually represent connections between “small baptismal worlds.” (for example, spouses, brothers, sisters, parents, and children appear to be rarely welcomed into the Church by the same baptizer).

Categories
Uncategorized

Time Visualization to Help Project

Matthew Bloch, Lee Byron, Shan Carter, and Amanda Cox (The New York Times) – “The Ebb and Flow of Movies: Box Office Receipts 1986-2008

 

I chose this stacked flow visualization regarding ticket sales because I think it would be rather illustrative in the context of my data. I believe I could select one or more journalistic publications and collect information regarding the volume of articles written about Penn State over the course of several months, beginning in November 2011. In order to separate out layers as shown above, I could potentially differentiate articles by subject (ex: Sandusky, Paterno, football program, university, etc.) and see how volume ebbs and flows over time. I believe this visualization would help to depict the nuance of the narrative of scandal I am looking to depict. For example, I believe certain subjects will have their volume increase at certain points, showing turning points in the scandal (ex: Paterno’s volume will likely increase around the time of his death or the publication of the Freeh report).

 

 

Florence Nightingale – “Diagram of the Causes of Mortality in the Army in the East”

 

Florence Nightingale’s radial diagram could also be useful in visualizing the time dimension of my data (article publication date) as the pedals can expand to show volume, but also be broken up into actors, similarly to the layers of a stream graph. In addition, the circular layout of the visualization implies a repetitive aspect that would be interesting to incorporate should I choose to compile volume data for more than a particular year.

Categories
Assignment 3

Adams Assignment Three

Palladio Gallery view (unfortunately image URL’s were not behaving)

For this assignment, I continued my investigation of journalistic sources involving the firing of Joe Paterno at Penn State University in 2011. In addition to the university newspaper sources I analyzed for our previous assignment, I added several non-collegiate journalistic sources to my corpus (articles from The New York Times, The Chicago Tribune, The Seattle Times, The Los Angeles Times, The Denver Post, and The Dallas Morning News). In terms of metadata, I chose to classify my data into the following categories: institution, institution location, institution geolocation (lat,long), institution connection to Penn State, article author, author gender, article title, article word count, date of publication, article classification upon original publication (ex: ‘Opinions’), and a quantitative sentiment valuation for the article (taken from Jigsaw and altered by adding one to the value given in Jigsaw to put all scores in a numerical set ranging from one to two). While my metadata certainly contained values for many constituent aspects of my articles, I believe each could be visualized in an epistemological way. When it comes to sentiment and word count (my two numerical categories), each of these facets could be visualized in such a way that aids in generating arguments regarding quantity and quality of discourse surrounding Joe Paterno’s firing. In addition, these values could be paired with author gender or geolocation to determine which areas (in general, as there is still subjective interference in article selection) or genders are speaking more or less and more or less pleasantly about the event. Although publication date did not turn out being too helpful for my data visualization, a larger corpus may foster some sort of conclusions tying article publication (denoting temporal proximity to the actual firing) to geographic location, institution classification, or institution connection to Penn State. I also believe it is important to keep article title and author name present in certain visualizations, as these things can both become pertinent to discursive analysis (understand particular author bias if a name is familiar and get a sense of article tone by looking at title).

 

Palladio Table view showing multiple layers of metadata

In both the Table and Gallery views in Palladio (Left: Table View, Above: Gallery View) I found the most important function of the platform to be keeping researchers from being too far divorced from the metadata and corpus in question. One of Johanna Drucker’s most poignant comments from her chapter mentions that “almost all information visualizations are reifications of mis-information” (Drucker 245). In my opinion, these two windows in Palladio prevent this from happening on some level. While they do not necessary alleviate concerns regarding data as capta (constructed and gathered by biased subjects), it does allow for those who engage with a corpus and set of metadata in the platform to see for themselves exactly where visualizations are coming from. In the Table view, this is accomplished by presenting metadata exactly as it was uploaded. The Gallery window (when linked to URL’s) not only does the same thing, but can allow (at least in my case) for researchers to return to the component articles/documents of a corpus with the click of a button. Essentially, each of these views in Palladio helps to increase familiarity with a data set/corpus for the benefit of better understanding visualizations in other views.

 

 

Palladio Graph view visualizing article titles authored by women (and sentiment value in node size)
Palladio Graph view visualizing article titles authored by men (and sentiment value in node size)

 

 

 

 

 

 

In the Graph view (Pictured Twice Above), I chose to visualize connections between author gender (M, F, and N/A in the case of collective publication) and article title with node sizing governed by Jigsaw sentiment score. While node sizes were not remarkably illustrative in an argumentative, epistemological, or hermeneutic sense, this could potentially be attributed to the small range of achievable sentiment scores keeping nodes at similar sizes. In addition, because the metadata being visualized comes from a journalistic corpus, it is understandable that wild swings in sentiment from one article to the next are not necessarily evident in the visualization (most articles must conform to a customary journalistic prosaic style). However, when node sizes are combined with the sentiment conclusions rooted in article title (as this conveys tone), one can use these visualizations to begin to discern a connection between gender and article “feel.” It appears to me that female authors, based on article title in particular, were more willing to tie the Paterno firing back to Sandusky’s sexual abuse explicitly, whereas male authors were more likely to use their platform to honor Paterno or frame the event through a less criminal lens.

 

 

Palladio Map view showing institution geolocation and sentiment value

I also found the Map view (Above) to be quite interesting in terms article sentiment and publishing institution geolocation. When I began to construct this corpus, I theorized that article sentiment would become more negative as proximity to Penn State decreased. However, according to the map, it seems that some of the highest Jigsaw sentiment scores (denoting more “positive” vocabulary in the source document) were found further from State College on the map (except in the southern United States). This could have had something to do with closer, Big Ten affiliated, schools being more saddened by the symbolic loss of one of their conference’s most recognizable icons. It is also important to remember when looking at this visualization that the documents in question are journalistic, and, for that reason, may not be the most perfect texts to analyze in terms of sentiment (although some articles are opinion pieces).

 

All in all, I believe my visualizations presented above can be perceived as flawed in some sense, but also intellectually constructive in another. Importantly, Palladio’s provision of the Table and Gallery views is helpful for preventing researchers from being barred from a conception of the “translation” process that occurs in the visualizations in the Graph and Map windows. That being said, I believe my visualizations, at least in terms of sentiment analysis, are hindered by forces of concealment and reduction that come along with an imperfect (and hidden) Jigsaw sentiment computation (in some sense these visualizations can be seen as simple representations or mis-representations of sentiment). However, what I do believe my visualizations were successful in doing is putting on display some of the ambiguity or more nuanced aspects of my corpus. For this reason, I believe some of my visualizations would be considered successful knowledge creators to Drucker, for their “denaturalizing” effect on viewers. These visualizations are certainly novel ways of interrogating journalistic discourse surrounding Joe Paterno’s firing, and thus can help to serve as stepping stones for new understanding.

When it comes to Drucker’s discussion of spatialization creating meaning, I believe the map view can be used as a tool for knowledge creation. Due to the fact that article sentiment is tethered to proximity to the focal event, one can begin to create arguments regarding geographic location governing attitudes toward the Paterno firing in local newspapers. Article titles could also be added to these visualizations (found by hovering over a point) to help with this aforementioned pattern recognition.

Categories
Assignment 2

Adams Assignment Two

Working with Jigsaw and Voyant

a.) After accepting the fact that corpus construction and visualization was, indeed, an iterative process (and that I would have to add nuance to my corpus throughout the semester), I decided to begin collecting journalistic articles from university newspapers regarding the firing of Joe Paterno. Although it appears that each article does not necessarily focus on Paterno’s firing, I decided it would be interesting to incorporate each publication’s first mention of the event (whether that was the overt focus of the article or not). I chose to gather articles from institutions that are connected to Penn State University either by proximity (UPenn, Bucknell, Pitt), athletic conference affiliation (Ohio State, Michigan, etc.), or level of football prestige (Oklahoma, Georgia, etc.). As my research with this corpus continues, I hope to remain cognizant of the subjectivity lurking behind article selection. My role as an archival author is just one factor that allows for the visualizations produced from my corpus to remain perspectival (simply by adding technology into the equation, essentialist understandings of this journalistic prose cannot be reached).

 

b.) Throughout my time working with the Voyant platform, I have appreciated it not only for its ability to foster more intimate understanding of meaningful entities, but also provide individuals with the opportunity to easily generate comparative research questions at the document level. In the two Voyant visualizations embedded (one being a Bubblelines visualization and the other a StreamGraph) one can begin to analyze occurrences of key entities as they relate to one another within specific documents and in the corpus as a whole. In particular, the raw word frequencies displayed in the StreamGraph visualization allowed me to to understand how different university news outlets framed the Paterno firing (ex: some did not even mention the board responsible for his firing and others). Two other interesting conclusions that can be drawn from the visualizations are that the victims of Sandusky’s abuse (the cover-up of which led to Paterno’s demise) rarely appear in this journalistic prose, but also that external campus news outlets (non-State College) do not shy away from connecting Paterno and Sandusky, leading to an implicit association shrouded in negativity.

 

c.) One of the most meaningful visualizations I was able to produce using the Jigsaw platform was the WordTree displaying appearances of “Paterno” throughout my corpus. By expanding my capacity for entity contextualization, this tool afforded me the opportunity to begin to discern community level sentiment displayed in nominal data consisting of journalistic prose of publications external to PSU. Meirelles effectively summarizes the value of a WordTree visualization for analysis in her sixth chapter, “Besides preserving the context in which the term occurs, the method also preserves the linear arrangement of the text” (Meirelles 200). While functioning in this way, the WordTree visualization manifests itself as a more intermediate step in distancing oneself from a text. By quickly allowing me to understand the context in which JoePa’s name showed up, the Jigsaw platform helped me to see that university newspapers outside of State College did not shy away from criticizing Joe Paterno for his role in the Sandusky scandal, but also felt the need to show appreciation for his accomplishments as a coach (things that may aid in blinding the Penn State community to his wrongdoing). The image of Jigsaw’s Document View exemplifies the platform’s ability to further remove researchers from their corpus and garner an understanding of entity frequency and classification all while summarizing documents (although the selection of these statements by the program may also be an example of non-objectivity in data provision).

 

d.) In terms of commonalities, each of these programs has the ability to generate meaningful visualizations that allow for the contextualization of entities (ex: seeing connected words, etc.) and understanding of word frequency (which can be extrapolated to fit perspectival analysis). Obviously, each platform also possesses characteristics which its counterpart lacks. For example, Jigsaw can easily generate visualization of sentiment while Voyant offers remarkably unique and varying visualizations of word frequency (ex: bubble sizes, word sizes, stream width). It has also become rather apparent to me that individuals choosing to work in the Jigsaw platform must have a more intimate understanding of their corpus in order to create visualizations, as more “work” must be done to produce meaningful images. However, once one is familiar with the platform, I am under the impression that Jigsaw allows for more advanced visualization of entity context and document connection. All in all, Voyant provided me a more preferable method for discerning patterns regarding word frequency (allowing me to address sentiment and entity relationships). Jigsaw, on the other hand, allowed me to produce detailed WordTree visualizations that worked as a sort of additional approach in bolstering conclusions regarding outsider perception of Paterno (bigger picture as opposed to entity-level).

 

e.) Simply put, by engaging with both my corpus and the textual analysis platforms Jigsaw and Voyant, I have learned there are many opportunities for argumentation and subjective involvement in digital humanities research. In this way, I have learned to appreciate the role of technology in expanding the capabilities of nominal data analysis, but also understand that the implementation of digital methods into digital humanities does not generate essentialist interpretations (I now actually see where I could have gone in different interpretive directions). As Clement puts it, “Sometimes the view facilitated by digital tools generates the same data human beings (or humanists) could generate by hand…At other times, these vantage points are remarkably different from that which has been afforded within print culture and provide us with a new perspective on texts that continue to compel and surprise us by being so provocative and complex — so human” (Clement 12). What can be gleaned from this quotation, and my work on this assignment is that digital humanities does, in fact, involve a give and take between subjective and objective construction. By utilizing differential reading strategies and accepting the interaction of close and distant reading, researchers such as myself will be properly equipped to read and contextualize readily accessible and more distant digital findings (and acknowledge subjective intervention in text selection, platform construction, etc.).

 

Voyant StreamGraph Visualization

 

Voyant Bubblelines Visualization

 

Jigsaw WordTree Visualization

 

Jigsaw Document View Visualization
Categories
Assignment 1

Kyle Adams Assignment One

Comparison of “The Genealogy of Pop/Rock Music” and “IBM Watson News Explorer”

I chose these visualizations for their collaborative function as exemplar of the fact that networks can manifest themselves in remarkably different ways. When analyzed in tandem, “The Genealogy of Pop/Rock Music” and the “IBM Watson News Explorer” exemplify several distinctions between static and dynamic visualization. As a dynamic visualization, the latter allows individuals to interact with a massive news/media database that simultaneously sorts articles by location, keyword, actors, etc. By giving those engaging with the platform the freedom to “play” with the visualization of sources (ex: alter how they are presented and what is presented) as well as updating information in real time, this IBM interface demonstrates characteristic elements of dynamic visualization that the more static (cannot be altered or “played with” but does include a temporal dimension) musical genealogy does not. Both visualizations, by allowing for the studying of patterns of connections between elements (musicians, news sources, or whatever else) allow data to be viewed in novel ways. However, only the news explorer provides individuals with the opportunity to actively interact with the data in multiple ways (must also keep in mind that the genealogy’s multiple dimensions allow for pluralistic interpretive interaction with data). Although essentially static, the musical genealogy does serve a heuristic function for its ability to depict artistic longevity and influence that may not immediately be apparent to researchers (operates similarly to Priestly’s timeline). The IBM platform utilizes more dynamic qualities to create new knowledge about the interconnectedness of major global figures, organizations, etc. (and even can be used as a way to interrogate the media as an institution). In either case, the visualization does a magnificent job of contributing to novel ways of understanding their fields and many others (understand both relational and attribute data).

 

Reebee Garofalo’s “The Genealogy of Pop/Rock Music”
Screen Capture of IBM Watson News Explorer

 

 

 

 

 

 

 

 

 

 

 

 

Analysis of “Six Degrees of Francis Bacon”

I selected this visualization based on its direct ties community network designs discussed by Meirelles and Lima. This network visualization entitled “Six Degrees of Francis Bacon” utilizes dynamic characteristics such as a mutable structure and variable presentations of biographical and connective information that can be accessed at the discretion of users (both things static visualizations do not necessarily present). More specifically, this visualization allows users to both interpretively interact with the data as well as more actively mutate the visualization (can be radial, force-directed, etc.). By visualizing myriad informational dimensions and including analytical ambiguity, this community network succeeds in allowing individuals to view the data from different perspectives and arrive at unique and valuable humanistic conclusions. “Six Degrees of Francis Bacon,” much like other community networks we have encountered, effectively contributes to new understanding of intellectual, political, and general social influence in England (shows connections, timelines, biographical information). If this particular visualization did not exist, it would seem nearly impossible to readily generate conclusions regarding the impact or magnitude of this particular social network tethered to its central node, Francis Bacon.

 

Screen Capture of “Six Degrees of Francis Bacon” Network Visualization

 

 

 

 

 

 

Categories
practice Uncategorized

Two Bad Visualizations

 

Principle of Graphical Integrity Violated:

  1. Representation of numbers should be directly proportional to numerical quantities
    1. This visualization utilizes a “lie factor” significantly higher than one to emphasize statistical imbalance (due to teams using five starters, it may also be more informative to show how Arkansas’ additional forty minutes of action was divided amongst its starters).
    2. Potentially utilizing “lie factor” to generate a justification for a predicted outcome.

 

Principle of Graphical Integrity Violated and Other Concerns:

  1. Clear, detailed, and thorough labeling should defeat graphical distortion and ambiguity
    1. Obviously, data is represented in an ambiguously relative fashion (with actual numbers being hidden)
  2. Difficult to say whether or not other principles are violated due to the lack of data labels or a key (cannot say whether or not data has been classified properly, etc.)
    1. May not even be an appropriate visualization based on data being represented, as shading is not necessarily always the best way to visualize singular quantities.