Abstract
In the Digital Humanities, there is a fast-growing body of research that uses
data visualization to explore the structures of language. While new techniques
are proliferating they still fall short of offering whole language
experimentation. We provide a mathematical technique that maps words and symbols
to ordered unique numerical values, showing that this mapping is one-to-one and
onto. We demonstrate this technique through linear, planar, and volumetric
visualizations of data sets as large as the Oxford English Dictionary and as
small as a single poem. The visualizations of this space have been designed to
engage the viewer in the analogic practice of comparison already in use by
literary critics but on a scale inaccessible by other means. We studied our
visualization with expert participants from many fields including English
studies, Information Visualization, Human-Computer Interaction, and Computer
Graphics. We present our findings from this study and discuss both the
criticisms and validations of our approach.
1 INTRODUCTION
One of the goals of the literary critic is to analyze language and its embedded
complexity. For example, when literary critics examine a poem’s form, they
consider many characteristics of the words it contains, including the
similarities and differences in orthography, sound, the visible pattern it
produces, rhythmic structure, and countless others. Nearly all of this
information is available through visual inspection of the poem and contained in
what may already be our greatest visualization technique — the written word.
However, this same inspection carries with it the biases introduced by the
semantic meanings of the words themselves: it is difficult to pay attention to
the structural parts of the word “apple” without imagining the fruit it
represents.
One possible method of aiding in the process of literary criticism is to provide
an alternate representation [
Simon 1988] of the words contained
within the text to help separate the meaning from the structure, or to provide a
different vantage point with which to examine the work. By providing this
alternate view, structural aspects of the poem that may be difficult to
recognize when obfuscated by the meaning of the words themselves may be easier
to analyze and observe. We caution that this understanding will likely not be
valuable unless a connection can be bridged between the alternate representation
and the original piece of literature, and we stress that this approach is to be
considered an augmentation to existing practices. The goal of the research
presented in this paper is to provide an information visualization that provides
alternate representations of literature that will allow the critic to recover
the original work for analogic purposes.
There have been a variety of examples in the digital humanities and information
visualization disciplines that provide alternate representations of language,
such as Word Clouds [
Viegas 2008] and Word Trees [
Wattenberg 2008], but the authors are not aware of any examples
that have as a priority the ability to discern the text itself from the
visualization. Instead, these examples provide a means of examining text with
prior knowledge of the contents therein. Nonetheless, it is an open question
whether this change in representation would prove useful to the trained literary
critic.
In this paper, we present a design study of an information visualization that is
a recoverable representation of language. Specifically, we present the design of
a visualization we have named Language DNA (L-DNA) that visually encodes any
symbol system, in our case the letters and phonemes of words in the English
language, and a qualitative evaluation with several literary critics and
designers to investigate the need and use of such a visualization.
2 RELATED WORK
Our study is an examination of how to visualize language in ways that can build
ontologies of words based on the needs of the literary critic. Much work has
been done in the area of text visualization but none have yet approached a level
that can produce whole language interactions. Our work builds on previous text
visualization techniques while expanding on the scope of information that can be
shown in our system, Language DNA.
2.2 Visualizing Language in the Digital Humanities
Within the domain of digital humanities, there are three main streams of
visualization that tend to pervade the literature. All three types are based
in text analysis but approach the problem from different directions. The
first is the GIS type tools for organizing spatial data in the humanities
[
Jessop 2006]
[
Jessop 2008]
[
Wood 1992], specific examples of this type of work include
Bingenheimer et al.’s visualization of the biographies of Buddhist monks
[
Bingenheimer 2009], as well as Valley of the Shadow [
Valley 2015], in which there are examples of multiple types of
quantitative and qualitative data presented in a GIS format.
The second category involves tools used to augment reading, text analysis
tools that highlight relationships in texts, often with visualizations. This
category includes work such as Clement’s distant reading of Gertrude Stein
[
Clement 2008], and Wattenberg et al.’s work on using tree
visualizations for visual concordances [
Wattenberg 2008].
Projects such as TAPoR [
Tapor 2015] also provide data mining
tools for literary and textual analyses. The aim of these projects is to
allow for interesting portions of large texts to be marked up and visualized
for comparison.
The third category of research is the straight visualization projects which
are of two types: The first are the projects that use toolkits such as
Voyant tools [
Voyant 2015] to accomplish their research task
and the second are the projects that create bespoke visualizations specific
to their research problem. The second type of these includes Moretti’s work
on visualizing interactions and character movements in literature [
Moretti 2005], Bingerheimer et al.’s social network
visualization from TEI data, which uses text encoded data to produce
networked relationships among textual elements [
Bingenheimer 2006], and Gould’s use of network tools for
historical data analysis [
Gould 2003]. This can also be seen
in the TextArc project, which uses radial graphs to show contents and
relationships between words in texts [
Paley 2002], and Writing
without Words [
Posavec 2015], Posavec’s beautiful graphs of
sentence structures and themes in Kerouac’s “On the Road”. Perhaps the most
exciting and cognate study is Alexander’s work on the Oxford Historical
Thesaurus and mapping sonic relationships in language over time [
Alexander 2012]. Also, we find a movement to projects that use
art to approach these problems, highlighting that aesthetics may be itself a
critical tool. Installations such as TextRain [
TextRain 2015],
an interactive art piece where gallery goers interact with falling text, and
Word Collider [
WordCollider 2015], an artistic project modeled
after pictures from CERN’s Large Hadron Collider, where parts of words when
“smashed” together produce effects like those seen in images from
particle accelerators.
The development of our technique was motivated by the fact that existing
visualizations and text analysis tools, while usually aesthetic pleasing,
cannot easily be used for the types of analyses expected or desirable for
literary criticism or linguistics. Our Language DNA is an attempt to build a
system that can handle multiple levels of information to display complex
structural and content-related relationships within texts.
3 MOTIVATION — A LITERARY CRITIC’S PERSPECTIVE
In his book
Anatomy of Criticism, Northrop Frye asks
the question: “what if criticism is a science as well as an
art?”
[
Frye 2000] The development of the visualization technique presented here was driven
by this exact question. Our intention is to facilitate a different approach to
Frye’s domain-specific question about literary criticism: can we use
mathematical and visualization techniques to incorporate science into a literary
criticism? These questions have a long history in the humanities and were born
out of the general notion that arose in the renaissance when the likes of Locke,
Bacon, and Descartes defined a practice that separated and solidified science
from the humanities. The 20th century structuralists suggest that this position
is reconcilable, that all literature and language is systematic. This idea —
that language is systematic — is approachable via visualization.
The visualization algorithm we present can accommodate as much or as little
information that a critic could want, giving the possibility of visualizing as
little as a single letter, as much as entire corpora. If we are to imagine a
space where the literary critic or linguist can experiment using a
language-based mathematics, it must have these three characteristics:
3.1 Consistency and Reversibility
First, a visualization algorithm is needed to encode language such that we
can create a space that is both consistent and reversible. In mathematical
terms, this would be referred to as one-to-one and onto. The need for
consistency and reversibility arises from the requirement in the analysis
process of preserving the ability for human interpretation of the words. In
order for the visualization to be “readable” by a critic, each word must be
consistently mapped and that mapping must be reversible into the original
work. Most existing visualization techniques distort the original texts
without providing an avenue for reconstituting them. This is problematic
when studying things like poetics where the spatial component of the text is
integral to its meaning.
3.2 Infinite Plotting Space
Second, it is important that the algorithm uses a plotting space that is
infinite. Imagine that we were to approach the problem of metaphor. As an
example, in theory we do not know if the chain of meaning created by
metaphoric relations between words is finite. It stands to reason then, that
without an understanding of the full requirements of a system that an
infinite space is a safe decision. If we are to start to approach these
types of questions our lack of knowledge should not limit the size of
possibilities. Specifically, the critic must be able to analyze words and
literature that are perhaps not known to the visualization designer. With
our technique, we can encode anywhere from one letter to an entire language,
to the entire literatures of one language, and even multiple languages, in a
space where each point belongs to one individual piece of original
information. The infinite space means that in creating “experiments”
the literary critic is not limited to our present understanding of
language.
3.3 Layering
The third requirement is the need for the ability to layer symbols in order
to make comparisons. For instance, it should be possible to overlay a poem
within the context of the entire language or other poetry. The need for
layering arises from the analogic basis of most types of literary critical
and linguistic inquiries. This should extend to any types of symbols, as
comparisons are not always rooted in the Latin alphabet. Based on what we
recognize as the possible requirements of such a system, we designed our
Language DNA visualization with the three characteristics of consistency and
reversibility, infinite plotting space, and layering in mind.
4 DEFINING LANGUAGE DNA
An important property of our mapping is that the words be recoverable from the
visual space. Mathematically, this requires that the mapping be a bijection
(i.e., that words both map to a unique place in the visual space, and that each
point in the visual space maps back to a word). As an example of the technique
we introduce a mathematical translation of words to numbers that relies on the
lexicographical ordering of letters. This is essentially a mapping of
alphabetical order and is one of possibly infinite ways to group the data. We
have chosen this technique as a first demonstration because we are all familiar
with the way we order a dictionary, but we must stress that we can map the data
many different ways. We define the mapping 𝑔 so that each letter is
mapped to its position in the alphabet, as follows:
𝑔 ∶ 𝐴 → ℤ
where 𝐴 is the set of alphabetical characters {𝑎, 𝑏 … ,
𝑧} and:
𝑔(𝑎) = 1, 𝑔(𝑏) = 2, … ,
𝑔(𝑧) = 26
Note that this mapping is currently written using base 10 numbers for the
integers (1 to 26, with an implied 0 for no character), but our mapping requires
a base 27 representation (or more generally base N+1, where N is the number of
characters in the language), which for convenience we will symbolically
represent as follows:
110 = 𝑎27 (i.e., 1
in base 10 is represented as ‘a’ in base 27)
210 = 𝑏27
...
2610 = 𝑧27
Thus, we can define our mapping of words to a one-dimensional number line as
follows:
𝑓: 𝑊 → (0,1)
where 𝑊 is the set of alphabetical words (e.g.,“apple”, “dog”,
“the”, etc.) such that for each 𝑤 𝜖 𝑊,
𝑤 = 𝑥1 𝑥2 ... x𝑛,
and:
𝑓 (𝑤) = 0.(𝑥1)𝑔(𝑥2) ...
𝑔(𝑥𝑛)
For example, for 𝑤 = “dog”, we have x1 = ‘d’, 𝑥2 = ‘o’,
𝑥3 = ‘g’, therefore:
𝑓 ("dog") = 0.𝑑𝑜𝑔
Note that this is a base 27 number, but could be converted to base 10:
0.𝑑𝑜𝑔27 = 4x27-1 + 15x27-2 + 7x27-3 = 0.169079910
If we relax the restriction that each word needs to end (i.e., we allow words to
have an infinite sequence of letters), it becomes clear that 𝑓 is a bijection,
since every word generates a unique base 27 representation (one-to-one: the
property that if two words map to the same number, they must be the same word)
and each number between 0 and 1 can be converted to base 27 to recover the
sequence of letters (onto: the property that every number has a word that can
map onto it).
4.2 Using the 2D Visual Space
The mapping above describes how an arbitrary word can be mapped onto a number
line, which already allows the visual mapping of words onto an axis in 1D
space (similar to a lexicographical axis). Here, we describe a method,
inspired by Cantor’s Diagonalization to map an individual word onto 2D space
directly (and more generally onto n-dimensional space).
We can split the word in two by considering every other character, for
instance InFoViS would become IFVS and
noi. Thus, we can take the base 27 representation of the
mapped word and create two dimensions as follows:
𝑓2:𝑊 →
(0,1)2
𝑓2(𝑤) = 𝑓2(𝑥1𝑥2 ...
𝑥2𝑛) = (𝑓(𝑥1𝑥3 ...
𝑥2𝑛-1) ,
(𝑓(𝑥2𝑥4 ... 𝑥2𝑛))
For example, if our word is “applesauce” (Fig. 3):
𝑓2 ("applesauce") =
(0.𝑎𝑝𝑒𝑎𝑐,
0.𝑝𝑙𝑠𝑢𝑒)
which, in base 10 would be:
𝑓2 ("applesauce") =
(0.0592410,0.6100587)
This mapping can easily be extended to n-dimensions by taking every nth character in the base
27 representation of (𝑤). f2 is also clearly a bijection, because every word
can be split into alternating characters to generate two base 27
representations (one-to-one) and each pair of numbers between 0 and 1 can be
converted to base 27 to recover the two parts of the word, which can then be
reassembled (onto). Thus, every word in the English language can be mapped
onto a 2D space using f2,
and every 2D point can be mapped to a “word”, where a word is a
sequence of possibly infinite letters which may well not have associated
semantics. Note that this mapping does not account for things like homonyms,
but with a simple addition to the mapping we could easily differentiate
words for any number of their ontological characteristics.
4.3 A Note on Scale
Since whole natural languages are immense, it is important to discuss scale,
both of what is being visualized and the size of the resulting
visualization. We can base a visualization size calculation on the number of
words being visualized, and then determine the length of a 1D L-DNA
visualization that draws at a density of a single pixel for each word or
unit (it is important to remember that these calculations are for
orthography, they will change depending on the symbol system used). Two
measures are needed to accomplish this: the smallest and the largest
distance between two words. Since our algorithm already normalizes words in
1D to be between 0 and 1, we can assume that the difference between the
largest and smallest words is approximately 1.0 (with the words ‘a’ and
‘zygote’, this is already correct to 1 decimal place). In our analysis of
words from the Oxford English Dictionary (OED), the two closest words using
our algorithm are “abandoner” and “abandoning”, with the first seven letters
in common and the next being very close in the alphabet. The difference in
values from our algorithm for these two words is:
0.abandoning27 - 0.abandoner27 = 1.37 x 10-1110
Thus, to present a number line from 0 to 1 with numbers only 1.37×10-11 apart represented as different pixels would
require:
1.0 ÷ (1.37 x 10-11) = 73.1 billion
pixels
Note that in 2D, our algorithm fairs far better. This same pair of words
would be broken down into two pairs of coordinates:
(0.aadnn27, 0.bnoig27) and (0.aadnr27, 0.bnoe27)
which has at most four letters in common for each dimension and would require
only:
1.0 ÷ (0.aadnr27 -
0.aadnn27) = 1.0 ÷
(2.79x10-7) = 3.6 million pixels
To put this into perspective, a 1D visualization using our algorithm would
require the width of 38.1 million 1080p screens (1920 × 1080 pixels) placed
side-by-side, and a 2D visualization would require 6.2 million 1080p screens
arranged to form a rectangle.
5 ILLUSTRATING L-DNA
We start with three examples to demonstrate how L-DNA can be used to visualize
language. The first is a dictionary mapping for the English language, the second
is a view of multiple languages, and the third is a mapping of English phonemes
to illustrate the applicability of this approach to any set of symbols
[2].
5.1 Visualizing the Oxford English Dictionary
Fig. 4 shows all 370,624 words parsed with criteria that eliminate diacritics
and punctuation the Oxford English Dictionary
[3] rendered using the algorithm described above
(i.e., using the coordinates provided by
f2. The result is a mapping that privileges the
first two letters of each word. That is, the x-axis can be read as an
alphabetical ordering of the first letters of words, and the y-axis can be
read as an alphabetical ordering of the second letters. This makes the top
left box “AA”, where you would find words such as “Aardvark” (note
that few words in English begin with two A’s, which is why this box is quite
sparse). This property is recursive, so that within each box, the third and
fourth letters are similarly privileged. For example, in the “BA” box,
there is an “NA” box, which has another “NA” box that contains the
word “BANANA”. This initial visualization shows how we can start to
understand where each word belongs within the 2d whole.
Because our algorithm privileges the spelling of words, this 2D
representation can be thought of as a form of 2D orthography (specifically
spelling rules). It is essentially a two-dimensional layout of alphabetical
order. This version of L-DNA reveals a “bird’s-eye view” of the
language that was not previously available to the literary critic, linguist,
or lexicographer; a critic could previously flip through a dictionary’s
pages or even a list of ordered English words, but this visualization
instead provides a new 2d spatial location for each word in this
dictionary.
5.2 The Multiple Language L-DNA View
Our second example compares multiple languages (English, French, German, and
Spanish). Fig. 5 shows these four languages each represented in 1D on the 0
to 1 number line using our algorithm, stacked for comparison. Visual
inspection reveals a similar sparseness in the ‘Q’ portion of the line for
all languages (i.e., few words in any of these languages begin with ‘Q’ and
any letter other than ‘U’), but additional sparseness in French, German, and
Spanish exists near the end of the alphabet (‘W’, ‘X’, ‘Y’).
Fig. 6 also shows a side-by-side comparison of multiple languages in 2D, and
Fig. 7 overlays these four languages. Fig. 8 shows a close-up of the AL
region of the overlaid image. These side-by-side comparisons or overlays
allow for elementary analogic comparisons and can be expanded on with more
complex symbol encoding.
The above images were generated with the constraint that we only had access
to open source dictionaries [
Dict 2015] for languages other
than English (for which we have university-wide access to the OED). The
French (red) has 197,954 words, the English (blue) has 370,624 words, the
German (yellow) 425,501 words, and the Spanish (green) has 160,442
words.
5.3 English Phonemes in L-DNA
We chose the next example (Fig. 9), English phonemes, to demonstrate the
robustness of the technique to arbitrary symbolic representations of
language, and to create an analogue between the spellings of words and the
sounds of words. The mapping is organized in like sound units: vowels (e.g.,
“AA, AE”), semivowels (e.g., “W,Y”), stops (e.g.,
“B,D,K”), affricates (e.g., “C, H, JH”), fricatives (e.g., “D,
SH, V”), aspirates (e.g., “HH”), liquids (e.g., “L, R”),
and nasals (e.g., “M, N, NG”).
By organizing words into phonemes, some interesting observations can be made.
It appears clear that a portion of the phonemes are used primarily for the
first syllable and another distinct set is used primarily for the second
syllable. This can be observed through the densely populated top and right
columns, with the majority of the bottom-left part of the image containing
almost no English words. In addition, the top-right corner is mostly empty,
with the exception of a few very dense groups, representing the few phonemes
that are used for both the first and second syllables.
5.4 Poetry
The final example that we created was to insert a single poem into the space
that we created for English words and phonemes. This is a first step in
being able to use these spaces for analogic comparisons.
Some interesting patterns can be observed in the poem through visualizing it
in this manner. Firstly, in terms of orthography (Fig. 10) it becomes
possible to identify visual rhymes by cluster groups within the image. In
Fig. 11, this same phoneme visualization can be used to identify rhyming
patterns within a poem. As the phonemes group together it is possible to see
the types of sounds being repeated in the piece. Although this is easy to do
with a 16 line sonnet, it becomes much more difficult with a poem of any
significant length (e.g. Milton’s 10,000 + lines of verse in “Paradise Lost”) and this technique could help to
highlight “macro-structures” in poetry. Each diagram below is laid out
in two dimensions. This is a decision made during the encoding process and
can be n-dimensional based on the amount of information you wish to build
into the model. In this case we have chosen to show 2d representations for
simplicity. In Figure 10 we present what we label our alphabetical order
visualization where we represent words by their spellings. A visualization
of spelling alone may not lead to many insights, but is useful for simple
demonstrations. One area where this simple encoding could be used would be
to visually compare irregular spellings in Elizabethan drama. It would
provide quick visual access to the places in texts that differed and needed
an editor’s attention. The real power in this method comes from being able
to encode as many connections as desired. Work has been done in the digital
humanities and computer science in the last few years in word embedding
models and the consistency of our method could aid in the process of
detecting connections in texts by using vectors. In figure 11, we have
graphed phonemes in two dimensions. Any highlights that form vertical lines
are showing alliteration in the poem. An example is in the line from Donne’s
poem: “Or like a thiefe, which till deaths doome be
read”. The words “deaths” and “doom” line up vertically
to indicate alliteration in this particular encoding. If we wanted to
visualize rhymes we would simply encode the phonemes in reverse, privileging
the final phoneme and we would generate a similar graph. The n-dimensional
nature of the models allows for as much or as little data coding as needed,
including relations between words. The only limitations on the questions
that can be asked are the imagination of the analyst.
6.0 QUALITATIVE STUDY
We wanted to discuss this project with a cross-section of scholars to develop a
better understanding of how people understood L-DNA and whether or not they saw
potential uses for their research. Our goal was to gain insight into whether
this technique could inspire reactions and possibly spark interest in the
approach.
6.1.1 Participants
We intentionally sought participants from a variety of disciplines. We had 14
participants which included 1 visual artist, 3 literary critics, 1
rhetorician, 2 digital media critics, 1 database programmer, 1 business
analyst, 1 linguist, 2 interaction designers, 1 graphic designer, and 1
marketing specialist.
6.1.2 Procedure
Each participant took part in a thirty-minute interview and was first shown
L-DNA visualizations that we had intentionally left void of any legends or
axis labels, so that we could ask questions about their initial
interpretations. After showing the image in Fig. 4, the mapping of the OED
in two dimensions, we asked what they thought the image might be. We then
showed participants Fig. 7, four languages plotted in the same space, and
the interviewer gave a thorough explanation of the how the algorithm works
and what they were seeing. We took time to make sure they were comfortable
with the explanation and asked about their understanding. We then showed
them Fig. 8 to be able to further explain L-DNA and asked questions based on
the participants’ understanding. We also asked participants to complete a
post-interview questionnaire. Five questions were asked on a 5-point scale,
with an opportunity to provide free form answers:
Q1. Once explained to what extent is the visualization readable?
Q2. Do you think the white space has meaning?
Q3. Is the white space necessary?
Q4. Does representing languages by colour and words as points work well?
Q5. Does this spatial representation of language trigger new ideas?
The following three questions asked for free form answers only:
Q6. What are your initial interpretations of this visualization?
Q7. Can you imagine a more suitable or readable structure?
Q8. Please provide any criticism you have about this visualization.
7.0 RESULTS: SCALE-BASED QUESTIONS
The scale questions were answered as follows. For Q1, 6 out of 14 participants
told us the visualization was clear after the explanation (5 out of 5 on the
scale). FOR Q2, 9 out of 14 people said that the white space carried meaning to
them (5 out of 5) and 6 participants thought that this white space was
completely necessary. For Q4, 7 out of 14 people ranked a 5 out of 5 for the
visualization approach. 10 out of 14 participants said that the visualization
triggered new ideas for research (5 out of 5).
8.0 DISCUSSION: FREE-FORM QUESTIONS & INTERVIEWS
We have formulated our discussion around the free-form questions. Since our
participants were experts from a variety of fields and domains, the similarities
in answers in some cases are particularly interesting. In other places it is the
difference in answers that encourages us as researchers in terms of the
potential of L-DNA as a tool for approaching questions about language. In this
discussion we include the questions and a discussion of the general themes that
arose in the responses.
8.1 The Power of Representation
Interestingly, when shown the images without any labels, participants tended
to engage in metaphoric comparisons of what they were seeing. The omission
of a legend led each participant to find something that was cognitively
analogous to what they were being shown. For example, some responses were:
“Is it zoomable? It has a DNA look or sort of ummm a matrix data flow
and I feel the urge to zoom it. It looks like I am really far away from
an unbound book, like how a book would be printed in sheets. It has that
type of aesthetic.”
“Well, it reminds me of DNA, like a screening test, it also looks like
a stamp, like someone has stamped something. It is a very tactile image,
I want to touch it.”
“It’s kind of like DNA. Like, uh, people show these images that
visualize DNA.”
In response to this process, five out of nineteen (26%) of our participants
from varying backgrounds and fields, with no explanation of what they were
looking at, described Fig. 4 as appearing like DNA — the inspiration for the
name of our technique. This result also demonstrates the power of
representation held within L-DNA. Some of the responses received from
participants included references to stamped or fading paper, city grids,
abstract art, and digital clock faces. Because we were trying to create a
space that could handle an ontology of words, the DNA metaphor was highly
applicable based on the implications of describing parts of a long chain of
information.
8.2 White Space
In our interpretation of the study data, the questions about whitespace
(Q2-Q3) produced perhaps the most interesting results. It was during this
question that most of the participants began to hypothesize about the
“space” in the languages they use every day. Essentially the parts
of the visualization with an absence of marks inspired thought, because they
were in stark contrast to the actual dots drawn on the screen. This result
strongly indicates the analogic or comparative possibilities of this
technique — the literary critic can begin to understand what makes a word
English by recognizing what is not a word, or by investigating what words
poets or writers use that push the boundaries of language. This result is
encouraging for the domain-specific problems of literary criticism. One of
our requirements is a space for analogues and with the whitespace in this
simple mapping there is significant affect. The response of our participants
to the relationship between the whitespace and the space occupied by words
creates a relationship that gives insight into what sets of symbols we use
and which we do not use in our language. The sheer size of the whitespace in
comparison forces an understanding of how few of the possible arrangements
of letters we actually use. With further work we think it is possible to
show that more complex mappings will produce more complex analogues.
It was generally agreed upon that it was the relation of the empty space to
the marked space that created meaning and inspired insight into what the
visualizations were showing and what they could show.
“the white space gives you a sort of ground against which it makes
sense... without it it might be even less evident”
It was in the white, or the lack of spellings, that our participants saw the
potential for growth in the language, or commented on the enormous range of
letter combinations that we do not use in English. This is encouraging
because as we build “meaning” into these visualizations we expect the
response to be comparative, and we expect new interpretations will result
from these comparisons. Initial reasons were as follows:
“Every letter can start a word, but does that mean that there’s
combinations of letters that don’t turn into words... you don’t have
words that start with trsz is that why there would be space... Maybe it
means that language is primitive, not as evolved as it could be.”
“I’m almost more interested in the dots in the white space... if
there's one I want to know what that is...the outliers are more
interesting.”
“What it does is show what the common letters are and common
overlaps.”
For the researchers this result was surprising, but was explainable. Without
“meaning” being built into this version we were simply showing a
part of orthography (spelling) and, in this stripped down version of the
possibilities of the space, it was the comparison between what was empty and
what was marked that sparked the interest of our participants. This is the
exact response we were interested in and we anticipate with more complex
representations we will be able to see more complex analogies. Some of the
responses to this whitespace analogy are as follows:
“I think the uniformity of the gaps are startling. It seems oddly
uniform and consistent. I think that potentially symbolically the gaps
can sort of be a formative quality of language and words will start to
fill in those gaps.”
“Well, I guess it kind of goes to show how constrained we are in
language, which shows how some things just are not possible with
spelling, which is kind of cool. There is so much blank space
particularly along certain lines, you get some sense that our alphabet
constrains us, which is why we have poets.”
These responses were typical of all our participants and are very promising
for future work.
8.3 Reluctant Inspiration
When asked if the images they were seeing inspired any new thought processes,
they proved to be exciting to our participants and the answers that follow
show the breadth of their thinking:
“You could map to any narrative, I would like to see this map out, A
Tale of Two Cities, it looks like a genetic footprint, like a genetic
phonetics, you get to see this formal genetic blueprint, it is more like
an autopsy of the text.”
“Yes, from a design and a fine art point of view. I think it is, again,
if you look at it as a design issue does it solve any problems, not yet,
but then again you are deconstructing something that already has that
problem solved so you are raising questions instead of solving problems
and you are raising interesting and meaningful questions.”
“Well, I mean my initial instinct is YES (emphasis given by the
participant), but I am not sure what that would be yet, I think it could
lead to a lot of productive conversations about how language operates or
the way someone creatively uses language.”
“These are important questions. Literature departments have always
performed as if they were in the shadows of the sciences and…[i]t seems
to me that this kind of work, although seemingly scientific, should be
the domain of literature department. It would be very important work for
us to do.”
Interestingly the 1 person out of 14 who said no to being inspired (included
below) touched directly on the fact that the simplicity of our
representation of orthography was limiting but suggested in his negative
response that it could be interesting for literary criticism if we could
find a way to include meaning, which is fruitful ground for future work and
possible with the three criteria we laid out above for the development of
the technique:
“I’m not sure it’s useful for literary theory or criticism because it
seems to explicitly set aside the question of meaning in favor of
orthography”
“Oh yea, um this is only for spellings well things get really
interesting when you get into phrases, rhetorical aspects of language
organization, the whole question of nuance which we like to think we are
studying as literary scholars. Once you get into these things and away
from mere spelling and into points that have meanings you can start to
clump together interpretation to different meanings.”
8.4 Criticisms from Participants
Our participants were in some cases critical of the design of L-DNA. In
particular, the most common criticism centered on making the technique
interactive, which is a clear (and previously planned) next step in our
iterative design process. Another criticism was that the encoding we used
was arbitrary, and its meaning was not immediately clear:
“The mechanism by which you map words to spatial locations, it’s kind
of arbitrary (maybe that’s not the right word). Mostly you see the first
two letters, and when you zoom in you’re seeing the space of subsequent
letters. I guess what this mapping is lacking is the Meaning of words,
or the semantics. They end up being distances from one another but … the
distance between related concepts, synonyms. I’m not necessarily saying
this is a bad visualization because of that. It just doesn’t encode the
meaning of words. Which is much harder. But, this is a legitimate way of
putting words in a consistent space.”
“The way that it is right now, that it’s static, it’s getting in the
way of itself. It demands explanation. It needs to be paired up with
something very practical, like reading a word or reading a sentence and
how that gets paired up in the system.”
We see this interactivity as the next steps in developing an application for
these types of interactions and it is in that interactivity that the
objections to the arbitrariness of the design will be addressed. We
hypothesize that being able to investigate the space dynamically, and by
defining other symbol systems to encode, the literary critic will be able to
explore the types of meaning and associations being looked for by our
participants.
9 STUDY RESPONSE
From our qualitative study, the overwhelming result was the importance of the
white space in our visualizations to the entire group that was interviewed. This
reaction has influenced the direction of our future work and demonstrates that
this space can be used for the types of comparisons that we are interested in,
namely those that lead to interpretive possibilities. We recognize that we have
presented a simplified form, but the technique itself allows for infinite
complexity. Some of our participants talked about the need to include
information that gives meaning to relationships and that is the next step in
developing mappings that can solve the original problem of creating a space that
can be experimented in with language and literary theory. Our study confirmed
the idea that this technique has the potential to answer these much more complex
questions as they relate to the domain in question. The sheer breadth of answers
to our question in relation to inspiring new ideas is extremely promising and we
take it as a success in developing the type of space that can inspire
investigation.
9.1 Prioritizing Whitespace
In response to the discussion with participants about the interest in
whitespace and the lack of density in certain regions, we produced a density
and inverted density map to highlight the white spaces, shown in Fig. 12 and
Fig. 13.
This was in direct response to comments from our participants such as:
“Yeah. But, one thing I was going to say… the UNIFORM inclusion of the
whitespace is interesting. But I wonder if there’s a better
representation of density and overlap.”
In this iteration of our design, instead of rendering each word in the
language we instead cluster groups of words into the boxes representing
pairs of words (“AA”, “AB”, etc.), and render the box using
transparency that corresponds to the density of words therein. The inverted
version of this mapping highlights all of the non-English words, which were
clearly of interest to participants.
9.2 ygUDuh
We have discussed how the blank spaces are compelling and how the absence of
words in these spaces generally seems to intrigue people.
It is interesting to note that this space is and has been filled in many
interesting ways. For example, E. E. Cummings poem ygUDuh does not contain a
single English word, yet it can be read as English, where the “words”
take on the sonic characteristics of English when read aloud. For example,
the first few lines of the poem are:
ygUDuh
ydoan
yunnuhstan
That when read allowed becomes a phonetic map for a type of early 20th century urban slang exemplified by the
poem:
you gotta
you don't
you understand
In Fig. 14 we have plotted ygUDuh overlaid on the OED grid. Note how these
“non words” exist largely on the edge of the word spaces and the
white spaces. This may be because, while they are not English words — hence
the white space proximity — they have similar vowel and consonant structures
to English words. By seeing these words overlaid on the whole language, we
can see a visual representation of Cummins’ craft, of the attempt to make
non English words that sound like English when read aloud. The fact that all
of the non-words are situated on the edges of heavily populated space tells
us that these arrangements of letters that we try to make into words when
reading the poem are “closer” to English words than we think, at least
in terms of spelling.
9.3 Instant Messaging
Another example where new types of “words” or at least English
communications are evolving is in instant messaging, text messaging, and
social media. It seems that for ease and speed, we can give up many letters
— chiefly the vowels — in words and still retain meaning. Fig. 15 shows MSN
“words” overlaid on the OED visualization. It is interesting to
note that many of the new “words” (marked with red dots), fall in the
spaces where very little or no words exist. This demonstrates that even in a
type of shorthand, like the one used in instant messaging (e.g., “btw”,
“lol”, “ttyl”, etc.) that many of the newly created words are
spelled with letter combinations that simply don’t exist in the language.
This is partially a result of the volume of acronyms used in instant
messaging but it becomes obvious by “reading” the image that many of
the words used fall on the top line of each row suggesting (such as with row
A, and row I) that many of these “words” and acronyms begin with those
letters. In this way our technique produces visuals that allow us to ask
further questions about the organization of our data.
9.4 Interaction
We have also begun to integrate interactive elements into our visualization,
some of which were planned prior to our qualitative study, and some of which
were inspired by our results. In particular, we have already created a
version of L-DNA which incorporates a brushing technique that presents the
“words beneath the cursor”, both when dragging across words and
when dragging across the whitespace. We have also created a version of the
density map (Section 9.1) that allows zooming into the recursive letter
pairs. For example, it is possible to click or tap on the “BA” square,
then the “NA” square, then another “NA” square to then see the
word “BANANA” as shown in the static image of figure 3.
10 CONCLUSION
In this paper we have introduced L-DNA and presented the findings of a
qualitative study of its design. In L-DNA, we have developed a mapping of symbol
systems to visual space, which we have demonstrated using language. Our
formulation has several properties that are valuable for the analysis of
language and are not available in some other common visualizations of language.
L-DNA has the following important features:
- The L-DNA space is capable of handling any symbol string from a null
string to a string of infinite length. L-DNA can be used in 1, 2, or n
dimensions.
- L-DNA space is infinite in that between any two points (words) in the
space there exists another word — though it may not have semantic
meaning.
- The L-DNA space is one-to-one and onto (bijective). Every unique coding
maps to a unique position and, in reverse, words (or any original
information) can be recovered uniquely from the visual space.
We have mathematized language to make exploring and experimenting with language
easier, but the results of said experiments need to have the possibility of
being reversed out of the space to be able to assign meaning once again to the
language. We have also presented a qualitative study, which provided encouraging
results that indicate the power of the type of representation provided by L-DNA,
the benefit of the whitespace that it generates, and its possibilities to
provide inspiration (even if reluctantly), as well as some useful criticisms
that led to iterations in our design.
Works Cited
Alexander 2012 Alexander M., “Patchworks and
field-boundaries: visualizing the history of English”, Digital Humanities
(2012), Hamburg Germany.
Bingenheimer 2006 Bingerheimer M., Xishihu,
J. “Social network visualization from tei data”, Lit Linguist Computing 26 (3):
271-278 (2006).
Bingenheimer 2009 Bingenheimer, M., Hung,
J., and Wiles, S. “Markup meets GIS – Visualizing the ‘Biographies of Eminent
Buddhist Monks’”, In Banissi, E. et al. (eds), Proceedings of Information
Visualization IV Piscataway/NJ: IEEE Computer Society, pp. 550–4 (2009).
Chen 2006 Chen, C. “CiteSpace II: Detecting and
visualizing emerging trends and transient patterns in scientific literature”,
JIS&T 57, 359–377 (2006).
Chuang 2012 Chuang J., Ramage D., Manning C.,
Heer, J. “Interpretation and Trust: Designing Model-Driven Visualizations for
Text Analysis” ACM Human Factors in Computing Systems (CHI), 2012.
Clement 2008 Clement, T. E. “A thing not
beginning and not ending: using digital tools to distant-read Gertrude Stein’s
The Making of Americans.”, Literary and Linguistic Computing, 23(3): 361–81
(2008).
Collins 2009a Collins, C., Carpendale, S. and
Penn, G. "Docuburst: Visualizing document content using language structure."
Computer Graphics Forum. Vol. 28. No. 3. Blackwell Publishing Ltd (2009).
Collins 2009b Collins, C. Viegas, F. Wattenberg,
M.. "Parallel tag clouds to explore and analyze faceted text corpora." Visual
Analytics Science and Technology, 2009. VAST 2009. IEEE Symposium on. IEEE
(2009).
DeCamp 2005 DeCamp, P., Frid-Jimenez, A., Guiness,
J., & Roy, D. “Gist icons: Seeing meaning in large
bodies of literature”, In Proc. of IEEE Symp. on
Information Visualization (InfoVis 2005), Poster Session,
Minneapolis, USA, October, (2005).
Dict 2015 http://www.winedt.org/Dict/ (Accessed
April 19 2015)
Frye 2000 Frye, N., The Anatomy of Criticism,
Princeton University Press (2000): 7.
Gardiner 2010 Gardner, M. J., Lutes, J., Lund,
J., Hansen, J., Walker, D., Ringger, E., and Seppi, K. “The Topic Browser: An
interactive tool for browsing topic models”. In NIPS (Workshop on Challenges of
Data Vis), (2010).
Gould 2003 Gould, R. V. “Uses of network tools in
comparative historical research”, in Mahoney, J. and Rueschemeyer, D. (eds),
Comparative Historical Analysis in the Social Sciences. Cambridge: Cambridge
University Press, pp. 241–69 (2003).
Gretarsson, B., O’Donovan,
J., Bostandjiev, S., Llerer, T. H., Asuncion, A., Newman, D., and Smyth, P.
“TopicNets: Visual analysis of large text corpora with topic modeling”, Trans on
Intelligent System and Technology 3, 2 (2011).
Hall 2008 Hall, D., Jurafsky, D., and Manning, C.
D. “Studying the history of ideas using topic models”. In EMNLP, 363–371
(2008).
Hearst 1995 Hearst, Marti A. “TileBars: visualization of term distribution information in full text
information access.”
Proceedings of the SIGCHI conference on Human factors in
computing systems. ACM Press/Addison-Wesley Publishing Co.,
(1995).
Hetzler 1998 Hetzler, Beth, et al. “Multi-faceted insight through interoperable visual information
analysis paradigms.”
Information Visualization, 1998. Proceedings. IEEE
Symposium on. IEEE, (1998).
Jessop 2006 Jessop, M. “Dynamic Maps in
Humanities Computing”. Human IT, 8.3, 68–82 (2006).
Jessop 2008 Jessop, M. “The Inhibition of
Geographical Information in Digital Humanities Scholarship”, Literary and
Linguistic Computing, 23(1): 39– 50, (2008).
Lee 2010 Lee, Bongshin, et al. “Sparkclouds: Visualizing trends in tag clouds”. Visualization and Computer Graphics, IEEE Transactions on 16.6:
1182-1189, (2010).
Lin 1992 Lin, X. “Visualization for the document
space”. In Vis, 274–281 (1992).
Moretti 2005 Moretti, F, Graphs, Maps, Trees:
Abstract Models for a Literary Theory. London: Verso, 2005.
Paley 2002 Paley, W. Bradford. "TextArc: Showing
word frequency and distribution in text." Poster presented at IEEE Symposium on
Information Visualization. (2002).
Posavec 2015
http://www.stefanieposavec.co.uk/-everything-in-between/#/writing-without-words/
(Accessed April 19 2015)
Rohrdantz 2012 Rohrdantz, C., Niekler, A.,
Hautil, A., Butt, M., Keim, D. “Lexical semantics and
distribution of suffixes: a visual analysis”. EACL 2012 Proceedings of the EACL 2012 Joint Workshop of LINGVIS &
UNCLH, pages 7-15. (2012).
Simon 1988 Simon, Herbert A. “The science of design: creating the artificial.”
Design Issues: 67-82,(1988).
Smalheiser 2009 Smalheiser, N. R., Torvik, V.
I., and Zhou, W. “Arrowsmith two-node search interface: a tutorial on finding
meaningful links between two disparate sets of articles in MEDLINE” Computer
M&P in Biomedicine 94, 2, 190–197 (2009).
Tapor 2015 portal.tapor.ca (Accessed April 19
2015)
http://camilleutterback.com/projects/text-rain/(Accessed April 19 2015)
Valley 2015 Valley of the Shadow
http://valley.vcdh.virginia.edu/ (Accessed April 19 2015)
Viegas 2008 Viegas, F. B., and Wattenberg, M.
“TIMELINES: Tag clouds and the case for vernacular visualization. Interactions”
15, 49–52 (2008).
Viegas 2009 Viegas, Fernanda B., Martin
Wattenberg, and Jonathan Feinberg. “Participatory
visualization with wordle”. Visualization and
Computer Graphics, IEEE Transactions on 15.6: 1137-1144,
(2009).
Voyant 2015 http://voyant-tools.org/ (Accessed
April 19 2015)
Wattenberg 2008 Wattenberg, Martin, and
Fernanda B. Viégas. "The word tree, an interactive visual concordance."
Visualization and Computer Graphics, IEEE Transactions on 14.6: 1221-1228
(2008).
Wise 1995 Wise, J., Thomas, J., Pennock, K.,
Lantrip, D., Pottier, M., Schur, A., and Crow, V. “Visualizing the non visual:
spatial analysis and interaction with information from text documents”, in
InfoVis, 51–58,(1995).
Wood 1992 Wood, D., The Power of Maps. New York:
Guilford Press (1992).
WordCollider 2015
http://www.itsokaytobesmart.com/post/26993944713/word-colliderparticiple-
physics (Accessed April 19 2015)
Yatani 2011 Yatani, K., Novati, M., Trusty, A.,
& Truong, K. N. “Review spotlight: a user interface for summarizing
user-generated reviews using adjective- noun word pairs”. In Proceedings of the
2011 annual conference on Human factors in computing systems (pp. 1541-1550).
ACM (2011).
van Ham 2009 van Ham, F., Wattenberg, M., and
Viegas, F. B. “Mapping text with phrase nets”, in InfoVis, 1169–1176
(2009).