- Open Access
Automated deception detection of 911 call transcripts
© Burns and Moffitt; licensee Springer 2014
- Received: 7 July 2014
- Accepted: 1 August 2014
- Published: 12 August 2014
This study is a successful proof of concept of using automated text analysis to accurately classify transcribed 911 homicide calls according to their veracity. Fifty matched, caller-side transcripts were labeled as truthful or deceptive based on the subsequent adjudication of the cases. We mined the transcripts and analyzed a set of linguistic features supported by deception theories. Our results suggest that truthful callers display more negative emotion and anxiety and provide more details for emergency workers to respond to the call. On the other hand, deceivers attempt to suppress verbal responses by using more negation and assent words. Using these features as input variables, we trained and tested several machine-learning classification algorithms and compared the results with the output from a statistical classification technique, discriminant analysis. The overall performance of the classification techniques was as high as 84% for the cross-validated set. The promising results of this study illustrate the potential of using automated linguistic analyses in crime investigations.
- Automated Linguistic Analysis
- Deception Detection
- 911 Calls
In part due to a natural “truth bias,” humans (including those with special training) can generally detect deception at a rate only slightly better than chance, at around 54% ,. This inability to accurately separate truth from deception can have serious consequences; this is particularly true in the case of law enforcement. Not only do criminal investigators have to be concerned with deception that is not detected (false negatives), but they must also take into account the serious outcome of labeling truth tellers as deceivers (false positives). Thus, there is a critical need for more reliable and accurate methods of identifying deception, especially in the earliest contact between a suspected perpetrator and law enforcement. In this study, we evaluate linguistic cues extracted from transcriptions of 911 homicidea calls as potential indicators of deception.
As a dataset, 911 calls may be ideal for deception detection research because they occur in a real-world setting, are relatively unrehearsed, occur soon after the crime in question, and are part of emotionally charged situations ,. In contrast, in laboratory settings where deception is often sanctioned, the consequences of lying and getting caught are minor, and incentives to deceive are artificial (e.g., ,). Thus, as Mann et al.  suggest, these low-stakes laboratory environments may not induce feelings of guilt or elicit behavior found in real settings; as a result, they may adversely affect researchers’ ability to accurately judge credibility, and therefore diminish the external validity of the results. Thus, it is important to add real-world scenarios to the collection of deception research data.
Deception studies using real-world person-of-interest statements begin to address the research gap identified above –. Person-of-interest statements are written explanations of crimes by a person who has not been formally charged with a crime but who is “of interest” to law enforcement in an investigation. Studies have been conducted using pre-polygraph interviews, which also involve real-world data . However, relative to 911 calls, the time between the crime and the statements in pre-polygraph interviews is much longer, giving deceivers time to rehearse their responses. Additionally, such datasets may be influenced by investigative procedures or the deceiver’s contact with acquaintances between the incident and the statement ,. Thus, 911 calls may be considered a less biased dataset because the interaction with authorities occurs much sooner after the crime and the individual’s words have been less affected by outside influences. Therefore, this dataset provides researchers an unusual opportunity for an unfiltered look at deception.
A subset of 911 calls reporting homicides are available on the Internet because they are high-profile crimes. The calls can be corroborated for ground truth by examining the associated court outcomes. Establishing ground truth is one of the most difficult aspects of research into deception detection . Because initial homicide reports, subsequent investigations, legal proceedings, and judgments are covered widely by news media, it is possible to substantiate ground truth to a high degree of certainty for these types of 911 calls. Although 911 homicide calls represent an extremely small fraction of total 911 calls, they can serve as a proxy for calls made reporting other high-profile crimes such as arson, bomb threats, and sexual assault.
Data mining techniques, including text mining, linguistic feature mining, and classification by text features, can be used to analyze the “caller side” of the transcripts of these calls. Text mining involves looking for hidden patterns or cues in texts, while linguistic feature mining refers to dissecting texts with respect to specific linguistic categories, such as words associated with positive affect. Text mining is a multidisciplinary research area that combines approaches used in the fields of computer science, linguistics, mathematics, communication, and psychology. It focuses on using computing power to process unstructured human language in spoken or written form ; furthermore, text mining has been used to process text data to discover linguistic cues or features in order to classify documents, including fraudulent versus non-fraudulent financial statements , or deceptive versus truthful statements in instant messages, email exchanges, or person-of-interest statements –,.
In this study, we applied linguistic feature mining to evenly matched (i.e., deceptive matched with truthful) transcripts of 911 homicide calls via Linguistic Inquiry and Word Count (LIWC). LIWC 2007, a general-purpose psychosocial linguistic dictionary comprising 4,500 words and word stems, has been used by researchers to quantify linguistic cues for deception –.
Using a truthful/deceptive matched convenience data set of fifty 911 calls, we identify linguistic cues based on deception theories that may be used to discriminate between deceptive and truthful 911 calls;
We extract useful information from unstructured text comprised of transcribed 911 calls to demonstrate that the largely unexploited data of 911 calls can be analyzed for further investigative work; and
The classification results achieve up to 84% accuracy.
The remainder of this paper proceeds as follows: we first review deception theories and previous research involving automated or manual deception detection; we then examine advantages and disadvantages of using 911 calls as a data source and discuss linguistic cues of deceptive and truthful 911 callers as we develop our research question and hypotheses; next we describe our methodology, and finally we present and discuss the results, research contributions, limitations of the study, and future research directions.
Deception is defined as purposefully concealing the truth, either by omission or commission. In the present study, our general hypothesis is that 911 callers who deceive exhibit systematic differences in the words they use compared to 911 callers who are telling the truth. Possible underlying causes for the differences in deceptive speech are described by deception theories, including four factor theory , interpersonal deception theory (IDT) , information manipulation theory , and reality monitoring . Deception detection researchers rely on theory to identify strategically employed clues that can discriminate between those who deceive and those who do not. In this study, we rely on four factor theory and IDT because they best fit the interpersonal context of a 911 call.
Four factor theory  delineates four processes or factors that underlie deceivers’ behaviors. Control, the first process, describes how deceivers control or suppress their behavior to try to conceal their deception. For example, in 911 calls, deceivers will manage the linguistic and paralinguistic features of their interaction with the dispatcher in order to appear as truthful as possible and not to induce suspicion. The second factor, arousal, refers to various autonomic arousal responses of the deceiver’s central nervous system that coincide with the deceptive behavior or story. Felt emotion, the third factor, encompasses various emotions that deceivers experience, specifically guilt, anxiety, and/or satisfaction in pulling the wool over others’ eyes (i.e. “duping delight”). For instance, because of the negative feelings associated with guilt, deceivers try to disassociate themselves from their crime by referring to others rather than to the self through a greater use of through the use third-person pronouns. Anxiety may also impair the quality of the control that deceivers use to conceal their deception. Finally, due to the fourth factor, cognitive processing, deceivers have an increased cognitive load as they fabricate and maintain lies. This factor ties into a proposition of IDT, specifically that high cognitive load may be detrimental to a liar’s performance and increase the chances of detection. Because of this, deceivers in 911 calls may shorten their responses and use a smaller set of words.
According to IDT ,, deception is goal-oriented and strategic. Arising out of interpersonal communication and deception research, IDT predicts the behaviors of senders and receivers in an interactive contextb. The theory acknowledges the “superordinate role” of the context and relationship within which the interaction occurs. For example, the situational factors of a 911 homicide call will influence how deceptive exchanges play out, and consequently, the hypotheses regarding these exchanges. IDT proposes that the behavior of the deceiver will vary systematically with the spontaneity of the interaction (i.e., lying to a 911 operator requires more dexterity than lying in a written letter) and the immediacy of the context. In 911 calls, the interaction is spatially non-immediate, but temporally immediate. IDT also predicts that deceivers are strategic in managing the information they send, their image, and their behavior; as a result, however, they experience nonstrategic byproducts including dampened affect, noninvolvement, and performance decrements. In the current study, callers should experience increased cognitive loads relative to most contexts, making deception even more difficult. The increased stakes and spontaneity of the conversation may impair deceptive performance.
These theories lend support to the feasibility of using linguistic analysis to carry out deception detection. To discern these deceptive patterns in communication, many recent studies involving automated linguistic cue analyses have leveraged a general-purpose, psychosocial dictionary such as LIWC ,-,-. LIWC contains predefined categories composed of words related to a particular construct, such as Anxiety or Negative Emotion. Depending on the context in which deception occurs, deceivers have been found to display elevated uncertainty and affect, share fewer details, provide more spatiotemporal details, and use less diverse and less complex language than truth tellers ,,. Researchers have also documented cases in which deceivers use more words, group references and use more informal, non-immediate language than truth tellers ,,.
Researchers who have conducted manual content analyses have also documented linguistic markers of deception ; for example, perpetrators of homicide or kidnapping may use past tense (vs. present or future) when discussing the victim, and deceivers may try to distance themselves from the crime by using “they” rather than “I” in statements . Law enforcement researchers , were successful in manually coding verbal indicators to classify 911 homicide callers as “guilty” or “innocent.” In these studies, the key verbal indicators of guilt included extraneous information, inappropriate politeness, a lack of plea for help, and evasion. In an earlier study, Olsson  analyzed emergency calls made to report fires in London, UK. He found that hoax calls could be discriminated from truthful ones based on how the caller described or implied his/her relationship to the emergency and the urgency and cooperation of the caller.
The original mode of communication (i.e., written vs. spoken language) is particularly important with respect to linguistic cues that distinguish between truth tellers and deceivers. As noted in IDT, deception is strategic and the spontaneity of the interaction is important ,. Thus, for example, a team tasked with writing the text to include with financial statements has months in which to develop a document that may include strategic misrepresentation or obfuscation via increasing word count and employing words of more than three syllables ,. On the other hand, a person calling a 911 operator after committing a crime has far less time for strategic wordsmithery and may use fewer, simpler words to suppress or control verbal cues for deception.
Contrary to media hype regarding highly publicized crimes, most 911 calls are not full of drama, excitement, and/or rich descriptions of the crime, crime scene, or victim(s). The 911 operator is trained to elicit information to deploy the right emergency resources as quickly as possible. Although a dramatic outpouring of emotion may be observed in some calls, most 911 calls comprised a great deal of mundane information gathering and information passing to clarify names, addresses, and directions; give life-saving instructions (such as the steps to perform cardiopulmonary resuscitation (CPR)); and ask questions that only require a one-word answer (e.g., “is the victim breathing?”).
There are several key advantages to using automated linguistic analysis techniques to detect deception in transcripts of 911 homicide calls. First, 911 calls represent the initial contact between a caller and an emergency response team, including law enforcement, leaving callers little time to prepare or to settle down for the encounter. Moreover, as noted above, the caller’s statement has not yet been corrupted by contact with criminal investigators or lawyers . Furthermore, because callers do not perceive 911 operators to be members of law enforcement, deceptive callers may exhibit less controlled behavior and more cues of deception; consequently, 911 callers may become more engaged in interpersonal communication and be less guarded because 911 operators interview rather than interrogate. Due to the temporal immediacy of the crime in relation to the 911 call, there may also be more active stress on the caller, in turn causing the caller to display more deceptive cues. Another advantage for comparison of truthful and deceptive responses in emergency calls is that 911 operators use a structured interview style that was similar across calls.
On the other hand, there are a number of issues with 911 calls that make them a less than ideal source for linguistic analysis. First, these calls can be very short. In the dataset used for this study, calls ranged from less than thirty seconds to over ten minutes in length. On average, the calls were about three to four minutes. Another problem is that there can be a lot of dead air time while the 911 operator puts the caller on hold to coordinate with the rest of the emergency team; such gaps cannot be used in linguistic analyses. Third the sound quality of the calls can make them difficult to transcribe. Fourth, dispatchers, operators, callers, and rescue workers may talk simultaneously. Fifth, certain sounds such as sobs, shrieks, or gasps of pain cannot be easily transcribed and/or analyzed using current linguistic feature dictionaries. Finally, a caller may be poised to offer what could prove to be valuable clues to future investigators, but 911 operators may have to interrupt them to elicit the best information for timely dispatching of the right resources. Consequently, to accomplish their primary mission, the operators who perform both call-taking and dispatching functions may restrict “telling” cues or information .
The strengths of the dataset ultimately outweigh the potential weaknesses of the dataset. Thus, we analyzed cues in transcribed 911 calls using the same approach that other researchers have adopted to analyze unstructured texts with a view to distinguish between deceivers and truth tellers. Based on IDT, four factor theory, and previous research, we considered that deceivers may try too hard to cover up what they perceive to be deceptive cues. For example, in a face-to-face context, deceivers may try to limit fidgeting and/or posture shifts, thereby displaying more “stillness” during a lie in an attempt to inhibit overt signs of deception . In the circumstances of this study, where callers are not visible, deceptive 911 callers may restrict their verbal responses, exhibiting low affect or shortened responses when compared to truth tellers. This may be observed in a higher rate of negation and assent words (e.g., “no” and “yes”) that may be used by deceivers to limit and control answers. Deceivers may also feel emotions such as guilt, anxiety, and/or duping delight . To lessen these feelings, they may attempt to distance themselves from the situation. The use of first-person singular constructs implies that the speaker “owns” the statement. However, because liars try to distance themselves  from the crime or bad situation, they include fewer self-references in retelling stories . Thus, we should find that deceptive 911 callers use more third-person plural to share the blame.
Four factor theory posits that cognitive effort is required to not only lie, but also to maintain the lie. Vrij, Fisher, Mann, and Leal  expand upon this claim and outline all of the tasks a deceiver undertakes that increase cognitive load, including developing a plausible lie, self-monitoring for credibility, monitoring the listener, and remembering the details of the lie while concealing the truth. In short, they argue that lying requires strategic intent and more cognitive effort than truth telling. To mitigate these effects, a liar may rehearse his or her story to keep the facts and details in order. However, because deceptive 911 callers have often had little opportunity to rehearse between the time of the crime and the time of the call, they may face extreme cognitive overload. Thus, these callers may compensate by supplying shorter, controlled statements, sharing fewer details with emergency responders (such as those that would be helpful in locating the physical address of the victim), and asking the 911 operators to “hold on” or “wait” when the operator gives instructions.
Based on these theories and building on previous research on deception detection using linguistic cue analysis, our research question is as follows:
Can automated linguistic analysis techniques accurately classify deceptive versus truthful callers in transcripts of 911 homicide calls?
To define the types of cues that can be examined for deception or truthfulness in 911 calls, we suggest that truth tellers will exhibit more immediacy through greater use of first-person singular and first-person plural words. On the other hand, we expect deceivers will show more non-immediacy, a distancing from what is said, by referencing others in the third-person singular or plural. To control verbal output, or to suppress reactions or answers, deceivers will tend to answer more frequently with shorter, simpler “yes” or “no” answers. Because deceivers tend to suppress reactions, we expect that truth tellers will display more felt emotion. On the other hand, we anticipate that deceivers will use more swear words because instances of swearing can be perceived as more credible . Thus, deceivers may include swear words as a way to appear to be emotionally connected to an incident while suppressing their true emotions. Relative to deceivers, truth tellers have a lower cognitive overload that allows them to give more location details, such as house numbers and generic information about location. Truth tellers want to provide many clues to get rescue teams to their location as quickly as possible, and will therefore provide specific addresses and phone numbers more clearly, as well as giving more details about the location, such as whether it is a house or apartment building. Finally, due to cognitive overload deceivers may be reluctant, or find it difficult to follow life-saving instructions given by the dispatcher even though not doing so would seem suspicious.
Formally stated, the hypotheses in this study are as follows:
Deceptive 911 callers will display:
(a) higher use of third-person plural, (b) higher use of third-person singular, (c) more assent terms, (d) more negation terms, (e) more emotionally-charged swearing, (f) more inhibition words, and
(g) lower use of first-person plural, (h) lower use of first-person singular, (i) less negative emotion, (j) less anxiety, (k) lower use of numbers, and (l) lower use of generic location details than truthful 911 callers.
Our dataset represented a convenience sample obtained from publicly available sources found on the Internet. The majority of the calls came from Dispatch Magazine On-Line (911dispatch.com). This resource contains a tape library of public domain 911 and other emergency calls that have been collected since 2006. The website contains documentation about the calls that enables the user to determine the outcome of the case.
Because we do not have control over the chain of custody of these 911 calls, we cannot state to what degree they were edited. For the most part, personally identifying information has been redacted, but we have no reason to believe that the calls were edited otherwise. However, the inability to state this conclusively is a potential limitation of this convenience dataset. Still, the archive presents a unique opportunity to access this type of real-world data.
The final dataset of 50 transcribed 911 calls was equally split between truthful and deceptive callers. The size of the dataset was restricted by the number of publicly available deceptive calls for which ground truth could be established. To establish ground truth, we corroborated subsequent arraignment, prosecution, and/or admission of guilt via news articles about the crimes based on the information at the source website and other websites as necessary. Once we had identified 25 deceptive calls, we randomly chose 25 calls from our set of transcribed truthful calls to create a matched set. After transcribing the calls and removing the 911 operators’ side of the conversations, we analyzed the caller side of the transcripts using LIWC. LIWC normalizes the data by dividing category counts by the number of words in each document.
Constructs and corresponding LIWC categories
1st person plural, 1st person singular
3rd person singular; 3rd person plural
Negative emotion; Anxiety
Lack of felt emotion
Lack of cognitive overload
Numbers; Leisure (location-related)
The first step for identifying significant linguistic cues was to run a one-tailed independent sample t-test for each linguistic cue to compare truth tellers with deceivers. We considered each 911 call transcript to be an independent observation, since it represented one call placed by a unique caller.
Results of Classification Algorithms
Logistic model tree induction
Each machine-learning algorithm builds a model based on a different set of theoretical premises. Logistic tree model induction is a classifier for building logit models using regression functions as base learners . A simple naïve Bayes is a probabilistic classifier based on Bayes’ theorem. A neural network is a “black box” that performs a classification using hidden layers. A random forest is a type of decision tree that applies decision rules to divide an overall dataset into smaller classification sets. Using these theoretically diverse algorithms, we reduce the likelihood of relying on the results of one algorithm that over-learns the data and fails to generalize to a broader population.
Analysis of constructs/variables (LIWC categories) in transcripts
Associated variables (LIWC categories)
Truthful Std Dev
Deceptive Std Dev
1st person plural
T > D
D > T
1st person singular
T > D
D > T
3 rd person singular
D > T
T > D*
3rd person plural
D > T
D > T*
D > T
D > T*
D > T
D > T*
T > D
T > D*
T > D
T > D*
Extreme swearing (Sexual)
D > T
T > D
T > D
T > D*
T > D
T > D*
D > T
D > T*
On average, compared to truthful calls, deceptive 911 calls exhibited greater use of “they” (t(50) = 1.802, p < .05). They also involved more negation (t(50) = 2.031, p < .05) and assent terms (t(50) = 1.905, p < .05). Furthermore, they displayed a higher rate of inhibition terms (t(50) =2.428, p < .05). In contrast, truthful callers used more numeric words (t(50) = -2.417, p < .05) and leisure (location-related) words (t(50) = -2.109, p < .05). The transcripts for truthful calls contained more negative emotion words (t(50) = -1.915, p < .05), terms for anxiety (t(50) = -1.975, p < .05), and leisure (location-related) terms (t(50) = -2.109, p < .05). Table 2 reports the results of the classification algorithms for both a training set and a cross-validation set.
The overall performance of the classification techniques was very strong and ranged from 70% to 84% for the cross-validation tests. The results yielded predictive models with much higher accuracy than that of unaided humans, which, as mentioned above, is 54% ,. The best classification method used was discriminant analysis, followed by logistic model tree induction.
Variables (LIWC Categories) with Examples from Transcripts
3rd person singular
T > D
She's right on the floor. She's not breathing.
3rd person plural
D > T
Yes, they said, they said if they heard anything they were going to my house.
D > T
No, nothing, he's gone.
D > T
Okay, they’re here.
T > D
There was a fight. It was terrible.
T > D
I found out about an hour ago and I've been in a panic ever since.
T > D
Five seventeen West Doty Street
T > D
I see her in her garage right now.
D > T
[conversation while the operator is trying to give CPR instructions] Hold on, I have to throw up, please hold on.
The results suggest that truthful callers display more negative emotion and anxiety than deceivers, who tend to display flat affect. Although we had hypothesized that deceivers would use more swear words as an attempt to appear more credible through a faked emotional response, we actually discovered that truth tellers used more extreme swearing (the mean for truthful swearing = .0652; the mean for deceptive swearing = .0000). Emotionally charged swearing was another way for truth tellers to convey negative emotion or frustration during the calls. This finding corresponds to previous research that demonstrated that the primary reason that people swear is to express negative emotions or frustration ,.
Honest callers also tended to refer to others in third-person singular. To aid emergency responders, truth tellers used more numbers related to addresses and/or phone numbers and used names of locations, such as “apartment” or “garage”.
Deceivers used third-person plural at a higher rate, perhaps to distance themselves from an incriminating situation. However, contrary to our hypotheses, they also demonstrated more immediacy than truth tellers by using both first-person singular (the mean for truthful first-person singular = .2964; the mean for deceptive first-person singular = 1.1760) and first-person plural pronouns (the mean for truthful first-person plural = 9.4136; the mean for deceptive first-person plural = 10.6788).
Deceivers’ use of negation and agreement words may have represented their need to suppress or contain their own verbal responses and/or affect. Finally, deceivers tended to tell the 911 operator to “wait” or “hold [on]” (inhibition terms) at a higher rate than truth tellers. This occurred in the 911 calls when the operator asked them to do something they were reluctant to do, such as CPR.
The aim of this research was to expand the understanding of how we can analyze 911 call transcripts for crime analysis and solving. Thus, this study makes three major contributions. First, it has advanced deception detection research by applying linguistic feature mining methods to a unique corpus, namely, transcripts of 911 homicide calls, which represents extremely raw and largely unrehearsed human communication. We determined that deceivers use language and linguistic cues differently than truth tellers in a high-stakes, real-world corpus of 911 homicide calls. Thus, this study represents a successful proof of concept of using automated linguistic analysis to classify deceptive versus truth telling 911 homicide calls accurately, quickly, and objectively in comparison to manual methods of content analysis, which involve extensive training, time-consuming analyses, and subjectivity. Through analysis of 911 calls, law enforcement could detect deceptive, guilty perpetrators earlier in the investigative process and use that information for crime analysis. Second, we extracted useful information from unstructured text comprising transcribed 911 calls to demonstrate that these largely unexploited data can be analyzed for further investigative work. Third, this study provided strong classification results of up to 84% accuracy (cross-validated). These results approach the highest reported accuracy of field-based polygraph tests which is 92% ,. Combined with speech-to-text software, automated deception detection techniques could be used to monitor 911 calls in real time. Although the mission of the 911 operators would remain the same, automated monitoring of 911 calls could enable law enforcement and rescue workers to focus efforts and resources more successfully on post-hoc crime solving and analysis.
Despite these contributions discussed above, the convenience sample of 50 archived 911 homicide calls downloaded from the Internet represents a limited dataset. The size of the current dataset was restricted by the number of publicly available deceptive calls for which ground truth could be established. Furthermore, the 911 calls were taken from the Internet, not the original source. As a result, some information—such as names—may have been blocked out to protect the privacy of the caller when the 911 call was released to the public. An additional limitation arises from the fact that we did not control the chain of custody over the 911 calls. Thus, we are not able to establish whether or to what degree the 911 calls had been edited. Moreover, only certain states currently release 911 calls to the public, so the geographic origin of the calls could be skewed. Finally, the calls were restricted to those placed by English-speaking callers located in the U.S. To counter these limitations in the future, audio files and/or transcripts should be collected directly from law enforcement in subsequent research to validate our results, and sampling from diverse geographical locations should be carried out.
Accurate, credible assessment decisions are critical for law enforcement personnel, who may not detect deception, may act on false positives, or may use incorrect information for crime analyses. Moreover, other clues for deception, such as vocalic cues, should be studied in conjunction with linguistic cues for a practical decision support tool that includes combined analyses. In the future, an integrated system for deception detection that can be used in real time in high-stakes 911 calls could be added to law enforcement’s overall crime-solving and crime-analysis strategy. Text mining tools could also be used to analyze 911 calls linguistically as part of a portfolio of crime-solving techniques to enable law enforcement and rescue workers to focus their efforts and resources more successfully.
This study’s findings provide critical knowledge about how deceivers communicate during typically unrehearsed verbal exchanges and expand the usefulness of deception models from a low-stakes, laboratory setting into a high-stakes, real-world environment where there are serious consequences not only for the deceiver, but also for law enforcement. The next step for this research will be to validate these results using a larger real-world dataset. Additionally, when we use a larger dataset, we can conduct testing of revised hypotheses for 1st person plural (D > T), 1st person singular (D > T), and swearing (T > D) to establish if new results are significant in the revised directions. Current and future quantitative models and decision aids could assist law enforcement in detecting deception.
aIn this paper, the term “homicide” includes not only murders, but also accidents that were later deemed by investigators to be homicides, reports of kidnapping that masked an underlying homicide or neglectful death, and murder-suicides.
bA unique characteristic of this dataset that may not be true in others is that 911 deceivers are treated implicitly as truthful by the dispatcher. A 911 dispatcher operates under the assumption that the caller is telling the truth. Therefore, a deceptive caller may never change his story based on his perception that the operator is suspicious of it.
We are pleased to acknowledge the generous support from a Center for Identification Technology Research (CITeR) Grant for funding the data collection and analysis phases of this project.
- Bond CF, DePaulo BM: Accuracy of deception judgments. Personal. Soc. Psychol. Rev. 2006, 10: 214–234. 10.1207/s15327957pspr1003_2View ArticleGoogle Scholar
- Aamodt M, Custer H: Who can best catch a liar? Forensic Examiner 2006, 15: 6–11.Google Scholar
- Adams SH, Harpster T: 911 homicide calls and statement analysis. FBI Law Enforce. Bull. 2008, 77: 22–31.Google Scholar
- Harpster T, Adams SH, Jarvis JP: Analyzing 911 homicide calls for indicators of guilt or innocence. Homicide Stud. 2009, 13: 69–93. 10.1177/1088767908328073View ArticleGoogle Scholar
- Zhou L, Burgoon JK, Twitchell DP, Qin T, Nunamaker JF Jr: A comparison of classification methods for predicting deception in computer-mediated communication. J. Manag. Inf. Syst. 2004, 20: 139–165.Google Scholar
- Zhou L, Burgoon JK, Nunamaker JF Jr, Twitchell DP: Automating linguistics based cues for detecting deception in text based asynchronous computer mediated communication: an empirical investigation. Group Decis. Negotiation 2004, 13: 81–106. 10.1023/B:GRUP.0000011944.62889.6fView ArticleGoogle Scholar
- Mann S, Vrij A, Bull R: Suspects, lies, and videotape: an analysis of authentic high-stake liars. Law Hum. Behav. 2002, 26: 365–376. 10.1023/A:1015332606792View ArticleGoogle Scholar
- Fuller CM, Biros DP, Adkins M, Burgoon JK, Nunamaker JF Jr, Coulon S: Detecting Deception in Person-Of-Interest Statements. In Lecture Notes in Computer Science. Springer, Berlin/Heidelberg; 2006:504–509.Google Scholar
- CM Fuller, DP Biros, D Delen, Exploration of Feature Selection and Advanced Classification Models for High-Stakes Deception Detection, in Proceedings of the 41st Hawaii International Conference on System Sciences (HICSS) (Waikoloa, Big Island, HI, 2008) CM Fuller, DP Biros, D Delen, Exploration of Feature Selection and Advanced Classification Models for High-Stakes Deception Detection, in Proceedings of the 41st Hawaii International Conference on System Sciences (HICSS) (Waikoloa, Big Island, HI, 2008)Google Scholar
- Fuller CM, Biros DP, Wilson RL: Decision support for determining veracity via linguistic-based cues. Decis. Support. Syst. 2008, 46: 695–703. 10.1016/j.dss.2008.11.001View ArticleGoogle Scholar
- Jensen ML, Bessarabova E, Adame B, Burgoon JK, Slowik SM: Deceptive language by innocent and guilty criminal suspects: the influence of dominance, question, and guilt on interview responses. J. Lang. Soc. Psychol. 2011, 30: 357–375. 10.1177/0261927X11416201View ArticleGoogle Scholar
- Sandoval VA: Strategies to avoid interview contamination. FBI Law Enforce. Bull. 2003, 72: 1–12.Google Scholar
- Nijholt A, Arkin RC, Brault S, Kulpa R, Multon F, Bideau B, Traum DR, Hung H, Santos E Jr, Li D, et al.: Trends and controversies. IEEE Intell. Syst. 2012, 27: 60–75. 10.1109/MIS.2012.116View ArticleGoogle Scholar
- Humpherys S, Moffitt KC, Burns MB, Burgoon JK, Felix WF: Identification of fraudulent financial statements using linguistic credibility analysis. Decis. Support. Syst. 2011, 50: 585–594. 10.1016/j.dss.2010.08.009View ArticleGoogle Scholar
- KC Moffitt, MB Burns, What does that mean? Investigating Obfuscation and Readability Cues as Indicators of Deception in Fraudulent Financial Reports, in Fifteenth Americas Conference on Information Systems (San Francisco, CA, 2009), p. 2009 KC Moffitt, MB Burns, What does that mean? Investigating Obfuscation and Readability Cues as Indicators of Deception in Fraudulent Financial Reports, in Fifteenth Americas Conference on Information Systems (San Francisco, CA, 2009), p. 2009Google Scholar
- Hancock JT, Curry LE, Goorha S, Woodworth M: On lying and being lied to: a linguistic analysis of deception in computer-mediated communication. Discourse Process. 2008, 45: 1–23. 10.1080/01638530701739181View ArticleGoogle Scholar
- Newman ML, Pennebaker JW, Berry DS, Richards JM: Lying words: predicting deception from linguistic styles. Personal. Soc. Psychol. Bull. 2003, 29: 665–675. 10.1177/0146167203029005010View ArticleGoogle Scholar
- Pennebaker JW, Francis ME, Booth RJ: Linguistic Inquiry and Word Count. Erlbaum Publishers, Mahway, NJ; 2001.Google Scholar
- Tausczik YR, Pennebaker JW: The psychological meaning of words: LIWC and computerized text analysis methods. J. Lang. Soc. Psychol. 2010, 29: 24–54. 10.1177/0261927X09351676View ArticleGoogle Scholar
- Zuckerman M, DePaulo BM, Rosenthal R: Verbal and Nonverbal Communication of Deception. In Advances in Experimental Social Psychology. Edited by: Berkowitz L. Academic, New York, NY; 1981:60.Google Scholar
- Buller DB, Burgoon JK: Interpersonal deception theory. Commun. Theory 1996, 6: 203–242. 10.1111/j.1468-2885.1996.tb00127.xView ArticleGoogle Scholar
- McCornack SA: Information manipulation theory. Commun. Monogr. 1992, 59: 1–16. 10.1080/03637759209376245View ArticleGoogle Scholar
- Johnson M, Raye C: Reality monitoring. Psychol. Rev. 1981, 88: 67–85. 10.1037/0033-295X.88.1.67View ArticleGoogle Scholar
- Burgoon JK, Buller DB: Interpersonal deception: III: effects of deceit on perceived communication and nonverbal behavior dynamics. J. Nonverbal Behav. 1994, 18: 155–184. 10.1007/BF02170076View ArticleGoogle Scholar
- Fuller CM, Biros DP, Burgoon JK, Adkins M, Twitchell DP: An analysis of text-based deception detection tools. AMCIS 2006 Proceedings 2006. Paper 418 Paper 418Google Scholar
- Qin T, Burgoon JK, Blair JP, Nunamaker JF Jr: Modality Effects in Deception Detection and Applications in Automatic Deception Detection. Proceedings of the 38th Hawaii International Conference on Systems Sciences (HICSS) 2005.Google Scholar
- Zhou L, Twitchell DP, Qin T, Burgoon JK, Nunamaker JF Jr: An Exploratory Study into Deception Detection in Text-Based Computer-Mediated Communication. Proceedings of the 36th Hawaii International Conference on Systems Sciences (HICSS '03) 2003.Google Scholar
- DePaulo BM, Lindsay JJ, Malone BE, Muhlenbruck L, Charlton K, Cooper H: Cues to deception. Psychol. Bull. 2003, 129: 74–118. 10.1037/0033-2909.129.1.74View ArticleGoogle Scholar
- Zhou L, Burgoon JK, Twitchell DP: A Longitudinal Analysis of Language Behavior of Deception in E-Mail. In Intelligence and Security Informatics: First NSF/NIJ Symposium, ISI 2003, Tucson, AZ, USA, June 2–3, 2003 Proceedings. Edited by: Chen H, Moore R, Zeng D, Leavitt J. Springer, Berlin /Heidelberg; 2003:102–110. Lecture Notes in Computer Science Lecture Notes in Computer Science 10.1007/3-540-44853-5_8View ArticleGoogle Scholar
- Driscoll LN: A validity assessment of written statements from suspects in criminal investigations using the scan technique. Police Stud: Int. Rev. Police Dev. 1994, 17: 77–78.Google Scholar
- Olsson J: Forensic Linguistics: An Introduction to Language, Crime, and Law. Continuum International Publishing Group, London; 2004.Google Scholar
- Tracy SJ: When questioning turns to face threat: an interactional sensitivity in 911 call-taking. West. J. Commun. 2002, 66: 129–157. 10.1080/10570310209374730View ArticleGoogle Scholar
- Ekman P: Mistakes when deceiving. Ann. N. Y. Acad. Sci. 1980, 364: 269–278. 10.1111/j.1749-6632.1981.tb34479.xView ArticleGoogle Scholar
- Vrij A: Detecting Lies and Deceit: Pitfalls and Opportunities. Wiley, Chichester, West Sussex, England; 2008.Google Scholar
- Vrij A, Fisher R, Mann S, Leal S: A cognitive load approach to lie detection. J. Investig. Psychol. Offender Profiling 2008, 5: 39–43. 10.1002/jip.82View ArticleGoogle Scholar
- Rassin E, van der Heijden S: Appearing credible? Swearing helps! Psychol. Crime Law 2005, 11: 177–182. 10.1080/106831605160512331329952View ArticleGoogle Scholar
- Sumner M, Frank E, Hall M: Speeding up Logistical Model Tree Induction. 9th European Conference on Principles and Practice of Knowledge Discovery in Databases 2005, 675–683.Google Scholar
- Rassin E, Muris P: Why do women swear? An explanation of reasons for and perceived efficacy of swearing in Dutch female students. Personal. Individ. Differ. 2005, 38: 1669–1674. 10.1016/j.paid.2004.09.022View ArticleGoogle Scholar
- Jay T, Janschewitz K: The pragmatics of swearing. J. Politeness Res. Lang. Beh. Cult. 2008, 4: 267–288.Google Scholar
- Crewson PE: Comparative analysis of polygraph with other screening and diagnostic tools. No. DODPI01-R-0003. Research Support Service, Ashburn, VA; 2001.Google Scholar
- Honts CR, Raskin DC: A field study of the validity of the directed lie control question. J. Police Sci. Adm. 1988, 16: 56–61.Google Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.