Dynamic activity involving social networks often has distinctive temporal patterns that can be exploited in situations involving incomplete information. Gang rivalry networks, in particular, display a high degree of temporal clustering of activity associated with retaliatory behavior. A recent study of a Los Angeles gang network shows that known gang activity between rivals can be modeled as a self-exciting point process on an edge of the rivalry network. In real-life situations, data is incomplete and law-enforcement agencies may not know which gang is involved. However, even when gang activity is highly stochastic, localized excitations in parts of the known dataset can help identify gangs responsible for unsolved crimes. Previous work successfully incorporated the observed clustering in time of the data to identify gangs responsible for unsolved crimes. However, the authors assumed that the parameters of the model are known, when in reality they have to be estimated from the data itself. We propose an iterative method that simultaneously estimates the parameters in the underlying point process and assigns weights to the unknown events with a directly calculable score function. The results of the estimation, weights, error propagation, convergence and runtime are presented.
In this work we focus our attention on data sets of events involving rival gangs on a social network. Each event in the data set corresponds to a crime that occurs at a specified time and involves a pair of rival gangs. A subset of these events are unsolved crimes in which one or both of the rival gangs is not known. The method developed in this paper could be broadly applied to any social network involving activities in time between pairs of nodes on the network. However the interest in the problem came about by examining data from the Hollenbeck Division of the Los Angeles Police Department, home to 29 street gangs with a well-known rivalry network -.
Unlike other methods used to address incomplete data relating to social networks [4, 5], the question at hand is not if a rivalry exists, but rather to which rivalry a violent event belongs. This structure of between gang rivalries can be viewed as a social network  often embedded in space [1, 2]. Violent events involving gangs tend to be dyadic, and so we can formulate these events as a realization of a stochastic process occurring on the edges of the rivalry network. For each edge in the network there exists a different stochastic process. In our analysis however, we use identical parameters to generate synthetic data. The method does not assume that the underlying parameters generating each process are identical.
The first step to inferring the affiliation of the violent events is to understand the underlying stochastic process. This requires us to capture the behavior of criminal activity through computational means, much like in -. Recently methods have been proposed in the literature to mathematically model gang violence. The authors in  employ an agent-based model to investigate the geographic influences in the formation of the gang rivalry structure observed in Hollenbeck. These authors consider the long-term structure of the rivalry network embedded in space. In terms of the rivalry violence, a shorter timescale must be considered.
Violence among gangs exhibits retaliatory behavior . In other words, given an event has happened between two gangs, the likelihood that another event will happen shortly after is increased. A problem such as this is modeled naturally by a self-exciting point process. It is interesting to note that these models were first used to analyze earthquakes -. Since then, they have been used to model financial contagion in credit markets [16, 17], viral videos on the web , terrorist activity in Indonesia , and the spread of infectious disease . In this analysis we limit the scope of our model to include time only, thus providing a baseline model.
The authors in  and  have successfully modeled the pairwise gang violence as a Hawkes process . All of the events are associated with exactly one rivalry, or edge of a social network. The violence on each edge, k, is assumed to have the conditional intensity
In this Hawkes process, the intensity λk(t|Hτ,k) depends on the history of the process , where Mkis the number of events for process, k. In this framework, the window of time, [0,T], observed for each process in the network is the same. However, the number of events in each process, Mk, is stochastic, and therefore varies from process to process. In practice the final time, T, is determined by the end of the data collection period. Further, the edges of the window introduce boundary effects that are adjusted for in the parameter approximation, see Equations 10 and 11.
The background rate of the process is defined by the constant μk. In the context of gang rivalries, background events can be thought of as random occurrences between rival gangs that trigger retaliatory events. The expected number of offspring for any event is determined by the constant αk, and the decay of the intensity back to the background rate is ωk. Offspring events, in this context, could be interpreted as retaliatory events. Larger values for μkand αkproduce more background and offspring events respectively. Larger values of ωk do not influence the total number of events, but rather the amount of clustering in time.
The authors of  produce a mathematical framework to solve the incomplete data problem observed in gang violence data sets. In their work they use an optimization strategy that computes the weights to infer the rivalry affiliation of the incomplete data. In this formulation the authors prove that their optimization has a unique solution under mild constraints. This is substantial contribution in inferring the affiliation of the unknown violent events. However, the authors of  assume that the process parameters are known, an assumption that is often not feasible in practice. Further, finding the weights requires solving a computationally expensive optimization problem.
We propose an iterative method that (A) estimates the process parameters assuming the data is generated by the process defined by Equation 1 and (B) infers the process affiliation of simulated data via a direct method of computation. We iterate between (A) and (B) until the estimates for the unknown events converge. We call this the Estimate & Score Algorithm (ESA). The details of the ESA are described in Section “The Estimation & Score Algorithm (ESA)”. The ESA is tested on simulated data in Section “Results”, with analysis of the estimation of the parameters in the presence of incomplete data (see Subsection “Estimation analysis”) and comparison of the proposed score functions with that of the Stomakhin-Short-Bertozzi (SSB) method in  (see Subsection “Updating Weights analysis”). In Subsection “Runtime Analysis” there is an analysis of the runtime between the Stomakhin-Short-Bertozzi and the Forward Backward score functions used to update the weights (see Subsection “Runtime Analysis”). Subsection “Convergence Results” contains an analysis of the convergence of the Estimation & Score Algorithm. This method solves the more realistic problem of estimating the process and the weights. Further, the computation for the weight updates is more direct and therefore avoids performing the costly optimization scheme used in . This is a novel piece of work with many exciting extensions. A final discussion of the results and future work is presented in Section “Discussion and Future Work”. As in  we do not use field data in this paper, rather we generated point process data using similar parameters as observed in the field data for Hollenbeck . By using simulated data to test the algorithms we have actual ground truth evaluate the performance of the method.
The data is assumed to lie on a known social network containing K processes, where each of the K processes is a pairwise rivalry between two gangs. From this set of events, there are a total of N events where the time is known, but the processes affiliation is not known. These events are referred to as unknown events. Each of the N unknown events are placed into each of the K processes. Since the process affiliation is not known for all of the events in the network, each event is given an associated weight, Si,k. Here Si,kis the i th element of the k th process. If the event is known Si,k=1. If Si,k is unknown then it is assigned a number between 0 and 1 by our algorithm. We enforce the constraint that .
A simplified representation of our problem formulation can be found in Figure 1. The known events are represented by circles and the unknown event is represented by a triangle. Here we can see that since we do not know the affiliation of the triangle event, it is placed in all of the other processes. We emphasize that this represents our lack of information about which rivalry it belongs to.
As indicated in Figure 1, for each process in the network events ei,k are indexed by increasing time, . Ordering the events in such a way has the consequence that the first unknown element in time, for example, may have different indexes for different processes. In Figure 1 the triangle index in first process is the third event, e1,3. However the triangle in the K th process is the second event, eK,2. One can easily keep track of the local index of a unknown event for each process.
The Estimation & Score Algorithm (ESA)
The proposed Estimation & Score Algorithm can be broken into three basic stages: initialization, parameter estimation, and updating the weights. This method is succinctly described in Figure 2.
For this paper, there were two ways of initializing the Estimate & Score Algorithm. The first is used to infer rivalry affiliation given field data. After importing the data, the unknown events are identified and placed into each of the of the K processes. The weights, Si,k, must also be initialized. If the event is known, then Si,k=1. If the event is unknown then .
An alternate initialization utilizes simulated data in order to test the components of the Estimate & Score Algorithm. In this case, data is generated from K independent Hawkes processes with given μk, αk, and ωk. From these data, choose N events at random from the network to mark as unknown. Place these N unknown events into each of the other processes. Initialize the weights such that for known events Si,k=1 and for unknown events Si,k=1/K. This initialization process is used in this paper to test the method and produce the results in Section “Results”.
In the presence of no unknown events, there are both parametric  and nonparametric - ways to model the underlying stochastic process on each edge of the social network. For this work, we chose a parametric form for the triggering density to validate the model but the results could easily be extended to the nonparametric case. We note that, as is usual with nonparametric estimates, speed would be compromised for the sake of flexibility.
For this paper, the data is assumed to be a realization of Equation 1, where the parameters are estimated using a method similar to the Expectation Maximization (EM) algorithm . An EM-like approach is taken because of the branching structure present in a Hawkes process. In such a process each event can be associated with a background or response event. However, given a realization from this process it is not immediately obvious whether an event is a background or response event. We can view this information as a hidden variable that we must estimate. In this way, every event in each of the K processes is assigned a probability . The probability that event i is a background event is denoted , and probability that event i caused event j is denoted . This assumes that ti<tj. From this EM estimation, the approximation for each of the variables is altered to include the weights for the unknown events. In fact, in the case where all the events are known, the estimation formulas are the same. This section derives the EM estimates when in the presence of incomplete data.
The classical log-likelihood function for a general point process with a fixed window [0,T] is
Incorporating the branching structure into the log-likelihood function, the event association is added as a random variable, χi,j such that
This branching allows us to separate those events associated with the background μk and the response . This leads to the altered log-likelihood function
Taking the expectation of ℓk(Hτ,k|μk,αk,ωk) with respect to χi,jresults in
In the EM algorithm, the quantity Eχ[ℓk(Hτ,k|μk,αk,ωk)] is maximized with respect to each of the variables μk,αk,ωk given the data Hτ,k. This leads to the EM estimates
Where is defined by
for ti<tj. The EM algorithm then becomes a matter of iterating between estimating the probabilities and the parameters. It has been proven that this algorithm will converge under mild assumptions . Further, Equation 6 adjusts for boundary effects.
In the presence of events with unknown process affiliation in the network, we assign weights to the contribution of each event to the log-likelihood function. Specifically, each of the unknown events in process k have a weight Si,k, such that ∑kSi,k=1. For the known events Si,k=1. These weights are incorporated for each process via
Note that Lk(Hτ,k|μk,αk,ωk) is no longer an EM log likelihood in the presence of unknown data. Maximizing Lk(Hτ,k|μk,αk,ωk) with respect to each of the parameters the estimates become
When all of the events are known, i.e. Si,k=1 when unknown event i, k belongs to process k and is zero otherwise, these estimates become identical to the EM parameter estimates.
At the start of the Estimation & Score algorithm all of the weights for the unknown events are Si,k=1/K. Once the parameters are estimated using the altered EM algorithm described in Equation 11, the weights, Si,k, are updated, see Figure 2. Here we present four different score functions and the Stomakhin-Short-Bertozzi method , used to define, qi,k, the intermediate process affiliation. Each of these score functions synthesize information from different portions of the data set. Given an event early in the data set, a score function that uses future events would be ideal. On the other hand, for later events a score function using previous events is desired. Similar considerations should be made if there are portion of the data with more incomplete data. After all of these intermediate weights, qi,k, have been calculated, they are re-normalized as a probability via . For simplicity we consider a response function of the form, .
Ratio Score Function
The Ratio score function considers the ratio of the background rate μk and the sum of all the future events, ∑i<jgk(tj−ti). Mathematically the score is determined by
Lambda Score Function
The Lambda score function uses only previous information by taking the ratio of the intensities evaluated at the unknown event time ti.
This method is motivated by the Hawkes process defined in Equation 1.
Probability Score Function
The Probability score function uses the approximation of the branching structure of the underlying process. The idea behind this method is events that are background events with no corresponding response events should not belong in the process. An event that is a background with many response events or an event that is a response to another event should be part of that process.
Forward Backward Score Function
This method is the ratio of the summation of the response for the events in the future and the past, over the background rate μk.
The Estimation & Score Algorithm is tested for accuracy on simulated data from the Hawkes process defined in Equation 1. An analysis of the parameter estimation method outlined in Subsection “Parameter Estimation” is conducted in Subsection “Estimation analysis”. A comparison of the score functions when assuming the true parameters is found in Subsection “Updating Weights analysis”. Subsection “Runtime Analysis” provides a comparison of the runtime between the Forward Backward score function and the Stomakhin-Short-Bertozzi method. A example of convergence of the Estimate & Score Algorithm is provided in Subsection “Convergence Results”.
There are many ways we could allow the unknown events to influence our estimates of the underlying parameters for each process. There are two extremes. On the one hand, we could exclude all of the unknown events from the parameter estimation. This would be equivalent to setting the Si,k=0 for all unknown events. On the other hand, we could include all of the unknown events in the estimation of the parameters for each process. This would be equivalent to letting Si,k=1 for all i and k. Another possible estimation method is some combination of these two. We propose this as a way of allowing the unaffiliated events to play some role in the estimation process. The naive choice is allowing each event to play the same role in each process. This amounts to setting Si,k=1/K for the unknown events. We compare these three choices to the estimations obtained by the Estimate & Score Algorithm (ESA) using the Forward Backward score function. Finally, we want to compare all four of these possible estimation techniques to the best we could possibly do. In this case, that would mean we knew all the affiliations for the events (i.e. there are no unknown events).
Figures 3, 4, 5 displays the results for the μk, αk, and ωk estimates for the five cases: Si,k=0 for unknown events (dash-triangle), Si,k=1 for unknown events (dash-square), Si,k=1/K for unknown events (dash-x), the results using ESA (dash-circle), and the estimates you get when you know all the affiliations for the unknown events (solid). These results with standard deviations are displayed in Tables 1, 2, 3. In each of the three figures, the estimates are plotted vs the number of unknown events. Each network has five processes with the true parameters μk=0.01, ωk=0.1, and αk=0.5. Different networks are created with 15, 30, 45, 60, and 75 unknown events. We estimate the parameters using each of the five methods explained above. This procedure is repeated 100 times with different random seed values and then the average estimate is calculated.
Notice in the estimates for μkin Figure 3 and Table 1, the ESA performs the best compared to the true value and has only a slight reduction in accuracy as the number of unknown events increases. On average the other three estimates seem to degrade more rapidly as the number of unknown events increases. When Si,k=1, the estimates for μk are far above the true value and growing as the number of unknown events increases. This follows from the fact that letting Si,k=1 means we are effectively adding events to the network. Take the case when K=2. Assume that each process has 1000 events, and there are 100 unknown events from each process. When we estimate the parameters for the first process, we will use the 900 events we know plus the 200 unknown events from the network. We will get the identical number of events in our estimation for process two. This creates 200 new events and thus biases the estimates for μk. This motivated the idea of equal weighting for each unknown event, and that choice is validated by the estimates for μk. A similar argument shows why Si,k=0 (i.e. ignoring all the unknown events) has the lowest estimate for μk at each level of incomplete data.
In the estimates for the branching ratio αk, the ESA on average yields the best estimates and maintains its accuracy in the presence of unknown events. It is interesting to note that equal weighting performs worse here than if we let Si,k=0 for all unknown events. Using the ESA overcomes this drawback. Again, setting Si,k=1 for all unknown events performs the worst. This could stem from the fact that most of the unknown events are being labeled background and thus this estimation technique underestimates the branching ratio because fewer events are considered offspring. Notice that the estimate for ESA (dash-circle) tracks the best possible estimate (solid) well while the other three start to trail off as more and more information is labeled as unknown.
Finally, in Figure 5 and Table 1, it is shown that the ESA estimate (dash-circle) for ωktracks the behavior of the best estimate (solid) closer than the other methods. Including all of the unknown events (dash-square) provides the poorest estimate for ωk. For the other three estimation techniques we see that they are all comparable.
Updating Weights analysis
To understand the strengths and weaknesses of each of the five score functions, defined in Subsection “Updating weights”, the score functions were evaluated for 100 incomplete events using the true values for μkαk, and ωk when taking the Top 1, Top 2, and Top 3 best inferences. For comparison to , the true parameters were taken to be μk=0.01, ωk=0.1, and αk=0.5. Due to the stochastic nature of the processes, for each level of process number 100 random networks were tested. The average results of this analysis are found in Figure 6. The number correctly identified by the each of the score functions is on the vertical axis. The horizontal axis displays the number of processes in the network.
From Figure 6 it is clear that the Stomakhin-Short-Bertozzi score function in solid dark blue, and the Forward Backward score function (cyan dashed diamond) perform nearly identically when looking at the Top 1, Top 2, and Top 3 inferences. These functions look both forward and backward in time from the incomplete event, and are therefore able to identify clusters of events in time. The Probability (black dashed asterisk) and Ratio (solid green square) score functions don’t do nearly as well the Stomakhin-Short-Bertozzi and Forward Backward score functions, but better than the Lambda (magenta dashed circle) score function. The Lambda score function appears to perform close to chance (dark green solid plus) for the Top 1, Top 2, and Top 3 inferred process affiliation. Due to the success of the Forward Backward score function and the Stomakhin-Short-Bertozzi method, only these are used for further analysis.
The analysis comparing the score functions assumed that the true parameters were known. However, when applying this method in practice there will be error in the estimated parameters. This estimation error will propagate through to the score functions. To understand how deviations of the estimated parameters influence the score functions pairwise combinations of the parameters were increased and decreased by 90% from the target values μk=0.01, ωk=0.1, and αk=0.5 in 10% increments. In particular the Forward Backward and SSB score functions are computed for pairwise combinations of μ in the range of [0.001,0.019], ω in the range of [0.01,0.19], and α in the range of [0.05,0.95]. Further, in these pairwise combinations, the third parameter is kept at the target value. Notice that a 90% change is larger than the errors observed in the parameter estimates in Subsection “Estimation analysis”.
To examine the propagation of errors of the parameters to the score functions one event from a network with 10 processes is chosen to be unknown. The score function S1,true with the target parameters, μk=0.01, ωk=0.1, and αk=0.5 for the true process is calculated. Then, on the same network, the parameters are offset by
The offset score function is calculated from these offset parameters. The difference between is taken for each pairwise combination of parameters. Again, due to the stochastic nature of the processes, each analysis was done for 100 runs and the average difference in score functions is recorded. The results of this analysis are displayed in Figure 7 with those of the Forward Backward score function (left), and those for the Stomakhin-Short-Bertozzi score function (right). In general the Stomakhin-Short-Bertozzi score function is more sensitive to the changes than the Forward Backward score functions for the μk and αkparameters. Changes in the Forward Backward score functions are minimal for most changes of parameters except for small values of ωk. As ωk decreases then the approximated Forward Backward score function decreases, causing a positive difference. As seen in Subsection “Estimation analysis”, Figure 5, when estimating ωk, there is a tendency to over, not under estimate the parameter, and so this does not appear to occur within these parameters. The changes in the Stomakhin-Short-Bertozzi score function depend on all of the pairwise changes of the parameters. As μkincreases the computed Stomakhin-Short-Bertozzi decreases. On the other hand, as ωk or αkincrease the score function increases. This analysis shows that though the Stomakhin-Short-Bertozzi method and the Forward Backward score functions preform similarly when the parameters are known exactly, under the influence of estimation error the Stomakhin-Short-Bertozzi score function varies more than the Forward Backward score function.
Though the Forward Backward score function and the SSB method produce comparable results in terms of accuracy, there is a sizable difference in the time it takes to update the weights using these methods.
The Forward Backward score function is designed to be direct, meaning calculates the weights using available information without need for iteration. The Stomakhin-Short-Bertozzi method, however, determines the weight by solving a optimization problem. A closed form solution for the maximized weights is not known to these authors, so the weights are found by numerically approximating the weights that maximize Equation 14. In the implementation of the Stomakhin-Short-Bertozzi we employ a gradient ascent method which requires 4-11 iterations to reach convergence with a tolerance of 0.001. The direct methods, Forward Backward, Probability, Ratio, and Lambda score functions, are on the same order of operations as one iteration of the gradient ascent used to solve Equation 14. Specifically, one iteration of the gradient ascent method and calculating the direct score functions are O(N·K·M) where N is the number of unknown events, K is the number of processes and M is the expected number of events in process k. The expected number of events in process k can be further analyzed via,
The run time of both the Forward Backward function and the Stomakhin-Short-Bertozzi method are empirically examined in Figure 8. Both score functions were calculated with 20 networks for each level of number unknown events and number of processes with the known parameter values of μk=0.01, ωk=0.1, and αk=0.5. All of the run times are calculated in milliseconds. It can be seen that the average run time needed to compute the Forward Backward function at every level of N and K is substantially less than that of the Stomakhin-Short-Bertozzi method. Also, it is clear from this figure that the time needed to calculate both of these methods increases as N and K increase.
The Estimation & Score Algorithm converges quickly when either the Forward Backward score function or Stomakhin-Short-Bertozzi method are used. Figure 9 displays the parameter estimates for a typical run of the Estimation & Score Algorithm for both the Forward Backward (left) and Stomakhin-Short-Bertozzi (right). Both score functions produce qualitatively similar results, and it appears that the rate of convergence is comparable for both cases. The estimated weights for one unknown data event for this typical run versus the iteration for each process are plotted in Figure 10. The weights plotted are obtained from the Forward Backward score function (left) and the Stomakhin-Short-Bertozzi method (right). It is interesting to note that both methods of weighting choose the same process affiliation as the most likely. Further tests were conducted with a variable initial weighting. These runs showed similar behavior as initializing the Estimate & Score Algorithm with Si,k=1/K, implying that the Estimate & Score Algorithm is robust to small perturbations of the initial weighting.
Discussion and Future Work
In this paper we propose an effective method for simultaneously estimating the parameters and assigning process affiliation in case of incomplete field data from self-exciting point processes on a network. This problem comes from the demand for law enforcement agencies to identify gang affiliation in the case of unsolved crimes in an area of highly complex gang rivalry activity. We present a new framework we name the Estimate & Score Algorithm for possible application to field data. By testing the method on simulated datasets we can understand its performance features and liabilities. The method is an iterative procedure in which process parameters are estimated alternately with the calculation of network affiliation probabilities. We identify several useful ‘score functions’ for calculating the network affiliations. We also compare the use of unknown events in the parameter estimation. One upshot of our analysis is that the inclusion of unknown events may increase the accuracy of the parameter estimation. Several score functions are considered and the Forward Backward score function shows the most promise with comparable results to that of the Stomakhin-Short-Bertozzi method of  in the parameter regime tested. The score function calculation is a direct method that does not rely on solving a variational problem, and thus is more computationally efficient than .
For future work, space often plays a role in understanding criminal activity [8, 30]-. Further, criminal behavior has non-random structure and can often be framed in terms of routine activity theory [34, 35]. In the case of gang violence, there is a strong spatial component [1, 10, 36]. One can extend the Estimate & Score Algorithm to include space. There is a precedence in the earthquake literature of adding space to self-exciting point processes [13, 15], however, in the case of gang violence, the spatial response may be different. Instead of retaliatory events clustering around prior events, it appears that the data is clustered around regions in space. A spatial model similar to that of  could be employed, where the triggering density in space is related to their respective gang set-space, or center of activity . Statistically when modeling spatial point processes one needs to tease out the difference between hot spots due to risk heterogeneity versus event dependence. The data given will be one realization of the underlying process, however using techniques such as prototyping , one could potentially reformulate the data into multiple realization of the same process and distinguish between these two phenomena.
There are other factors in the data that can be fused into the model, though more analysis would be required. For example, in earthquake modeling the magnitude of the earthquake is often included. To include such a factor to the intensity λk(t|Hτ,k) one would need to determine a numerical metric to define the impact of each event type. This is not a straightforward task and would require further investigation. Extending this model in this way could allow for the inclusion of events involving tagging, or other low level gang crimes, which could be a precursor to more extreme violent interactions between gangs. Including this data is outside of the scope of the current model but has a strong potential to enrich the overall data set allowing for better analysis.
It is important to note that there are other methods to approximate the underlying form of the self exciting process. For example the authors in  consider the general form of the intensity function λk(t|Hτ,k) to be
Using a non-parametric method, they are able to approximated the background function μ(t) and the response function g(t) for a broader class of functions. In this paper, the data was assumed to come from a Hawkes process with constant background rate and an exponential response to previous events. There are cases where the background rate is not constant . Further it is conceivable that the response function could be of a form other than an exponential decay. In this circumstances, the model for λ(t|Hτ) in Equation 1 would not be appropriate.
Finally, this method has a great potential in the field of policing. Once such a model has been calibrated correctly, the Estimation & Score Algorithm using the quicker Forward Backward score function can be used to infer the gang association in real time, while the investigation is on going. Given an accurate model of the underlying process, such a method could identify rivalries that have heightened activity.
RH led in the construction of the Estimation & Score algorithm, coding, analysis of the results and writing of the manuscript. EL contributed to the construction of the Estimation & Score algorithm, coding, analysis of the results, and writing of the manuscript. AB contributed to the analysis of the results, editing of the manuscript, and conception and design of the Estimation & Score algorithm. All authors read and approved the final manuscript.
Estimation & Score Algorithm
Radilm SM, Flint C, Tita GE: Spatializing Social Networks: Using Social Network Analysis to Investigate Geographies of Gang Rivalry, Territoriality, and Violence in Los Angeles. Annals of the Association of American Geographers 2010, 100(2):307–326. http://www.tandfonline.com/doi/abs/10.1080/00045600903550428
Hoff P: Multiplicative latent factor models for description and prediction of social networks. Computational & Mathematical Organization Theory 2009, 15: 261–272. http://dx.doi.org/10.1007/s10588–008–9040–4. [10.1007/s10588–008–9040–4]. [10.1007/s10588-008-9040-4]. 10.1007/s10588-008-9040-4
Brantingham P, Glässer U, Jackson P, Vajihollahi M: Modeling Criminal Activity in Urban Landscapes. In Mathematical Methods in Counterterrorism. Edited by: Memon N, David Farley J, Hicks DL, Rosenorn T. Springer Vienna; 2009:9–31. http://dx.doi.org/10.1007/978–3-211–09442–6\_2. [10.1007/978–3-211–09442–6_2]
Brantingham P, Glasser U, Kinney B, Singh K, Vajihollahi M: A computational model for simulating spatial aspects of crime in urban environments. In Systems, Man and Cybernetics, 2005 IEEE International Conference on Volume 4. IEEE; 2005:3667–3674.
Hegemann RA, Smith LM, Barbaro AB, Bertozzi AL, Reid SE, Tita GE: Geographical influences of an emerging network of gang rivalries. Physica A: Statistical Mechanics and its Applications 2011, 390(21–22):3894–3914. http://www.sciencedirect.com/science/article/pii/S037843711100447X
Veen A, Schoenberg FP: Estimation of Space–Time Branching Process Models in Seismology Using an EM Type Algorithm. Journal of the American Statistical Association 2008, 103(482):614–624. http://pubs.amstat.org/doi/abs/10.1198/016214508000000148
Ogata Y: Space-Time Point-Process Models for Earthquake Occurrences. Annals of the Institute of Statistical Mathematics 1998, 50: 379–402. http://dx.doi.org/10.1023/A:1003403601725. [10.1023/A:1003403601725]
Crane R, Sornette D: Robust dynamic classes revealed by measuring the response function of a social system. Proceedings of the National Academy of Sciences 2008, 105(41):15649–15653. http://www.pnas.org/content/105/41/15649.abstract 10.1073/pnas.0803685105
Mohler GO, Short MB, Brantingham PJ, Schoenberg FP, Tita GE: Self-Exciting Point Process Modeling of Crime. Journal of the American Statistical Association 2011, 106(493):100–108. http://pubs.amstat.org/doi/abs/10.1198/jasa.2011.ap09546 10.1198/jasa.2011.ap09546
Stomakhin A, Short MB, Bertozzi AL: Reconstruction of missing data in social networks based on temporal patterns of interactions. Inverse Problems 2011, 27(11):115013. http://stacks.iop.org/0266–5611/27/i=11/a=115013 10.1088/0266-5611/27/11/115013
Sornette D, Utkin S: Limits of declustering methods for disentangling exogenous from endogenous events in time series with foreshocks, main shocks, and aftershocks. Physical Review E 2009, 79(6):61110.
Block R: Gang Activity and Overall Levels of Crime: A New Mapping Tool for Defining Areas of Gang Activity Using Police Records. Journal of Quantitative Criminology 2000, 16: 369–383. http://dx.doi.org/10.1023/A:1007579007011. [10.1023/A:1007579007011]
Nichols K, Schoenberg F, Keeley J, Bray A, Diez D: The application of prototype point processes for the summary and description of California wildfires. Journal of Time Series Analysis 2011, 32(4):420–429. 10.1111/j.1467-9892.2011.00734.x
This work was supported by NSF grant DMS-0968309, ONR grant N000141010221, ARO grants W911NF1010472 and W911NF1110332, and AFOSR MURI grant FA9550-10-1-0569. Further, we would like to thank Martin Short, Jeff Brantingham, and George Mohler for their helpful comments and conversations.
Authors and Affiliations
Department of Mathematics, University of California Los Angeles, 520 Portola Plaza, Los Angeles, CA, USA
Rachel A Hegemann, Erik A Lewis & Andrea L Bertozzi
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Hegemann, R.A., Lewis, E.A. & Bertozzi, A.L. An “Estimate & Score Algorithm” for simultaneous parameter estimation and reconstruction of incomplete data on social networks.
Secur Inform2, 1 (2013). https://doi.org/10.1186/2190-8532-2-1