<?xml version="1.0" encoding="UTF-8"?><?xml-model type="application/xml-dtd" href="https://jats.nlm.nih.gov/publishing/1.3/JATS-journalpublishing1-3.dtd"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "https://jats.nlm.nih.gov/publishing/1.3/JATS-journalpublishing1-3.dtd">
<article xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" specific-use="Marcalyc 1.3" dtd-version="1.3" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="index">7261</journal-id>
<journal-title-group>
<journal-title specific-use="original" xml:lang="es">Avances en Ciencias e Ingenierías</journal-title>
<abbrev-journal-title abbrev-type="publisher" xml:lang="es">ACI Avances en Ciencias e Ingenierías</abbrev-journal-title>
</journal-title-group>
<issn pub-type="ppub">1390-5384</issn>
<issn pub-type="epub">2528-7788</issn>
<issn-l>1390-5384</issn-l>
<publisher>
<publisher-name>Universidad San Francisco de Quito</publisher-name>
<publisher-loc>
<country>Ecuador</country>
<email>avances@usfq.edu.ec</email>
</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="art-access-id" specific-use="redalyc">726182578011</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artículos</subject>
</subj-group>
</article-categories>
<title-group>
<article-title xml:lang="en">A temporal approach to urban crime forecasting using recurrent neural networks</article-title>
<trans-title-group>
<trans-title xml:lang="es">
<bold>Un enfoque temporal para la predicción del crimen urbano usando redes neuronales recurrentes</bold>
</trans-title>
</trans-title-group>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name name-style="western">
<surname>Perez Leal</surname>
<given-names>Juan Pablo</given-names>
</name>
<xref ref-type="corresp" rid="corresp1"/>
<xref ref-type="aff" rid="aff1"/>
<email>juanpabloperezleal@gmail.com</email>
</contrib>
<contrib contrib-type="author" corresp="no">
<name name-style="western">
<surname>Ríos Gutiérrez</surname>
<given-names>Andrés Sebastián</given-names>
</name>
<xref ref-type="aff" rid="aff2"/>
<xref ref-type="aff" rid="aff3"/>
</contrib>
<contrib contrib-type="author" corresp="no">
<name name-style="western">
<surname>Romo-Bucheli</surname>
<given-names>David</given-names>
</name>
<xref ref-type="aff" rid="aff4"/>
</contrib>
</contrib-group>
<aff id="aff1">
<institution content-type="original">Escuela Ingeniería de Sistemas e Informática, Facultad de ingenierías fisicomecánicas, Universidad Industrial de Santander, Bucaramanga, Colombia.</institution>
<country country="CO">Colombia</country>
<institution-wrap>
<institution content-type="orgname">Universidad Industrial de Santander</institution>
<institution-id institution-id-type="ror">https://ror.org/00xc1d948</institution-id>
</institution-wrap>
</aff>
<aff id="aff2">
<institution content-type="original">Departamento de Estadística, Universidad Nacional de Colombia, Bogotá, Colombia.</institution>
<country country="CO">Colombia</country>
<institution-wrap>
<institution content-type="orgname">Universidad Nacional de Colombia</institution>
<institution-id institution-id-type="ror">https://ror.org/059yx9a68</institution-id>
</institution-wrap>
</aff>
<aff id="aff3">
<institution content-type="original">Escuela de Matemáticas, Universidad Industrial de Santander</institution>
<country country="CO">Colombia</country>
<institution-wrap>
<institution content-type="orgname">Universidad Industrial de Santander</institution>
<institution-id institution-id-type="ror">https://ror.org/00xc1d948</institution-id>
</institution-wrap>
</aff>
<aff id="aff4">
<institution content-type="original">Escuela de Matemáticas, Universidad Industrial de Santander</institution>
<country country="CO">Colombia</country>
<institution-wrap>
<institution content-type="orgname">Universidad Industrial de Santander</institution>
<institution-id institution-id-type="ror">https://ror.org/00xc1d948</institution-id>
</institution-wrap>
</aff>
<author-notes>
<corresp id="corresp1">
<email>
<italic>juanpabloperezleal@gmail.com</italic>
</email>
</corresp>
</author-notes>
<pub-date pub-type="epub-ppub">
<season>Enero</season>
<year>2025</year>
</pub-date>
<volume>17</volume>
<issue>1</issue>
<fpage>1</fpage>
<lpage>16</lpage>
<history>
<date date-type="received" publication-format="dd mes yyyy">
<day>29</day>
<month>11</month>
<year>2024</year>
</date>
<date date-type="accepted" publication-format="dd mes yyyy">
<day>09</day>
<month>12</month>
<year>2024</year>
</date>
<date date-type="pub" publication-format="dd mes yyyy">
<day>12</day>
<month>05</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Los autores conservan los derechos de autor y garantizan a la revista el derecho de ser la primera publicación del trabajo bajo una licencia Creative Commons Atribución-NoComercial 4.0 Internacional.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Revista ACI Avances en Ciencias e Ingenierías</copyright-holder>
<ali:free_to_read/>
<license xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<ali:license_ref>https://creativecommons.org/licenses/by-nc/4.0/</ali:license_ref>
<license-p>Esta obra está bajo una Licencia Creative Commons Atribución-NoComercial 4.0 Internacional.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p>This study investigates the use of Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks to predict crime patterns in Bucaramanga, Colombia. A temporal approach is presented, which starts by splitting the city into 17 communes. Using a dataset of robbery incidents from 2016 to January 2023, we developed individual time series models for each commune. Then, we used the Root Mean Squared Error (RMSE) as the evaluation metric in these regression tasks. The LSTM models consistently outperformed both the RNN and ARIMA models, a classical methodology for time series prediction, achieving lower RMSE scores. The LSTM model yielded an average RMSE of 2.875 (with a standard deviation of 1.657), which is considerably lower than that obtained by the RNN model 3.101 (1.82) and the ARIMA model 3.428 (2.57). These results show that LSTM better captures the complex temporal dependencies in the data. Future work should explore hybrid models and the incorporation of additional data sources to enhance predictive accuracy further.</p>
</abstract>
<trans-abstract xml:lang="es">
<title>Resumen</title>
<p>Este estudio investiga el uso de Redes Neuronales Recurrentes (RNN) y redes de Gran Memoria a Corto Plazo (LSTM) para predecir patrones de criminalidad en Bucaramanga, Colombia. Se presenta un enfoque temporal que comienza dividiendo la ciudad en 17 comunas. Utilizando un conjunto de datos de incidentes de robos desde 2016 hasta enero de 2023, se desarrollaron modelos de series de tiempo individuales para cada comuna. Posteriormente, se empleó el Error Cuadrático Medio (RMSE) como métrica de evaluación en estas tareas de regresión. Los modelos LSTM superaron de manera consistente tanto a los modelos RNN como a los modelos ARIMA, una metodología clásica para la predicción de series temporales, logrando menores puntajes de RMSE. El modelo LSTM obtuvo un RMSE promedio de 2.875 (con una desviación estándar de 1.657), considerablemente inferior al obtenido por el modelo RNN, con 3.101 (1.82), y el modelo ARIMA, con 3.428 (2.57). Estos resultados demuestran que LSTM captura mejor las complejas dependencias temporales en los datos. Trabajos futuros deberían explorar modelos híbridos y la incorporación de fuentes de datos adicionales para mejorar aún más la precisión predictiva.</p>
</trans-abstract>
<kwd-group xml:lang="en">
<title>Keywords</title>
<kwd>Crime prediction</kwd>
<kwd>Neural Networks</kwd>
<kwd>Long Short-Term Memory (LSTM)</kwd>
<kwd>Recurrent Neural Networks</kwd>
</kwd-group>
<kwd-group xml:lang="es">
<title>Palabras clave</title>
<kwd>Predicción de crímenes</kwd>
<kwd>redes neuronales</kwd>
<kwd>redes LSTM</kwd>
<kwd>Redes Neuronales Recurrentes</kwd>
</kwd-group>
<counts>
<fig-count count="5"/>
<table-count count="1"/>
<equation-count count="12"/>
<ref-count count="23"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>redalyc-journal-id</meta-name>
<meta-value>7261</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec>
<title>
<bold>INTRODUCTION</bold>
</title>
<p>Crime prediction is a critical area of study for urban planning and public safety. The ability to anticipate criminal activities enables authorities to allocate scarce resources more effectively, plan preventive measures, and enhance community safety. Traditional methods of crime prediction have often relied on statistical models and historical data analysis [1]. Some of them are related to stational ARIMA models [2, 3], and others to GARCH models [4]. However, with the advent of advanced computational techniques, there is a growing interest in leveraging machine learning algorithms to improve the accuracy and reliability of these predictions. In addition, the crime data in Bucaramanga is a large volume database as may be seen in [5], and therefore it is difficult to get a good fit by using parametric time series models [6].</p>
<p>In recent years, Recurrent Neural Networks (RNNs) and Long Short-Term Memory networks (LSTMs), have shown significant promise in modeling time series data due to their ability to capture temporal dependencies and patterns [7, 8, 9]. RNNs, with their recurrent connections, are designed to recognize sequences and trends over time, making them suitable for analyzing temporal data. LSTMs, an extension of RNNs, address the limitations of traditional RNNs by incorporating memory cells that can store information for long periods by overcoming the issue of vanishing gradients and enabling the network to retain long-term dependencies [10].</p>
<p>This paper focuses on the application of RNNs and LSTMs to predict crime patterns in the city of Bucaramanga. We aim to develop models that can identify and predict the temporal trends in crime occurrences with higher accuracy. The study will explore the effectiveness of these models in capturing the intricate patterns of criminal activities, considering various temporal factors such as time of day, day of the week, and seasonal variations.</p>
</sec>
<sec>
<title>
<bold>RELATED WORKS</bold>
</title>
<p>Predicting crime in cities has become a critical area of research for law enforcement and public safety agencies. By analyzing historical crime data, researchers can identify patterns and develop tools to anticipate future occurrences. Several existing research approaches are based on various machine learning for predicting crime. For instance, in [11], the authors utilize Chicago's public crime dataset for crime trend prediction based on a LSTM model. It also incorporates POI (Point of Interest) information data and employs a convolutional model to analyze and compare the time series distribution, spatial distribution, and prediction accuracy of various algorithm models. The experimental results demonstrate a hit rate of 20.11% for different experiments using the LSTM algorithm. Another example is applied in Atlanta, which is one of the cities with high crime rates in the United States. In [12], the authors show that crime events exhibit spatial aggregation and temporal dependence, indicating that criminal incidents are predictable. In the mentioned paper, crime data from Atlanta,  spanning  2009  to  2016,  is  used  to  identify  spatio-temporal  distribution  features  and  create  statistical  visualizations.  Forecast  of  daily  crime  occurrences  is  obtained  via  a  LSTM  model.  The  study  further  explores  the  impact  of  varying  spatio-temporal  scales  on  prediction  accuracy.  When  using  an  input  time  series  length  of  $50$  days  and  a  spatial  cell  size  of  $0.05$  degrees,  the  correlation  coefficient  (R  value)  between  predicted  and  observed crime data exceeds $0.87$. Finally, in [13] a spatio-temporal approach is applied to model monthly robberies in Buenos Aires (Argentina). The main objective of the study is to implement predictive models to associate specific corners and times with the occurrence of crimes, based on monthly robbery records in Buenos Aires from 2017 to 2020. Initially, time series models using only historical crime counts are employed to predict criminal incidents at each street corner. Additionally, predictions include covariates related to the surrounding environment  and  weather  conditions,  transforming  the  problem  into  a  spatio-temporal  study. The study shows that crime, being a human activity, does not follow random behavior but  rather  social  patterns  influenced  by  time,  weather,  and  geographical  zones;  in  some  cases, it shows an impact in the model of more than 13% for some features [13].</p>
</sec>
<sec>
<title>
<bold>MATERIALS</bold>
</title>
<p>Bucaramanga is the capital of the department of Santander, Colombia. It is located on a plateau at 960 meters above sea level, with a population of approximately 600,000. Nestled in the eastern Andes Mountains of Colombia, Bucaramanga is characterized by its undulating topography and dense urbanization. The city extends over a series of hills and valleys, resulting in variations in accessibility and visibility across different areas [14]. Administratively, Bucaramanga is divided into 17 communes, each further subdivided into neighborhoods. Leveraging the big sizes and geographical locations of these communes, we get a framework to identify patterns and trends in crime rates. The dataset used in this study encompasses robbery incidents recorded from 2016 to January 2023. This data was collected by Bucaramanga's city hall, in collaboration with the municipal police, and is publicly available on the Datos Abiertos platform [5]. The collected data was grouped in weeks of the year instead of daily, resulting in the count of robberies for each commune. The period associated with the training data spanned from January 1, 2016, to September 12, 2022. A total of 17 time series were obtained, one for each of the communes. Test data for each commune was also extracted and is associated with the period spanning from September 13, 2022, to November 20, 2022</p>
<p>Each record in the dataset includes the date of the robbery and the specific commune in which it occurred.</p>
<p>This dataset allows for the creation of a time series capturing the frequency of robbery incidents within each commune of Bucaramanga. Given the nature of the problem, it is expected that there will be days with zero reported incidents in some communes. This variability provides an opportunity to explore temporal patterns and anomalies in criminal activity across different areas of the city.</p>
<p>The collected data was grouped in weeks of the year instead of daily. The period associated with the training data spanned from January 1, 2016, to September 12, 2022. The models were trained for each commune independently, allowing them to learn the underlying patterns and trends specific to each commune in Bucaramanga. The test data is associated with the period spanning from September 13, 2022, to November 20, 2022. In Figure 1, a heatmap of Bucaramanga displaying the total count of robberies during the training period is shown. The crime distribution shows remarkable differences across the different communes in Bucaramanga.</p>
<p>
<fig id="gf1">
<label>FIGURE 1</label>
<caption>
<title>Map of Bucaramanga displaying the spatial distribution and count of robbery incidents from January 2016 to November 2022. Own elaboration</title>
</caption>
<alt-text>FIGURE 1  Map of Bucaramanga displaying the spatial distribution and count of robbery incidents from January 2016 to November 2022. Own elaboration</alt-text>
<graphic xlink:href="726182578011_gf7.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 1  Map of Bucaramanga displaying the spatial distribution and count of robbery incidents from January 2016 to November 2022. Own elaboration</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>FIGURE 1. </bold>Map of Bucaramanga displaying the spatial distribution and count of robbery incidents from January 2016 to November 2022. Own elaboration</p>
</sec>
<sec>
<title>
<bold>METHODOLOGY</bold>
</title>
<p>In this work, we compared several methodologies for crime prediction, utilizing an independent temporal approach for each of the communes of Bucaramanga. Our purpose is to evaluate the effectiveness of different time series models in forecasting crime patterns. These are the different models that are going to be employed in this study.</p>
<p>
<bold>ARIMA:</bold> Among the most commonly used tools for time series forecasting is the ARIMA model  (Autoregressive  Integrated  Moving  Average).  The  ARIMA  model  is  valuable  in  situations  where  the  data  exhibits  autocorrelation,  that  is,  correlation  between past  random  variables  and  the  present  time  random  variable.  Additionally,  it  can  handle  time  series  that  are  non-stationary,  i.e.,  those  that  show  trends  or  systematic  variations  over  time  [15].  It  is  important  to  note  that  the  ARIMA  model  is  a  dynamic  time  series  model,  meaning  that  future  estimates  are  explained  by  past  data  and  not  by independent variables [16,17]. The ARIMA model can be defined as a composition of three main components: Autoregression (AR), Integration (I), and Moving Average (MA). Autoregression  (AR)  refers  to  the  linear  dependency  of  a  current  observation  on  past  observations.  In  an  AR  model,  the  dependent  variable  is  linearly  regressed  on  its  past  values up to order p.</p>
<p>
<disp-formula id="e1">
<label>(1)</label>
<alternatives><mml:math id="mN104C1" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>c</mml:mi>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3D5;</mml:mi>             <mml:mn>1</mml:mn>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3D5;</mml:mi>             <mml:mn>2</mml:mn>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>2</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:mo>&#x22EF;</mml:mo>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3D5;</mml:mi>             <mml:mi>p</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mi>p</mml:mi>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3F5;</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee3.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>x</italic>
<sub>
<italic>t</italic>
</sub> is the dependent variable at time <italic>t</italic>, <italic>c</italic> is a constant, <italic>ϕ</italic>
<sub>
<italic>1</italic>
</sub>, <italic>ϕ</italic>
<sub>
<italic>2</italic>
</sub>, . . . , <italic>ϕ</italic>
<sub>
<italic>p</italic>
</sub> are the autoregression parameters, and <italic>ε</italic>
<sub>
<italic>t</italic>
</sub> is a random error term at time <italic>t</italic>. Generally, this follows a white noise (WN) process with variance <italic>σ²</italic>, for all <italic>t</italic>≥0, <italic>ε</italic>
<sub>
<italic>t</italic>
</sub>
<italic>WN</italic> (0, <italic>σ²</italic>) [17].</p>
<p>The Moving Average (MA) models the relationship between an observation and a residual error term generated by a moving average of errors in previous periods. In an MA model, the dependent variable is linearly regressed on past error terms up to order q. The relationship is mathematically defined as:</p>
<p>
<disp-formula id="e2">
<label>(2)</label>
<alternatives><mml:math id="mN1053B" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>c</mml:mi>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3B8;</mml:mi>             <mml:mn>1</mml:mn>         </mml:msub>         <mml:msub>             <mml:mi>&#x3F5;</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3B8;</mml:mi>             <mml:mn>2</mml:mn>         </mml:msub>         <mml:msub>             <mml:mi>&#x3F5;</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>2</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:mo>&#x22EF;</mml:mo>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3B8;</mml:mi>             <mml:mi>q</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>&#x3F5;</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mi>q</mml:mi>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x3F5;</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>,</mml:mo>         <mml:msub>             <mml:mi>&#x3F5;</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>&#x223C;</mml:mo>         <mml:mi>W</mml:mi>         <mml:mi>N</mml:mi>         <mml:mo>(</mml:mo>         <mml:mn>0</mml:mn>         <mml:mo>,</mml:mo>         <mml:msup>             <mml:mi>&#x3C3;</mml:mi>             <mml:mn>2</mml:mn>         </mml:msup>         <mml:mo>)</mml:mo>         <mml:mo>,</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee4.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>c</italic> is a constant, <italic>θ</italic>
<sub>
<italic>1</italic>
</sub>, <italic>θ</italic>
<sub>
<italic>2</italic>
</sub>, . . . , <italic>θ</italic>
<sub>
<italic>q</italic>
</sub> are the moving average parameters, and <italic>ε</italic>
<sub>
<italic>1</italic>
</sub>, <italic>ε</italic>
<sub>
<italic>2</italic>
</sub>, . . . , <italic>ε</italic>
<sub>
<italic>t-q</italic>
</sub> are the error terms at previous times. Integration (I) refers to the process of making a time series stationary, i.e., removing trends or systematic patterns that may affect the prediction. This is achieved by differencing the time series until it becomes stationary. The order of differencing is denoted as <italic>d</italic>. This difference is recurrently defined as:</p>
<p>
<disp-formula id="e3">
<label>(3)</label>
<alternatives><mml:math id="mN105A1" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msup>             <mml:mi>&#x394;</mml:mi>             <mml:mi>d</mml:mi>         </mml:msup>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:msup>             <mml:mi>&#x394;</mml:mi>             <mml:mrow>                 <mml:mi>d</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msup>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>&#x2212;</mml:mo>         <mml:msup>             <mml:mi>&#x394;</mml:mi>             <mml:mrow>                 <mml:mi>d</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msup>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>,</mml:mo>         <mml:mi>&#x394;</mml:mi>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:msup>             <mml:mi>&#x394;</mml:mi>             <mml:mn>1</mml:mn>         </mml:msup>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>&#x2212;</mml:mo>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mspace width="0.5em"></mml:mspace>         <mml:mtext>and</mml:mtext>         <mml:mspace width="0.5em"></mml:mspace>         <mml:mi>d</mml:mi>         <mml:mo>=</mml:mo>         <mml:mn>2</mml:mn>         <mml:mo>,</mml:mo>         <mml:mn>3</mml:mn>         <mml:mo>,</mml:mo>         <mml:mo>...</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee7.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>The ARIMA model combines these three components into a single framework and is denoted as ARIMA(<italic>p</italic>, <italic>d</italic>, <italic>q</italic>), where <italic>p</italic> is the order of autoregression, <italic>d</italic> is the order of differencing to get a stationary process (see [18]), and <italic>q</italic> is the order of moving average.</p>
<p>
<italic>Delay operator</italic> is defined, for all <italic>u</italic> an integer number, as <italic>L</italic>
<sup>
<italic>u</italic>
</sup>
<italic>x</italic>
<sub>
<italic>t</italic>
</sub>
<italic> := x</italic>
<sub>
<italic>t-u</italic>
</sub> [18]. Based on this operator, the ARIMA(<italic>p</italic>, <italic>d</italic>, <italic>q</italic>) model is defined by:</p>
<p>
<disp-formula id="e4">
<label>(4)</label>
<alternatives><mml:math id="mN10614" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>&#x3A6;</mml:mi>             <mml:mi>p</mml:mi>         </mml:msub>         <mml:mo>(</mml:mo>         <mml:mi>L</mml:mi>         <mml:mo>)</mml:mo>         <mml:mo>(</mml:mo>         <mml:mn>1</mml:mn>         <mml:mo>&#x2212;</mml:mo>         <mml:mi>L</mml:mi>         <mml:msup>             <mml:mo>)</mml:mo>             <mml:mi>d</mml:mi>         </mml:msup>         <mml:mi>x</mml:mi>         <mml:mo>=</mml:mo>         <mml:mi>c</mml:mi>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>&#x398;</mml:mi>             <mml:mi>q</mml:mi>         </mml:msub>         <mml:mo>(</mml:mo>         <mml:mi>L</mml:mi>         <mml:mo>)</mml:mo>         <mml:mi>&#x3F5;</mml:mi>         <mml:mo>,</mml:mo>         <mml:mspace width="0.5em"></mml:mspace>         <mml:mtext>with</mml:mtext>         <mml:mspace width="0.5em"></mml:mspace>         <mml:mi>&#x3F5;</mml:mi>         <mml:mo>&#x223C;</mml:mo>         <mml:mi>W</mml:mi>         <mml:mi>N</mml:mi>         <mml:mo>(</mml:mo>         <mml:mn>0</mml:mn>         <mml:mo>,</mml:mo>         <mml:msup>             <mml:mi>&#x3C3;</mml:mi>             <mml:mn>2</mml:mn>         </mml:msup>         <mml:mo>)</mml:mo>         <mml:mo>,</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee9.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>Φ</italic>
<sub>
<italic>p</italic>
</sub>
<italic>(L</italic>), <italic>Θ</italic>
<sub>
<italic>q</italic>
</sub>
<italic>(L</italic>) are polynomials, respectively called the autoregressive polynomial and the moving average polynomial. In consequence, <italic>d</italic> represents the order of the differences required for the stochastic process <italic>Φ</italic>
<sub>
<italic>p</italic>
</sub>
<italic>(L)x</italic>
<sub>
<italic>t</italic>
</sub> = <italic>c</italic> + <italic>Θ</italic>
<sub>
<italic>q</italic>
</sub>
<italic>(L</italic>)<italic>ε</italic>
<sub>
<italic>t</italic>
</sub> to be stationary.</p>
<p>
<bold>RRN Model:</bold> Recurrent neural networks (RNNs) are designed to recognize patterns in sequences of data, making them well-suited for tasks involving time series analysis. Unlike traditional feedforward neural networks, RNNs have connections that form directed cycles, enabling them to maintain a memory of previous inputs [19].</p>
<p>The core of an RNN is the recurrent cell, which processes the input sequence one step at a time while maintaining a hidden state that evolves over time. We can do a mathematical formulation of this for an input sequence <italic>
<bold>x</bold>
</italic> = (x<sub>1</sub>, x<sub>2</sub>,..., x<sub>
<italic>t</italic>
</sub>) where <italic>
<bold>x</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> represents the input at time step <italic>t</italic>. In our case, <italic>x</italic>
<sub>
<italic>t</italic>
</sub> would be the number of incidents at date <italic>t</italic> [20].</p>
<p>There is a hidden state <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> at time step <italic>t</italic> that is computed using the current input <italic>
<bold>x</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> and the previous hidden state <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t-1</bold>
</italic>
</sub>:</p>
<p>
<disp-formula id="e5">
<label>(5)</label>
<alternatives><mml:math id="mN106D1" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>&#x3C3;</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mrow>                 <mml:mi>h</mml:mi>                 <mml:mi>x</mml:mi>             </mml:mrow>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mrow>                 <mml:mi>h</mml:mi>                 <mml:mi>h</mml:mi>             </mml:mrow>         </mml:msub>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>b</mml:mi>             <mml:mi>h</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee26.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>hx</bold>
</italic>
</sub> is the weight matrix connecting the input to the hidden state, <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>hh</bold>
</italic>
</sub> is the weight matrix connecting the hidden state to itself from the previous time step, <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>h</bold>
</italic>
</sub> is the bias vector, and <italic>σ</italic> is a nonlinear activation function, typically a hyperbolic tangent or a ReLU function [21]. The output <italic>ŷ</italic>
<sub>
<italic>t</italic>
</sub> at time step <italic>t</italic> is computed by using the current hidden state <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>:</p>
<p>
<disp-formula id="e6">
<label>(6)</label>
<alternatives><mml:math id="mN1073A" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:mi>t</mml:mi>         <mml:mo>=</mml:mo>         <mml:mi>g</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mrow>                 <mml:mi>h</mml:mi>                 <mml:mi>y</mml:mi>             </mml:mrow>         </mml:msub>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>b</mml:mi>             <mml:mi>y</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee13.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>hy</bold>
</italic>
</sub> is the weight matrix connecting the hidden state to the output, <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>y</bold>
</italic>
</sub> is the bias vector, and <italic>g</italic> is typically a linear or softmax function depending on the task. In our case, <italic>g</italic> is a linear function as we are dealing with regression. The RNN model learns to adjust the weights <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>hx</bold>
</italic>
</sub>
<italic>
<bold>, W</bold>
</italic>
<sub>
<italic>
<bold>hh</bold>
</italic>
</sub>, and <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>hy</bold>
</italic>
</sub>, as well as the biases <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>h</bold>
</italic>
</sub> and <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>y</bold>
</italic>
</sub>, by minimizing a loss function over the training data. For time series prediction, a common choice is the mean squared error (MSE).</p>
<p>In the context of this study, we employ RNNs to predict crime patterns based on a time series dataset. Each time series contains two columns: one for the number of robbery incidents and the other for the corresponding date. The same process is applied separately for each of the 17 communes in Bucaramanga, resulting in 17 distinct models, each tailored to the specific crime dynamics of its respective commune.</p>
<p>
<bold>LSTM Model:</bold> LSTM networks are a type of RNN that address the limitations of traditional RNNs, particularly the issue of vanishing and exploding gradients. LSTMs introduce a more complex unit structure compared to traditional RNNs, featuring memory cells that can store information for extended periods. The key components of an LSTM cell include the cell state (<italic>
<bold>c</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>) and three types of gates: input gate, forget gate, and output gate. These gates regulate the flow of information into, out of, and within the cell [22].</p>
<p>In this context, given the number of incidents on date <italic>t</italic> for an input sequence <italic>
<bold>x</bold>
</italic> = (x<sub>1</sub>, x<sub>2</sub>,..., x<sub>
<italic>t</italic>
</sub>), the LSTM network processes this information in a structured manner.</p>
<p>The forget gate determines what portion of the previous cell state (<italic>
<bold>c</bold>
</italic>
<sub>
<italic>
<bold>t-1</bold>
</italic>
</sub>) should be retained based on the current input (<italic>
<bold>x</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>) and the previous hidden state (<italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t-1</bold>
</italic>
</sub>):</p>
<p>
<disp-formula id="e7">
<label>(7)</label>
<alternatives><mml:math id="mN107FF" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>f</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>&#x3C3;</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mi>f</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>U</mml:mi>             <mml:mi>f</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>b</mml:mi>             <mml:mi>f</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee14.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>σ</italic> is the sigmoid function, <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>f</bold>
</italic>
</sub> and <italic>
<bold>U</bold>
</italic>
<sub>
<italic>
<bold>f</bold>
</italic>
</sub> are weight matrices, <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t-1</bold>
</italic>
</sub> is the previous hidden state, and <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>f</bold>
</italic>
</sub> is the bias vector.</p>
<p>The input gate controls the amount of new information added to the cell state. It produces an intermediate candidate cell state (<italic>
<bold>C̃</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>) that is regulated by the input gate (<italic>
<bold>i</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>):</p>
<p>
<disp-formula id="e8">
<label>(8)</label>
<alternatives><mml:math id="mN10875" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>i</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>&#x3C3;</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mi>i</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>U</mml:mi>             <mml:mi>i</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>b</mml:mi>             <mml:mi>i</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>         <mml:mspace width="2em"></mml:mspace>         <mml:msub>             <mml:mover>                 <mml:mi>c</mml:mi>                 <mml:mo>~</mml:mo>             </mml:mover>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>tanh</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mi>c</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>U</mml:mi>             <mml:mi>c</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>b</mml:mi>             <mml:mi>c</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee16.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>
<bold>i</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> is the input gate output, <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>i</bold>
</italic>
</sub> is the weight matrix for the input <italic>
<bold>x</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>, <italic>
<bold>U</bold>
</italic>
<sub>
<italic>
<bold>i</bold>
</italic>
</sub> is the weight matrix for the previous hidden state <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t-1</bold>
</italic>
</sub>, <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>i</bold>
</italic>
</sub> is the bias vector for the input gate, <italic>
<bold>C̃</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> is the candidate cell state, <italic>
<bold>W</bold>
</italic>
<sub>
<italic>
<bold>c</bold>
</italic>
</sub> is the weight matrix for the input <italic>
<bold>x</bold>
</italic> related to the candidate cell state, <italic>
<bold>U</bold>
</italic>
<sub>
<italic>
<bold>c</bold>
</italic>
</sub> is the weight matrix for the previous hidden state <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t-1</bold>
</italic>
</sub>, and <italic>
<bold>b</bold>
</italic>
<sub>
<italic>
<bold>c</bold>
</italic>
</sub> is the bias vector for the candidate cell state. The cell state is updated by combining the previous cell state and the candidate cell state:</p>
<p>
<disp-formula id="e9">
<label>(9)</label>
<alternatives><mml:math id="mN1090E" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>c</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:msub>             <mml:mi>f</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>&#x2299;</mml:mo>         <mml:msub>             <mml:mi>c</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>i</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>&#x2299;</mml:mo>         <mml:msub>             <mml:mover>                 <mml:mi>c</mml:mi>                 <mml:mo>~</mml:mo>             </mml:mover>             <mml:mi>t</mml:mi>         </mml:msub>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee18.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where ⊙ denotes the element-wise multiplication.</p>
<p>The output gate determines the output of the LSTM cell based on the up-dated cell state (<italic>
<bold>c</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>) and the current input, producing the hidden state (<italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub>):</p>
<p>
<disp-formula id="e10">
<label>(10)</label>
<alternatives><mml:math id="mN10961" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:msub>             <mml:mi>o</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:mi>&#x3C3;</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>W</mml:mi>             <mml:mi>o</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>x</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>U</mml:mi>             <mml:mi>o</mml:mi>         </mml:msub>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>&#x2212;</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>         </mml:msub>         <mml:mo>+</mml:mo>         <mml:msub>             <mml:mi>b</mml:mi>             <mml:mi>o</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>     </mml:mrow>     <mml:mspace linebreak="newline"></mml:mspace>     <mml:mrow>         <mml:msub>             <mml:mi>h</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>=</mml:mo>         <mml:msub>             <mml:mi>o</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>&#x2299;</mml:mo>         <mml:mi>tanh</mml:mi>         <mml:mo>(</mml:mo>         <mml:msub>             <mml:mi>c</mml:mi>             <mml:mi>t</mml:mi>         </mml:msub>         <mml:mo>)</mml:mo>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee20.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>
<bold>o</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> is the output gate output and <italic>
<bold>h</bold>
</italic>
<sub>
<italic>
<bold>t</bold>
</italic>
</sub> is the hidden state at time step <italic>t</italic>. In this study, we employ LSTMs to predict crime patterns using time series data, where each time series consists of two columns: the number of robbery incidents and the corresponding date. This process is replicated for each of the 17 communes in Bucaramanga, resulting in 17 individual models tailored to the unique crime dynamics of each commune.</p>
<p>Each LSTM model is trained on the historical time series data of robbery incidents for a specific commune. The training process involves backpropagation through time (BPTT) to update the model parameters. Given the sparse nature of the data, with many days having zero incidents, the models must learn to identify underlying patterns and trends that contribute to the temporal dynamics of crime within each commune.</p>
<p>The tests were conducted using TensorFlow Keras in a Google Colab environment, utilizing Google cloud TPU v2 hardware. This setup provided high-performance computing for training deep learning models efficiently on large datasets.</p>
<p>
<bold>Metrics:</bold>
</p>
<p>
<bold>Metrics:</bold> The performance of the planned models in predicting crime patterns is evaluated using the Root Mean Squared Error (RMSE). RMSE is a commonly used metric for regression tasks, and it is defined as follows:</p>
<p>
<disp-formula id="e11">
<label>
<italic>    (11)</italic>
</label>
<alternatives><mml:math id="mN109CC" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:mi>R</mml:mi>         <mml:mi>M</mml:mi>         <mml:mi>S</mml:mi>         <mml:mi>E</mml:mi>         <mml:mo>=</mml:mo>         <mml:msqrt>             <mml:mrow>                 <mml:mfrac>                     <mml:mn>1</mml:mn>                     <mml:mi>n</mml:mi>                 </mml:mfrac>                 <mml:munderover>                     <mml:mo>&#x2211;</mml:mo>                     <mml:mrow>                         <mml:mi>i</mml:mi>                         <mml:mo>=</mml:mo>                         <mml:mn>1</mml:mn>                     </mml:mrow>                     <mml:mi>n</mml:mi>                 </mml:munderover>                 <mml:msup>                     <mml:mrow>                         <mml:mo>(</mml:mo>                         <mml:msub>                             <mml:mover>                                 <mml:mi>y</mml:mi>                                 <mml:mo>^</mml:mo>                             </mml:mover>                             <mml:mi>i</mml:mi>                         </mml:msub>                         <mml:mo>&#x2212;</mml:mo>                         <mml:msub>                             <mml:mi>y</mml:mi>                             <mml:mi>i</mml:mi>                         </mml:msub>                         <mml:mo>)</mml:mo>                     </mml:mrow>                     <mml:mn>2</mml:mn>                 </mml:msup>             </mml:mrow>         </mml:msqrt>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee23.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>n</italic> is the number of observations, (ŷ) is the predicted value for the <italic>i</italic>-th observation, and <italic>y</italic>
<sub>
<italic>i</italic>
</sub> is the actual value for the <italic>i</italic>-th observation. Lower RMSE values indicate a better fit of the model to the data, as it implies that the differences between the predicted and actual values are smaller.</p>
<p>We also computed the SMAPE metric defined as:</p>
<p>
<disp-formula id="e12">
<label>
<italic>(12)</italic>
</label>
<alternatives><mml:math id="mN10A1F" xmlns:mml="http://www.w3.org/1998/Math/MathML">     <mml:mrow>         <mml:mi>S</mml:mi>         <mml:mi>M</mml:mi>         <mml:mi>A</mml:mi>         <mml:mi>P</mml:mi>         <mml:mi>E</mml:mi>         <mml:mo>=</mml:mo>         <mml:mfrac>             <mml:mrow>                 <mml:mn>100</mml:mn>                 <mml:mo>%</mml:mo>             </mml:mrow>             <mml:mi>n</mml:mi>         </mml:mfrac>         <mml:munderover>             <mml:mo>&#x2211;</mml:mo>             <mml:mrow>                 <mml:mi>t</mml:mi>                 <mml:mo>=</mml:mo>                 <mml:mn>1</mml:mn>             </mml:mrow>             <mml:mi>n</mml:mi>         </mml:munderover>         <mml:mfrac>             <mml:mrow>                 <mml:mo>|</mml:mo>                 <mml:msub>                     <mml:mi>y</mml:mi>                     <mml:mi>t</mml:mi>                 </mml:msub>                 <mml:mo>&#x2212;</mml:mo>                 <mml:msub>                     <mml:mover>                         <mml:mi>y</mml:mi>                         <mml:mo>^</mml:mo>                     </mml:mover>                     <mml:mi>t</mml:mi>                 </mml:msub>                 <mml:mo>|</mml:mo>             </mml:mrow>             <mml:mfrac>                 <mml:mrow>                     <mml:mo>|</mml:mo>                     <mml:msub>                         <mml:mi>y</mml:mi>                         <mml:mi>t</mml:mi>                     </mml:msub>                     <mml:mo>|</mml:mo>                     <mml:mo>+</mml:mo>                     <mml:mo>|</mml:mo>                     <mml:msub>                         <mml:mover>                             <mml:mi>y</mml:mi>                             <mml:mo>^</mml:mo>                         </mml:mover>                         <mml:mi>t</mml:mi>                     </mml:msub>                     <mml:mo>|</mml:mo>                 </mml:mrow>                 <mml:mn>2</mml:mn>             </mml:mfrac>         </mml:mfrac>     </mml:mrow> </mml:math>
<graphic xlink:href="726182578011_ee24.png" position="anchor" orientation="portrait">
<alt-text/>
</graphic>
</alternatives>
</disp-formula>
</p>
<p>where <italic>n</italic> corresponds to the total number of observations, <italic>t</italic> is the index of the observation, <italic>y</italic>
<sub>
<italic>i</italic>
</sub> is the actual value at time <italic>t</italic>, and ŷ is the predicted value at time <italic>t</italic>.</p>
<p>The SMAPE metric outputs a percentage value that quantifies the accuracy of a prediction model. The value ranges from 0% to 200%, where 0% indicates a perfect prediction with no error, meaning the predicted values match the actual values exactly. As the SMAPE value increases, it indicates a larger discrepancy between the predicted and actual values, with 200% representing the maximum possible error when one value is positive and the other is zero. A SMAPE close to 100% suggests that the model's predictions are, on average, as far from the actual values as they are close, implying poor predictive performance.</p>
</sec>
<sec>
<title>
<bold>EXPERIMENTAL SETUP AND RESULTS</bold>
</title>
<p>The RNN implementation in this study is based on the foundational work by [23]. This basic RNN model serves as a reference point, as it does not include the advanced gating mechanisms introduced in more modern architectures like LSTM and GRU. While both LSTM and GRU enhance the basic RNN architecture by incorporating gates that effectively manage memory and filter information, we chose to focus exclusively on LSTM as it is representative of this kind of RNN.</p>
<p>After training the models using data from each commune, the results of the previously mentioned models were compared. Figure 2 (left) shows a table of RMSE results for each commune.</p>
<p>The LSTM model demonstrates superior overall performance across all models evaluated in this study. It consistently achieves lower RMSE scores compared to the RNN and ARIMA models, indicating its effectiveness in capturing the complex temporal patterns in the crime data. However, it is noteworthy that the RNN model occasionally outperforms the LSTM in certain specific communes. In contrast, the ARIMA model yields higher RMSE scores, underscoring the limitations of a purely statistical, regressive approach for this type of problem.</p>
<p>
<bold>Table 1. </bold>RMSE for each one of the models used. Best values are highlighted in bold.</p>
<p>
<table-wrap id="gt1">
<label>Table 1</label>
<caption>
<title>RMSE for each one of the models used. Best values are highlighted in bold.</title>
</caption>
<alt-text>Table 1  RMSE for each one of the models used. Best values are highlighted in bold.</alt-text>
<alternatives>
<graphic xlink:href="726182578011_gt2.png" position="anchor" orientation="portrait">
<alt-text>Table 1  RMSE for each one of the models used. Best values are highlighted in bold.</alt-text>
</graphic>
<table id="gt2-526564616c7963">
<thead style="display:none;">
<tr style="display:none;">
<th style="display:none;"/>
</tr>
</thead>
<tbody>
<tr>
<td style="background-color: #cc0000;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>Commune</bold>
</td>
<td style="background-color: #cc0000;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>ARIMA</bold>
</td>
<td style="background-color: #cc0000;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>RNN</bold>
</td>
<td style="background-color: #cc0000;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>LSTM</bold>
</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.915</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.722</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.761</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.486</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.387</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.534</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">3</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">7.162</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.684</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.216</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">4</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.338</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.266</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.292</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.269</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.416</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.158</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">6</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">4.184</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">3.733</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">3.852</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">7</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.841</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.620</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.591</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">8</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">0.854</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.109</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.108</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">9</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.262</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.517</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.578</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">10</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">4.656</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">4.354</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">4.016</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">11</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.205</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.106</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.358</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">12</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">8.643</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.693</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">4.893</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">13</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">8.177</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">6.697</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.867</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">14</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.169</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.121</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.006</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">15</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.917</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.485</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">5.501</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">16</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.501</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.302</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.500</td>
</tr>
<tr>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">17</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.703</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">2.501</td>
<td style="border: 1px solid #000000; padding: 0.07in 0.13in">1.641</td>
</tr>
<tr>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>Mean</bold>
</td>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">3.428</td>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">3.101</td>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>2.875</bold>
</td>
</tr>
<tr>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>St. dev.</bold>
</td>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">2.570</td>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">1.820</td>
<td style="background-color: #ffd5d5;  border: 1px solid #000000; padding: 0.07in 0.13in">
<bold>1.657</bold>
</td>
</tr>
</tbody>
</table>
</alternatives>
</table-wrap>
</p>
<p>
<fig id="gf2">
<label>FIGURE 2</label>
<caption>
<title>Boxplot comparison of RMSE for all the validated models. Own creation</title>
</caption>
<alt-text>FIGURE 2  Boxplot comparison of RMSE for all the validated models. Own creation</alt-text>
<graphic xlink:href="726182578011_gf8.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 2  Boxplot comparison of RMSE for all the validated models. Own creation</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>FIGURE 2. </bold>Boxplot comparison of RMSE for all the validated models. Own creation</p>
<p>Furthermore, the standard deviation of the RMSE scores reveals that the LSTM model not only provides more accurate predictions on average but also exhibits lower variability. This suggests a more stable and reliable predictive capability. Figure 2 and Table 1 offers a visual representation of the RMSE score distributions across the different models.</p>
<p>In addition to numerical evaluation, we can compare the performance of the models through visualizations of the time series predictions. Figures 3, 4, and 5 display the predicted values for ARIMA, RNN, and LSTM models alongside the actual data.</p>
<p>The graphical representation of the ARIMA model's predictions reveals a significant limitation: its inability to accurately predict spikes in the data. The ARIMA model tends to generate a single smooth approximation throughout the entire time series, failing to capture the abrupt changes and peaks that characterize real-world crime data. This indicates that the ARIMA model, being a purely statistical approach, lacks the flexibility needed to adapt to the complex and nonlinear patterns inherent in the crime time series.</p>
<p>In contrast, the RNN and LSTM models exhibit a markedly improved capacity for approximating the real data. Both models demonstrate a better alignment with the observed trends and are more adept at predicting some of the spikes in the crime incidents. The RNN and LSTM models, leveraging their recurrent structures, can retain and utilize information from previous time steps, allowing them to capture both short-term fluctuations and long-term trends more effectively.</p>
<p>The RNN model tends to produce smoother predictions, often missing the sharp fluctuations in the real data. This smoothing effect results in a more generalized prediction that fails to capture the variability present in the actual crime data. The RNN model struggles to detect sudden spikes in robbery counts, leading to underestimations during periods of high activity, particularly noticeable in Communes 2, 5, and 12. While the RNN model aligns with the general direction of the trends in some communes, it consistently underperforms in accurately following the actual data patterns, particularly in areas with more volatile crime rates. On the other hand, the LSTM model provides a much closer fit to the actual data compared to the RNN. It captures the fluctuations and trends more accurately, indicating its ability to model the temporal dependencies more effectively. Unlike the RNN, the LSTM model is better at detecting and replicating spikes in crime data, as seen in Communes 3, 7, and 13. This suggests that the LSTM model is more responsive to changes in the crime patterns. The LSTM model demonstrates a superior ability to track the real data across all communes. It reduces the prediction error and better follows the actual robbery counts, making it a more reliable model for this type of prediction.</p>
<p>Based on the analysis of SMAPE scores the models across the communes, both models exhibit similar overall performance, with average SMAPE scores of 48.02 for ARIMA and 48.59 for LSTM. However, closer inspection reveals that LSTM outperforms ARIMA in certain contexts. Specifically, in communes 1, 3, 7 and 12, the LSTM model shows superior predictive accuracy. These areas tend to have more stable or predictable crime patterns, or they exhibit higher volumes of crime, allowing LSTM's ability to capture long-term dependencies and handle complex temporal trends to shine. On the other hand, in communes where crime patterns have low variation or are closer to the mean, ARIMA tends to perform better, likely due to its strength in modeling simpler, more linear time series data.</p>
<p>
<fig id="gf3">
<label>FIGURE 3</label>
<caption>
<title>Time series of the predicted values for the ARIMA model. Own creation</title>
</caption>
<alt-text>FIGURE 3  Time series of the predicted values for the ARIMA model. Own creation</alt-text>
<graphic xlink:href="726182578011_gf9.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 3  Time series of the predicted values for the ARIMA model. Own creation</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>FIGURE 3. </bold>Time series of the predicted values for the ARIMA model. Own creation</p>
<p>
<fig id="gf4">
<label>FIGURE 4</label>
<caption>
<title>Time series of the predicted values for the RNN model. Own creation</title>
</caption>
<alt-text>FIGURE 4  Time series of the predicted values for the RNN model. Own creation</alt-text>
<graphic xlink:href="726182578011_gf10.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 4  Time series of the predicted values for the RNN model. Own creation</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>FIGURE 4. </bold>Time series of the predicted values for the RNN model. Own creation</p>
<p>
<fig id="gf5">
<label>FIGURE 5</label>
<caption>
<title>Time series of the predicted values for the LSTM model. Own creation</title>
</caption>
<alt-text>FIGURE 5  Time series of the predicted values for the LSTM model. Own creation</alt-text>
<graphic xlink:href="726182578011_gf11.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 5  Time series of the predicted values for the LSTM model. Own creation</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>FIGURE 5. </bold>Time series of the predicted values for the LSTM model. Own creation</p>
</sec>
<sec>
<title>
<bold>CONCLUSION</bold>
</title>
<p>In this study, we explored the potential of utilizing solely temporal data to understand and predict crime patterns in a Latin American urban context, specifically Bucaramanga.</p>
<p>This approach underscores a significant advancement in the field of urban crime analysis. By focusing on temporal dynamics, we demonstrated that effective predictions could still be achieved without the inclusion of spatial data, which is often considered crucial in such analyses. This finding is particularly valuable for regions where spatial data may be incomplete or unavailable. Our results highlight the robustness of temporal models, such as LSTM, in extracting meaningful patterns from time-series data alone, offering a promising avenue for enhancing public security strategies in similar urban settings. Integrating High-Performance Computing (HPC) enables us to leverage real-time data for scalable AI predictions, enhancing our capacity to handle extensive datasets efficiently and effectively in dynamic environments.</p>
<p>The results demonstrated that LSTM models consistently outperform both RNN and ARIMA models in predicting crime incidents across all 17 communes in our case study.</p>
<p>Visual comparison through time series plots further elucidated the strengths and weaknesses of each model. The ARIMA model's predictions were notably smooth and failed to capture the abrupt changes in crime incidents, indicating a significant shortfall in its predictive capability. In contrast, both the RNN and LSTM models exhibited a better fit to the actual data, more accurately reflecting the observed trends and spikes.</p>
<p>While the study demonstrates the effectiveness of LSTMs in crime prediction, there are limitations that warrant further investigation. Incorporating more features, such as socioeconomic indicators, weather conditions, and public events, could potentially improve the model's predictive power. Future work should explore the integration of spatial data, other data sources, and the development of hybrid models to capitalize on the strengths of different neural network architectures.</p>
</sec>
<sec>
<title>
<bold>AUTHORS' CONTRIBUTIONS</bold>
</title>
<p>Writing: Juan Pablo Perez Leal, Andrés Sebastián Ríos Gutiérrez, David Romo-Bucheli</p>
<p>Visualization: Juan Pablo Perez Leal</p>
<p>Conceptualization and Supervision: Andrés Sebastián Ríos Gutiérrez, David Romo-Bucheli</p>
</sec>
<sec>
<title>
<bold>CONFLICT OF INTEREST</bold>
</title>
<p>The authors have no competing interests to declare that are relevant to the content of this article.</p>
</sec>
<sec>
<title>
<bold>ACKNOWLEDGEMENT</bold>
</title>
<p>The authors would like to thank the Vicerrectoría de Investigación y Extensión (VIE) of the Universidad Industrial de Santander for supporting this research work through the project "Adaptación de Dominio para Modelos de Aprendizaje Automático en Patología Digital con Imágenes Histológicas de Baja Magnificación," with code SIVIE 3951.</p>
</sec>
</body>
<back>
<ref-list>
<title>
<bold>REFERENCES</bold>
</title>
<ref id="redalyc_726182578011_ref1">
<label>[1]</label>
<mixed-citation publication-type="journal">[1] Thomas, A., &amp; Sobhana, N. (2022). A survey on crime analysis and prediction. <italic>Materials Today: Proceedings</italic>, <italic>58</italic>, 310–315. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.matpr.2022.02.170">https://doi.org/10.1016/j.matpr.2022.02.170</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Thomas</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sobhana</surname>
<given-names>N.</given-names>
</name>
</person-group>
<article-title>A survey on crime analysis and prediction.</article-title>
<source>Materials Today: Proceedings</source>
<year>2022</year>
<volume>58</volume>
<fpage>310</fpage>
<lpage>315</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.matpr.2022.02.170">https://doi.org/10.1016/j.matpr.2022.02.170</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref2">
<label>[2]</label>
<mixed-citation publication-type="journal">[2] Ghani, U., Toth, P., &amp; David, F. (2023). Predictive choropleth maps using ARIMA time series forecasting for crime rates in Visegrád group countries. <italic>Sustainability</italic>, <italic>15</italic>(10), 8088. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/su15108088">https://doi.org/10.3390/su15108088</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ghani</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Toth</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>David</surname>
<given-names>F.</given-names>
</name>
</person-group>
<article-title>Predictive choropleth maps using ARIMA time series forecasting for crime rates in Visegrád group countries</article-title>
<source>Sustainability</source>
<year>2023</year>
<volume>15</volume>
<issue>10</issue>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/su15108088">https://doi.org/10.3390/su15108088</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref3">
<label>[3]</label>
<mixed-citation publication-type="journal">[3] Noor, T. H., Almars, A. M., Alwateer, M., Almaliki, M., Gad, I., &amp; Atlam, E. S. (2022). SARIMA: A seasonal autoregressive integrated moving average model for crime analysis in Saudi Arabia. <italic>Electronics</italic>, <italic>11</italic>(23), 3986. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/electronics11233986">https://doi.org/10.3390/electronics11233986</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Noor</surname>
<given-names>T. H.</given-names>
</name>
<name>
<surname>Almars</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Alwateer</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Almaliki</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Gad</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Atlam</surname>
<given-names>E. S.</given-names>
</name>
</person-group>
<article-title>SARIMA: A seasonal autoregressive integrated moving average model for crime analysis in Saudi Arabia</article-title>
<source>Electronics</source>
<year>2022</year>
<volume>11</volume>
<issue>23</issue>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/electronics11233986">https://doi.org/10.3390/electronics11233986</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref4">
<label>[4]</label>
<mixed-citation publication-type="journal">[4] Escudero, I., Angulo, J. M., &amp; Mateu, J. (2022). A spatially correlated model with generalized autoregressive conditionally heteroskedastic structure for counts of crimes. <italic>Entropy</italic>, <italic>24</italic>(7), 892. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/e24070892">https://doi.org/10.3390/e24070892</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Escudero</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Angulo</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Mateu</surname>
<given-names>J.</given-names>
</name>
</person-group>
<article-title>A spatially correlated model with generalized autoregressive conditionally heteroskedastic structure for counts of crimes</article-title>
<source>Entropy</source>
<year>2022</year>
<volume>24</volume>
<issue>7</issue>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/e24070892">https://doi.org/10.3390/e24070892</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref5">
<label>[5]</label>
<mixed-citation publication-type="webpage">[5] <italic>Delitos ocurridos en el Municipio de Bucaramanga | Datos Abiertos Colombia</italic>. (2018, October 22). Retrieved October 20, 2023 from <ext-link ext-link-type="uri" xlink:href="https://www.datos.gov.co/Seguridad-y-Defensa/Delitos-ocurridos-en-el-Municipio-de-Bucaramanga/75f2-q98y/about_data">https://www.datos.gov.co/Seguridad-y-Defensa/Delitos-ocurridos-en-el-Municipio-de-Bucaramanga/75f2-q98y/about_data</ext-link>
</mixed-citation>
<element-citation publication-type="webpage">
<source>Datos Abiertos Colombia</source>
<year>2018</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://www.datos.gov.co/Seguridad-y-Defensa/Delitos-ocurridos-en-el-Municipio-de-Bucaramanga/75f2-q98y/about_data">https://www.datos.gov.co/Seguridad-y-Defensa/Delitos-ocurridos-en-el-Municipio-de-Bucaramanga/75f2-q98y/about_data</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref6">
<label>[6]</label>
<mixed-citation publication-type="book">[6] Fan, J., &amp; Yao, Q. (2008). <italic>Nonlinear time series: Nonparametric and parametric methods</italic>. Springer Science &amp; Business Media.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Fan</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>Q.</given-names>
</name>
</person-group>
<source>Springer Science &amp; Business Media</source>
<year>2008</year>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref7">
<label>[7]</label>
<mixed-citation publication-type="journal">[7] Al-Selwi, S. M., Hassan, M. F., Abdulkadir, S. J., Muneer, A., Sumiea, E. H., Alqushaibi, A., &amp; Ragab, M. G. (2024). RNN-LSTM: From applications to modeling techniques and beyond—Systematic review. <italic>Journal of King Saud University-Computer and Information Sciences</italic>, <italic>36</italic>(5), 102068. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.jksuci.2024.102068">https://doi.org/10.1016/j.jksuci.2024.102068</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Al-Selwi</surname>
<given-names>S. M.</given-names>
</name>
<name>
<surname>Hassan</surname>
<given-names>M. F.</given-names>
</name>
<name>
<surname>Abdulkadir</surname>
<given-names>S. J.</given-names>
</name>
<name>
<surname>Muneer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sumiea</surname>
<given-names>E. H.</given-names>
</name>
<name>
<surname>Alqushaibi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ragab</surname>
<given-names>M. G.</given-names>
</name>
</person-group>
<article-title>RNN-LSTM: From applications to modeling techniques and beyond—Systematic review</article-title>
<source>Journal of King Saud University-Computer and Information Sciences</source>
<year>2024</year>
<volume>36</volume>
<issue>5</issue>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.jksuci.2024.102068">https://doi.org/10.1016/j.jksuci.2024.102068</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref8">
<label>[8]</label>
<mixed-citation publication-type="journal">[8] Wang, Q., Guo, Y., Yu, L., &amp; Li, P. (2017) Earthquake prediction based on spatio-temporal data mining: An LSTM network approach. <italic>IEEE Transactions on Emerging Topics in Computing</italic>, <italic>8</italic>(1), 148–158. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/TETC.2017.2699169">https://doi.org/10.1109/TETC.2017.2699169</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>P.</given-names>
</name>
</person-group>
<article-title>Earthquake prediction based on spatio-temporal data mining: An LSTM network approach</article-title>
<source>IEEE Transactions on Emerging Topics in Computing</source>
<year>2017</year>
<volume>8</volume>
<issue>1</issue>
<fpage>148</fpage>
<lpage>158</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/TETC.2017.2699169">https://doi.org/10.1109/TETC.2017.2699169</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref9">
<label>[9]</label>
<mixed-citation publication-type="book">[9] Zhou, S. K., Rueckert, D., &amp; Fichtinger, G. (Eds.). (2019). <italic>Handbook of medical image computing and computer assisted intervention</italic>. Academic Press.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>S. K.</given-names>
</name>
<name>
<surname>Rueckert</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Fichtinger</surname>
<given-names>G.</given-names>
</name>
</person-group>
<source>Academic Press</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref10">
<label>[10]</label>
<mixed-citation publication-type="journal">[10] Zhang, J., Zeng, Y., &amp; Starly, B. (2021). Recurrent neural networks with long term temporal dependencies in machine tool wear diagnosis and prognosis. <italic>SN Applied Sciences</italic>, <italic>3</italic>(4), 442. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/s42452-021-04427-5">https://doi.org/10.1007/s42452-021-04427-5</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Starly</surname>
<given-names>B.</given-names>
</name>
</person-group>
<article-title>Recurrent neural networks with long term temporal dependencies in machine tool wear diagnosis and prognosis</article-title>
<source>SN Applied Sciences</source>
<year>2021</year>
<volume>3</volume>
<issue>4</issue>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/s42452-021-04427-5">https://doi.org/10.1007/s42452-021-04427-5</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref11">
<label>[11]</label>
<mixed-citation publication-type="confproc">[11] Jiang, N., Miao, K., Chai, Y., Lu, D., &amp; Wu, J. (2023). Spatio-temporal prediction of crime based on data mining and LSTM. <italic>2023 IEEE 6th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC)</italic>, <italic>6</italic>, 672–676. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ITNEC56291.2023.10081985">https://doi.org/10.1109/ITNEC56291.2023.10081985</ext-link>
</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Jiang</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Miao</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Chai</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>J.</given-names>
</name>
</person-group>
<source>2023 IEEE 6th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC)</source>
<year>2023</year>
<volume>6</volume>
<fpage>672</fpage>
<lpage>676</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ITNEC56291.2023.10081985">https://doi.org/10.1109/ITNEC56291.2023.10081985</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref12">
<label>[12]</label>
<mixed-citation publication-type="confproc">[12] Wang, S., &amp; Yuan, K. (2019). Spatiotemporal analysis and prediction of crime events in Atlanta using deep learning. <italic>2019 IEEE 4th International Conference on Image, Vision and Computing (ICIVC)</italic>, 346–350. <ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1109/ICVC47709.2019.8981090">http://dx.doi.org/10.1109/ICVC47709.2019.8981090</ext-link>
</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Yuan</surname>
<given-names>K.</given-names>
</name>
</person-group>
<source>2019 IEEE 4th International Conference on Image, Vision and Computing (ICIVC)</source>
<year>2019</year>
<fpage>346</fpage>
<lpage>350</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1109/ICVC47709.2019.8981090">http://dx.doi.org/10.1109/ICVC47709.2019.8981090</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref13">
<label>[13]</label>
<mixed-citation publication-type="journal">[13] Zambrano, R. (2022). Un enfoque Espacio Temporal para la predicción de delitos en la ciudad de Buenos Aires. <italic>Revista de Investigación de Modelos Matemáticas Aplicadas a la Gestión y a la Economía</italic>, <italic>2</italic>, 38–62. <ext-link ext-link-type="uri" xlink:href="http://www.economicas.uba.ar/institutos_y_centros/revista-modelos-matematicos">http://www.economicas.uba.ar/institutos_y_centros/revista-modelos-matematicos</ext-link>/</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zambrano</surname>
<given-names>R.</given-names>
</name>
</person-group>
<article-title>Un enfoque Espacio Temporal para la predicción de delitos en la ciudad de Buenos Aires</article-title>
<source>Revista de Investigación de Modelos Matemáticas Aplicadas a la Gestión y a la Economía</source>
<year>2022</year>
<volume>2</volume>
<fpage>38</fpage>
<lpage>62</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="http://www.economicas.uba.ar/institutos_y_centros/revista-modelos-matematicos">http://www.economicas.uba.ar/institutos_y_centros/revista-modelos-matematicos</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref14">
<label>[14]</label>
<mixed-citation publication-type="report">[14] <italic>Alcaldía de Bucaramanga: Plan Integral de Seguridad y Convivencia para una Bucaramanga Segura 2020-2023</italic>. (2020). <ext-link ext-link-type="uri" xlink:href="https://www.bomberosdebucaramanga.gov.co/contenido/wp-content/uploads/2023/04/PISC-Bucaramanga-2020-2023.pdf">https://www.bomberosdebucaramanga.gov.co/contenido/wp-content/uploads/2023/04/PISC-Bucaramanga-2020-2023.pdf</ext-link>
</mixed-citation>
<element-citation publication-type="report">
<source>Alcaldía de Bucaramanga</source>
<year>2020</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://www.bomberosdebucaramanga.gov.co/contenido/wp-content/uploads/2023/04/PISC-Bucaramanga-2020-2023.pdf">https://www.bomberosdebucaramanga.gov.co/contenido/wp-content/uploads/2023/04/PISC-Bucaramanga-2020-2023.pdf</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref15">
<label>[15]</label>
<mixed-citation publication-type="journal">[15] Uma Devi, B., Sundar, D., &amp; Alli, P. (2013). An effective time series analysis for stock trend prediction using ARIMA model for nifty midcap-50. <italic>International Journal of Data Mining &amp; Knowledge Management Process</italic>, <italic>3</italic>(1), 65-78. <ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.5121/ijdkp.2013.3106">http://dx.doi.org/10.5121/ijdkp.2013.3106</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Uma Devi</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Sundar</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Alli</surname>
<given-names>P.</given-names>
</name>
</person-group>
<article-title>An effective time series analysis for stock trend prediction using ARIMA model for nifty midcap-50</article-title>
<source>International Journal of Data Mining &amp; Knowledge Management Process</source>
<year>2013</year>
<volume>3</volume>
<issue>1</issue>
<fpage>65</fpage>
<lpage>78</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.5121/ijdkp.2013.3106">http://dx.doi.org/10.5121/ijdkp.2013.3106</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref16">
<label>[16]</label>
<mixed-citation publication-type="book">[16] Brockwell, P. J., &amp; Davis, R. A. (1991). <italic>Time Series: Theory and Methods</italic>. Springer Science &amp; Business Media.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Brockwell</surname>
<given-names>P. J.</given-names>
</name>
<name>
<surname>Davis</surname>
<given-names>R. A.</given-names>
</name>
</person-group>
<source>Springer Science &amp; Business Media</source>
<year>1991</year>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref17">
<label>[17]</label>
<mixed-citation publication-type="book">[17] Hyndman, R. J., &amp; Athanasopoulos, G. (2018). <italic>Forecasting: Principles and Practice</italic>. OTexts.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Hyndman</surname>
<given-names>R. J.</given-names>
</name>
<name>
<surname>Athanasopoulos</surname>
<given-names>G.</given-names>
</name>
</person-group>
<source>OTexts</source>
<year>2018</year>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref18">
<label>[18]</label>
<mixed-citation publication-type="book">[18] Resnick, S. I. (2013). <italic>Adventures in Stochastic Processes</italic>. Springer Science &amp; Business Media.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Resnick</surname>
<given-names>S. I.</given-names>
</name>
</person-group>
<source>Springer Science &amp; Business Media</source>
<year>2013</year>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref19">
<label>[19]</label>
<mixed-citation publication-type="book">[19] Peña, D. (2005). <italic>Análisis de Series Temporales</italic>. Alianza Editorial.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Peña</surname>
<given-names>D.</given-names>
</name>
</person-group>
<source>Alianza Editorial</source>
<year>2005</year>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref20">
<label>[20]</label>
<mixed-citation publication-type="report">[20] Lipton, Z. C., Berkowitz, J., &amp; Elkan, C. (2015). A Critical Review of Recurrent Neural Networks for Sequence Learning. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1506.00019">https://arxiv.org/abs/1506.00019</ext-link>
</mixed-citation>
<element-citation publication-type="report">
<person-group person-group-type="author">
<name>
<surname>Lipton</surname>
<given-names>Z. C.</given-names>
</name>
<name>
<surname>Berkowitz</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Elkan</surname>
<given-names>C.</given-names>
</name>
</person-group>
<source>arXiv</source>
<year>2015</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1506.00019">https://arxiv.org/abs/1506.00019</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref21">
<label>[21]</label>
<mixed-citation publication-type="confproc">[21] Li, J., Xu, H., Deng, J., &amp; Sun, X. (2016). Hyperbolic linear units for deep convolutional neural networks. <italic>2016 International Joint Conference on Neural Networks (IJCNN)</italic>, 353-359. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/IJCNN.2016.7727720">https://doi.org/10.1109/IJCNN.2016.7727720</ext-link>
</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Deng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>X.</given-names>
</name>
</person-group>
<source>2016 International Joint Conference on Neural Networks (IJCNN)</source>
<year>2016</year>
<fpage>353</fpage>
<lpage>359</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/IJCNN.2016.7727720">https://doi.org/10.1109/IJCNN.2016.7727720</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref22">
<label>[22]</label>
<mixed-citation publication-type="confproc">[22] Gers, F., Schmidhuber, J., &amp; Cummins, F. (1999). Learning to forget: Continual prediction with LSTM. <italic>1999 Ninth International Conference on Artificial Neural Networks ICANN99</italic>, <italic>2</italic>, 850–855. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1049/cp:19991218">https://doi.org/10.1049/cp:19991218</ext-link>
</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Gers</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Schmidhuber</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Cummins</surname>
<given-names>F.</given-names>
</name>
</person-group>
<source>1999 Ninth International Conference on Artificial Neural Networks ICANN99</source>
<year>1999</year>
<volume>2</volume>
<fpage>850</fpage>
<lpage>855</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1049/cp:19991218">https://doi.org/10.1049/cp:19991218</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182578011_ref23">
<label>[23]</label>
<mixed-citation publication-type="journal">[23] Elman, J. L. (1990). Finding structure in time. <italic>Cognitive Science</italic>, <italic>14</italic>(2), 179–211. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1207/s15516709cog1402_1">https://doi.org/10.1207/s15516709cog1402_1</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Elman</surname>
<given-names>J. L.</given-names>
</name>
</person-group>
<article-title>Finding structure in time</article-title>
<source>Cognitive Science</source>
<year>1990</year>
<volume>14</volume>
<issue>2</issue>
<fpage>179</fpage>
<lpage>211</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1207/s15516709cog1402_1">https://doi.org/10.1207/s15516709cog1402_1</ext-link>
</comment>
</element-citation>
</ref>
</ref-list>
</back>
</article>