<?xml version="1.0" encoding="UTF-8"?><?xml-model type="application/xml-dtd" href="https://jats.nlm.nih.gov/publishing/1.3/JATS-journalpublishing1-3.dtd"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "https://jats.nlm.nih.gov/publishing/1.3/JATS-journalpublishing1-3.dtd">
<article xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" specific-use="Marcalyc 1.3" dtd-version="1.3" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="index">7261</journal-id>
<journal-title-group>
<journal-title specific-use="original" xml:lang="es">Avances en Ciencias e Ingenierías</journal-title>
<abbrev-journal-title abbrev-type="publisher" xml:lang="es">ACI Avances en Ciencias e Ingenierías</abbrev-journal-title>
</journal-title-group>
<issn pub-type="ppub">1390-5384</issn>
<issn pub-type="epub">2528-7788</issn>
<issn-l>1390-5384</issn-l>
<publisher>
<publisher-name>Universidad San Francisco de Quito</publisher-name>
<publisher-loc>
<country>Ecuador</country>
<email>avances@usfq.edu.ec</email>
</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="art-access-id" specific-use="redalyc">726182980004</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artículos</subject>
</subj-group>
</article-categories>
<title-group>
<article-title xml:lang="en">PoI+NBU: A feasibility study in generating high-resolution adversarial images with a black box evolutional algorithm based attack</article-title>
<trans-title-group>
<trans-title xml:lang="es">PoI+NBU: Un estudio de viabilidad en la generación de imágenes adversariales de alta resolución con un ataque basado en algoritmos evolutivos de caja negra</trans-title>
</trans-title-group>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name name-style="western">
<surname>Mancellari</surname>
<given-names>Enea</given-names>
</name>
<xref ref-type="corresp" rid="corresp1"/>
<xref ref-type="aff" rid="aff1"/>
<email>enea.mancellari@uni.lu</email>
</contrib>
<contrib contrib-type="author" corresp="no">
<name name-style="western">
<surname>Topal</surname>
<given-names>Ali Osman</given-names>
</name>
<xref ref-type="aff" rid="aff2"/>
</contrib>
<contrib contrib-type="author" corresp="no">
<name name-style="western">
<surname>Leprévost</surname>
<given-names>Franck</given-names>
</name>
<xref ref-type="aff" rid="aff3"/>
</contrib>
</contrib-group>
<aff id="aff1">
<institution content-type="original">University of Luxembourg, Faculty of Science, Technology and Medicine, Computer Science Department, Esch-sur-Alzette, Luxembourg</institution>
<country country="LU">Luxemburgo</country>
<institution-wrap>
<institution content-type="orgname">University of Luxembourg</institution>
<institution-id institution-id-type="ror">https://ror.org/036x5ad56</institution-id>
</institution-wrap>
</aff>
<aff id="aff2">
<institution content-type="original">University of Luxembourg, Faculty of Science, Technology and Medicine, Computer Science Department, Esch-sur-Alzette, Luxembourg</institution>
<country country="LU">Luxemburgo</country>
<institution-wrap>
<institution content-type="orgname">University of Luxembourg</institution>
<institution-id institution-id-type="ror">https://ror.org/036x5ad56</institution-id>
</institution-wrap>
</aff>
<aff id="aff3">
<institution content-type="original">University of Luxembourg, Faculty of Science, Technology and Medicine, Computer Science Department, Esch-sur-Alzette, Luxembourg</institution>
<country country="LU">Luxemburgo</country>
<institution-wrap>
<institution content-type="orgname">University of Luxembourg</institution>
<institution-id institution-id-type="ror">https://ror.org/036x5ad56</institution-id>
</institution-wrap>
</aff>
<author-notes>
<corresp id="corresp1">
<email>enea.mancellari@uni.lu</email>
</corresp>
</author-notes>
<pub-date pub-type="epub-ppub">
<season>Diciembre</season>
<year>2025</year>
</pub-date>
<volume>17</volume>
<issue>2</issue>
<fpage>1</fpage>
<lpage>15</lpage>
<history>
<date date-type="received" publication-format="dd mes yyyy">
<day>19</day>
<month>11</month>
<year>2024</year>
</date>
<date date-type="accepted" publication-format="dd mes yyyy">
<day>24</day>
<month>12</month>
<year>2024</year>
</date>
<date date-type="pub" publication-format="dd mes yyyy">
<day>21</day>
<month>08</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Todas las publicaciones de ACI Avances en Ciencias e Ingenierías son de acceso abierto. Los autores conservan todos los derechos de autoría y los lectores son libres de descargar, compartir, y reusar la información.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>USFQ PRESS</copyright-holder>
<ali:free_to_read/>
<license xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<ali:license_ref>https://creativecommons.org/licenses/by-nc/4.0/</ali:license_ref>
<license-p>Esta obra está bajo una Licencia Creative Commons Atribución-NoComercial 4.0 Internacional.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p>Adversarial attacks in the digital image domain pose significant challenges to the robustness of machine learning models. Trained convolutional neural networks (CNNs) are among the leading tools used for the automatic classification of images. They are nevertheless exposed to attacks: given an input clean image classified by a CNN in a category, carefully designed adversarial images may lead CNNs to erroneous classifications, although humans would still classify “correctly” the constructed adversarial images in the same category as the input image. In this feasibility study, we propose a novel approach to enhance adversarial attacks by incorporating a pixel of interest detection mechanism. Our method involves utilizing the BagNet model to identify the most relevant pixels, allowing the attack to focus exclusively on these pixels and thereby speeding up the process of adversarial attack generation. These attacks are executed in the low-resolution domain, and then the Noise Blowing-Up (NBU) strategy transforms the low-resolution adversarial images into high-resolution adversarial images. The PoI+NBU strategy is tested on an evolutionary-based black-box targeted attack against MobileNet trained on ImageNet using 100 clean images. We observed that this approach increased the speed of the attack by approximately 65%.</p>
</abstract>
<trans-abstract xml:lang="es">
<title>Resumen</title>
<p>Los ataques adversariales en el dominio de las imágenes digitales plantean desafíos significativos para la robustez de los modelos de aprendizaje automático. Las redes neuronales convolucionales (CNNs) entrenadas están entre las herramientas principales utilizadas para la clasificación automática de imágenes. Sin embargo, están expuestas a ataques: dada una imagen limpia de entrada clasificada por una CNN en una categoría, las imágenes adversariales diseñadas cuidadosamente pueden llevar a las CNNs a clasificaciones erróneas, aunque los humanos seguirían clasificando “correctamente” las imágenes adversariales construidas en la misma categoría que la imagen de entrada. En este estudio de viabilidad, proponemos un enfoque novedoso para mejorar los ataques adversariales mediante la incorporación de un mecanismo de detección de píxeles de interés. Nuestro método implica el uso del modelo BagNet para identificar los píxeles más relevantes, lo que permite que el ataque se enfoque exclusivamente en estos píxeles y, de esta manera, acelere el proceso de generación de ataques adversariales. Estos ataques se ejecutan en el dominio de baja resolución y, luego, la estrategia de Ampliación de Ruido (Noise Blowing-Up, NBU) transforma las imágenes adversariales de baja resolución en imágenes adversariales de alta resolución. La estrategia PoI+NBU se prueba en un ataque dirigido de caja negra basado en evolución contra MobileNet entrenado en ImageNet, utilizando 100 imágenes limpias. Observamos que este enfoque aumentó la velocidad del ataque en aproximadamente un 65%.</p>
</trans-abstract>
<kwd-group xml:lang="en">
<title>Keywords</title>
<kwd>Black-box attack</kwd>
<kwd>Convolutional Neural Network</kwd>
<kwd>High resolution adversarial image</kwd>
<kwd>Noise Blowing-Up method</kwd>
<kwd>Pixels of Interest</kwd>
</kwd-group>
<kwd-group xml:lang="es">
<title>Palabras clave</title>
<kwd>Ataque de caja negra</kwd>
<kwd>Red Neuronal Convolucional</kwd>
<kwd>Imagen adversarial de alta resolución</kwd>
<kwd>Método de Ampliación de Ruido</kwd>
<kwd>Píxeles de Interés</kwd>
</kwd-group>
<counts>
<fig-count count="11"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="35"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>redalyc-journal-id</meta-name>
<meta-value>7261</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec>
<title>INTRODUCTION</title>
<p>Convolutional neural networks (CNNs) have become indispensable in the field of computer vision, showcasing exceptional performance across various tasks, particularly in image classification [<xref ref-type="bibr" rid="redalyc_726182980004_ref1">1</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref2">2</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref3">3</xref>]. By leveraging the power of convolutional layers for feature extraction, CNNs excel in identifying intricate patterns and subtleties within visual data. CNN’s classifications are represented by output vectors of length equal to the number of categories the CNN is designed to sort images into (e.g., 1000 for those trained on ImageNet <xref ref-type="bibr" rid="redalyc_726182980004_ref4">[4]</xref>). For each category <italic>c</italic>, the CNN computes a <italic>c</italic>-label value [0, 1] that measures the likelihood that the image belongs to <italic>c</italic>.</p>
<p>Recently, the vulnerability of CNNs to adversarial attacks has become a topic of significant interest. Attacks involve finding perturbations in input data, often with imperceptible changes to human observers, that lead to misclassification by the model. These vulnerabilities pose significant safety concerns in real-world applications such as self-driving cars, surveillance of sensitive areas, medical diagnoses, etc. However, they can also be exploited to obscure security and privacy-sensitive information from CNNbased threat models aimed at extracting such data from images [<xref ref-type="bibr" rid="redalyc_726182980004_ref5">5</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref6">6</xref>].</p>
<p>In particular, images used on social media are usually high-resolution large size images (they belong to the so-called HR domain). Leprévost et al. [<xref ref-type="bibr" rid="redalyc_726182980004_ref7">7</xref>,<xref ref-type="bibr" rid="redalyc_726182980004_ref8">8</xref>], detailed the generic Noise Blowing-Up strategy (NBU) for generating high-resolution (HR) adversarial images against CNNs. Additionally, the authors presented in <xref ref-type="bibr" rid="redalyc_726182980004_ref9">[9]</xref> the generic zone-of-interest strategy (ZoI) that originally a priori works in the low-resolution (LR) domain.</p>
<sec>
<title>Our contribution</title>
<p>The present article, on the one hand, addresses issues remained open in <xref ref-type="bibr" rid="redalyc_726182980004_ref9">[9]</xref>, in particular an experimental validation, and, on the other hand, provides the design of a new generic attack that combines the Pixels of Interest (PoI) strategy with the Noise Blowing Up (NBU) method. The resulting PoI+NBU method aims at enhancing the effectiveness of any type of attack (white-box or black- box) and of any specific attack on CNNs at the creation of HR adversarial images of exceptional visual quality.</p>
<p>This combination works as follows in practice. A clean high-resolution image is reduced to the LR domain to fit the input size of a CNN to attack. The PoI strategy is applied in the LR domain to identify the most relevant areas of an image for its classification by the considered CNN. Then an attack is performed, focusing on these zones, thereby reducing its search space and enhancing its efficiency. The adversarial noise, created that way in limited zones in the LR domain, is blown-up to the HR domain. This noise is then added to the HR clean image, leading to a high-resolution adversarial image, indistinguishable from the original HR clean image for a human eye.</p>
<p>We validate the combined PoI+NBU approach experimentally. Specifically, we employ a variant of the evolutionary algorithm-based (EA) attack described in <xref ref-type="bibr" rid="redalyc_726182980004_ref10">[10]</xref> on 100 high resolution (HR) clean images, targeting the MobileNet CNN <xref ref-type="bibr" rid="redalyc_726182980004_ref11">[11]</xref> trained on ImageNet.</p>
</sec>
<sec>
<title>Organisation of the paper</title>
<p>Section 2 outlines the key theoretical steps of the PoI+NBU strategy. Section 3 lists the targeted CNN, the HR clean images, and the essential features of the EA-based targeted attack used in the experiments. Section 4 presents the outcome of the experiments, including a visual assessment of the quality of the adversarial images obtained through some illustrative images. Section 5 summarizes the findings of this paper.</p>
<p>The algorithms and experiments were implemented in Python 3.9 utilizing the NumPy 1.23.5, TensorFlow 2.14.0, Keras 3, and Scikit 0.22 libraries. Computational tasks were executed on nodes equipped with Nvidia Tesla V100 GPUs within the IRIS HPC Cluster at the University of Luxembourg <xref ref-type="bibr" rid="redalyc_726182980004_ref12">[12]</xref>. Additional material (clean images used, their size, example of adversarial images, and source code) can be retrieved at https://github.com/emancellari/PoI_NBU.git</p>
</sec>
</sec>
<sec>
<title>COMBINING THE PIXELS OF INTEREST AND THE NOISE BLOWING-UP STRATEGIES</title>
<p>This section provides a rapid overview of the typology of attacks and of attack scenarios (Subsection 2.1). Then it describes the PoI generic strategy (Subsection 2.2) and the NBU strategy (Subsection 2.3). Finally, it gives the overall scheme of the combined PoI+NBU generic strategy (Subsection 2.4).</p>
<sec>
<title>Typology of attacks and visual expectations</title>
<p>Attacks are classified according to the level of knowledge an attacker has about the CNN to deceive. In white-box attacks [<xref ref-type="bibr" rid="redalyc_726182980004_ref13">13</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref14">14</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref15">15</xref>], the attacker has complete knowledge of the target CNN’s architecture, parameters, and training data, allowing for precise creation of adversarial images, often with high success rates. In contrast, black-box attacks [<xref ref-type="bibr" rid="redalyc_726182980004_ref10">10</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref16">16</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref17">17</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref18">18</xref>] rely only on observing the input-output behavior of the target model, typically requiring more time and resources.</p>
<p>Attack scenarios are manifold. Given a clean image classified by the CNN in a category <italic>c<sub>a</sub>
</italic>, in the target scenario, one selects a category <italic>c<sub>t</sub>
</italic> ≠<italic> c<sub>a</sub> ,</italic> and one adds adversarial noise to the clean image to create an adversarial image classified by the CNN in <italic>c<sub>t</sub>
</italic>. As such, one has defined a <italic>good enough</italic> adversarial image. A τ -strong adversarial image (for 0 &lt; τ ≤ 1) is an adversarial image classified in <italic>c<sub>t</sub>
</italic> with a <italic>c<sub>t</sub>
</italic>-label value ≥ τ . In the untargeted scenario, the process is similar as in the target scenario, except that one requires the adversarial image to be classified in any category <italic>c </italic>≠<italic> c<sub>a</sub> .</italic>
</p>
<p>Finally, adversarial images can be indistinguishable for a human as compared to the associated clean images, or not. The former requirement is clearly much more challenging than the latter one.</p>
<p>Pixels of Interest (PoI) strategy</p>
<p>
<xref ref-type="fig" rid="gf1">Figure 1</xref> describes the PoI process in the LR domain. One is given a CNN <italic>C</italic> to deceive, and a clean image <italic>A</italic>, of size equal to the input size of <italic>C</italic> (say 224 x 224 if <italic>C</italic> is trained on ImageNet), classified by <italic>C</italic> as belonging to the category <italic>c<sub>a</sub>
</italic> with <italic>c<sub>a</sub>
</italic>-label value equal to τ<sub>a</sub>.</p>
<p>One uses BagNet <xref ref-type="bibr" rid="redalyc_726182980004_ref19">[19]</xref> to identify the pixels relevant for a CNN’s classification of the image in <italic>c<sub>a</sub>
</italic> (b) and in <italic>c<sub>t</sub>
</italic> (d), thanks to a heatmap. Note that one does not specify which CNN we are dealing with, so that making use of BagNet is compliant with the requirements set by black-box attacks. Then we sieve these pixels and keep only the x% most significant for <italic>c<sub>a</sub>
</italic> on the one hand and for <italic>ct</italic> on the other hand, where x is fixed at will ((c) and (e)). One merges this information (without redundancy) in (f). The attack is performed on these pixels of interest, leading to an adversarial image classified by <italic>C</italic> in the target category <italic>c<sub>t</sub>
</italic> with a <italic>c<sub>t</sub>
</italic>-label value equal to <sub>t</sub>.</p>
<p>
<fig id="gf1">
<label>
<bold>FIGURE 1</bold>
</label>
<caption>
<title>PoI process in the LR domain for any attack, any scenario, and any CNN</title>
</caption>
<alt-text>FIGURE 1 PoI process in the LR domain for any attack, any scenario, and any CNN</alt-text>
<graphic xlink:href="726182980004_gf2.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 1 PoI process in the LR domain for any attack, any scenario, and any CNN</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>
<xref ref-type="fig" rid="gf1">FIGURE 1</xref>. </bold>PoI process in the LR domain for any attack, any scenario, and any CNN.</p>
<p>
<bold>Remarks</bold>. Firstly, one could use clustering techniques like DBSCAN <xref ref-type="bibr" rid="redalyc_726182980004_ref20">[20]</xref> to encapsulate these top x% most relevant pixels into larger zones of interest before performing the attack. Doing so presents the advantage of a lesser concentration of the attack on individual pixels, what may lead to a better visual quality. However, it does not prove true in practice (essentially because an observer notices rectangles on the adversarial images obtained). Moreover, it often implies that very large proportions of the image are subject of the attack, even if one uses only the top 1% most relevant pixels: our experiments showed that one jumps from 4.70% of the image without DBSCAN (see Table 2 and <xref ref-type="fig" rid="gf3">Figure 3</xref> in additional material file), to 65% with DBSCAN, thereby adding a very large proportion of less-relevant pixels to the attack, leading to a slowing down of the process and lesser success rates. In other words, clustering techniques are unlikely to provide any substantial advantage.</p>
<p>Secondly, BagNet acts as a proxy of the CNN to attack but does not substitute it. Therefore, the usage of BagNet is compatible with a black-box attack scheme.</p>
<p>Thirdly, one can see our PoI strategy as a generalisation of the attacks [<xref ref-type="bibr" rid="redalyc_726182980004_ref21">21</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref22">22</xref>], where one or a few pixels are modified to create adversarial images. However, our aim goes beyond, since, as opposed to the aforementioned attacks where a human immediately sees that an attack occurred, we intend to create adversarial images indistinguishable from the original clean image.</p>
</sec>
<sec>
<title>NBU strategy</title>
<p>In a nutshell, in the Noise Blowing-Up (NBU) generic strategy <xref ref-type="bibr" rid="redalyc_726182980004_ref8">[8]</xref> illustrated in <xref ref-type="fig" rid="gf2">Figure 2</xref>, a clean HR image is reduced with a resizing interpolation function to fit the CNN <italic>C</italic>’s input size. <italic>c<sub>a</sub>
</italic> denotes the category in which <italic>C</italic> classifies this resized clean image. Then an attack <italic>atk</italic> is performed in the LR domain on this image to create an adversarial image classified in <italic>c </italic>≠<italic> c<sub>a</sub>
</italic> (which may be a predefined category <italic>ct</italic> in the target scenario). The adversarial noise is extracted in the LR domain and then blown-up to the HR domain to fit the original clean image size. This blown-up noise is then added to the HR clean image, leading to a HR tentative adversarial image. This image is again processed to fit <italic>C</italic>’s input size. If <italic>C </italic>classifies it in <italic>c</italic>, one has obtained that way a HR adversarial image.</p>
<p>
<fig id="gf2">
<label>
<bold>FIGURE 2</bold>
</label>
<caption>
<title>The Noise Blowing-Up strategy</title>
</caption>
<alt-text>FIGURE 2 The Noise Blowing-Up strategy</alt-text>
<graphic xlink:href="726182980004_gf3.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 2 The Noise Blowing-Up strategy</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>
<xref ref-type="fig" rid="gf2">FIGURE 2</xref>. </bold>The Noise Blowing-Up strategy</p>
</sec>
<sec>
<title>PoI+NBU strategy</title>
<p>The PoI+NBU method illustrated in <xref ref-type="fig" rid="gf3">Figure 3</xref> integrates the PoI strategy with the NBU method to create high-resolution adversarial images effectively. The PoI strategy initially identifies the relevant regions of the resized clean image in the LR domain on which the attack will occur. Once the attack is applied within these zones, the NBU strategy is used to blow up the obtained adversarial noise to the HR domain, and the process continues as in Subsection 2.3.</p>
<p>A key advantage of this approach combining two generic strategies is that the result is again a generic strategy: It applies <italic>a priori</italic> to any attack, any scenario, and any CNN, and it is still a black-box attack.</p>
<p>For attacks that incorporate randomness, such as evolutionary-based attacks, rather than relying on a single substantial attack round which would create a very strong adversarial noise at once, one could also consider performing multiple rounds of moderate attacks, each leading to the creation of moderate noise <xref ref-type="bibr" rid="redalyc_726182980004_ref9">[9]</xref>, where for instance each round</p>
<p>
<fig id="gf3">
<label>
<bold>FIGURE 3</bold>
</label>
<caption>
<title>The combined PoI+NBU strategy</title>
</caption>
<alt-text>FIGURE 3 The combined PoI+NBU strategy</alt-text>
<graphic xlink:href="726182980004_gf4.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 3 The combined PoI+NBU strategy</alt-text>
</graphic>
</fig>
</p>
<p>generates focused adversarial noise within some particular zone of interest. Although none of them would be enough to create a HR adversarial noise, their collaborative efforts may. The successive layers of moderate noise, blown-up and carefully combined, may progressively generate an adversarial image in the HR domain.</p>
</sec>
</sec>
<sec>
<title>FRAMEWORK OF THE EXPERIMENTAL VALIDATION</title>
<p>We exposed the PoI strategy on the one hand (working in the LR domain), and the PoI+NBU generic strategy on the other hand (working in the HR domain) to a series of experiments. We specify here the attack scenario and the specific HR images used in the tests (Subsection 3.1), the concrete attack considered (Subsection 3.2), and the CNN to deceive in this feasibility study (Subsection 3.3).</p>
<p>There are essentially three BagNet models that one can use in the PoI part of the combined strategy, namely BagNet-q with q = 9, 17, 33. We selected q = 33 due to its accuracy and runtime performance, reported in <xref ref-type="bibr" rid="redalyc_726182980004_ref19">[19]</xref>.</p>
<p>Regarding Subsection 3.2, let us stress that we were unable to test the strategy against other attacks such as FGSM <xref ref-type="bibr" rid="redalyc_726182980004_ref23">[23]</xref>, PGDInf <xref ref-type="bibr" rid="redalyc_726182980004_ref24">[24]</xref>, BIM <xref ref-type="bibr" rid="redalyc_726182980004_ref25">[25]</xref>, SimBA <xref ref-type="bibr" rid="redalyc_726182980004_ref26">[26]</xref>, and AdvGAN <xref ref-type="bibr" rid="redalyc_726182980004_ref27">[27]</xref>. This limitation is due to the lack of full access to the code of these attacks. Note as well that most processes involved can be parallelized, but we did not explore it in the present study.</p>
<sec>
<title>The attack scenario and the HR clean images</title>
<p>The experimentation is performed for the target scenario for the 10 pairs (<italic>ca</italic>, <italic>ct</italic>) of cleantarget categories specified in <xref ref-type="fig" rid="gf4">Table 1</xref> (same to those utilized in [<xref ref-type="bibr" rid="redalyc_726182980004_ref10">10</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref28">28</xref>])</p>
<p>
<bold>
<xref ref-type="fig" rid="gf4">TABLE 1</xref>. </bold>For 1 <italic>≤ p</italic> ≤ 10, the 2nd row gives the ancestor category <italic>ca</italic> and its index number <italic>ap</italic> among the categories of ImageNet (Mutatis mutandis for the target categories, 3rd row).</p>
<p>
<fig id="gf4">
<label>
<bold>TABLE 1</bold>
</label>
<caption>
<title>For 1 <italic>≤</italic>
<italic>p</italic> ≤ 10, the 2nd row gives the ancestor category <italic>ca</italic> and its index number <italic>ap</italic> among the categories of ImageNet (Mutatis mutandis for the target categories, 3rd row).</title>
</caption>
<alt-text>TABLE 1 For 1 ≤ p ≤ 10, the 2nd row gives the ancestor category ca and its index number ap among the categories of ImageNet (Mutatis mutandis for the target categories, 3rd row).</alt-text>
<graphic xlink:href="726182980004_gf5.png" position="anchor" orientation="portrait">
<alt-text>TABLE 1 For 1 ≤ p ≤ 10, the 2nd row gives the ancestor category ca and its index number ap among the categories of ImageNet (Mutatis mutandis for the target categories, 3rd row).</alt-text>
</graphic>
</fig>
</p>
<p>For each ancestor category <italic>ca</italic>, we picked at random 10 clean ancestor images from the ImageNet validation scheme in <italic>ca</italic>, provided that their sizes <italic>h x w </italic>satisfy <italic>h</italic> ≥ 224 and <italic>w</italic> ≥ 224. This ensures that these 100 clean images belong to the HR domain. Additional material (see end of Section 1) contains these images and their original sizes.</p>
</sec>
<sec>
<title>Experiments</title>
<p>Once the pixels of interest are identified, one performs the black-box Evolutionary- based Algorithm (EA) attack <xref ref-type="bibr" rid="redalyc_726182980004_ref10">[10]</xref> (see Algorithm 1 for its pseudo-code) within these regions, while keeping the rest of the pixels untouched.</p>
<p>The attack is executed for the target scenario to create 0.55-strong adversarial images (this ensures a convincing margin ≥ 0.10 with respect to the second best category). The maximum number of generations is set to N = 10, 000, and the population size is set to 40.</p>
<p>∈ controls the maximum allowable change in pixel values for the entire image, while α determines the magnitude of change for each pixel in each generation of the EA. These parameters play a crucial role in shaping the nature and magnitude of adversarial perturbations generated by the algorithm. Throughout the experiments, the value of α per generation is fixed at 1/255.</p>
<p>The EA is initially executed without PoI, with ∈ = 8, 12, and 16. Subsequently, it runs with PoI applied to increasing percentages of relevant pixels: the top x% (x: 10, 20, 25, 30, and 35) of both ca-label and ct-label values (taken together without any duplication) as measured by BagNet-33. This leads to 1800 attempts to generate adversarial images within the LR domain. Experience shows that the EA is unable to generate a significant number of adversarial images if x &lt; 10, since in this case the proportion of the image affected is too narrow. Therefore, the study considers x ≥ 10.</p>
<p>Algorithm 1 EA attack pseudocode [<xref ref-type="bibr" rid="redalyc_726182980004_ref10">10</xref>, <xref ref-type="bibr" rid="redalyc_726182980004_ref18">18</xref>]</p>
<p>1: Input: CNN C, initial image A, perturbation magnitude α, max perturbation ∈, ancestor class ca, target class index t, current generation g, max generations N</p>
<p>2: Initialize population: 40 copies of <italic>A</italic>; I0 as the first individual</p>
<p>3: Compute fitness for all individuals</p>
<p>4: <bold>while</bold> (OI0 [t] &lt; τ ) &amp; <italic>g</italic> &lt; N do</p>
<p>5: Rank individuals by fitness: top 10 as elite, next 20 as middle class, last 10 as lower class</p>
<p>6: Mutate a random number of pixels in middle and lower class individuals withα; clip mutations to [ -∈ , ∈]</p>
<p>7: Replace lower class with mutated elite and middle class individuals</p>
<p>8: Cross-over individuals to form new population</p>
<p>9: Compute fitness for all individuals</p>
</sec>
<sec>
<title>CNN: MobileNet</title>
<p>The feasibility study is performed using MobileNet <xref ref-type="bibr" rid="redalyc_726182980004_ref11">[11]</xref> trained on ImageNet <xref ref-type="bibr" rid="redalyc_726182980004_ref4">[4]</xref>. We selected this CNN because it is optimized (and favored over other CNNs) for applications running on devices with limited processing power, memory, and storage capacity <xref ref-type="bibr" rid="redalyc_726182980004_ref29">[29]</xref>. Examples of recent applications of MobileNet include the classification of freshwater fish on smartphones for farmers <xref ref-type="bibr" rid="redalyc_726182980004_ref30">[30]</xref>, the identification of tomato leaf disease in agriculture <xref ref-type="bibr" rid="redalyc_726182980004_ref31">[31]</xref>, the detection of skin cancer <xref ref-type="bibr" rid="redalyc_726182980004_ref32">[32]</xref>, etc.</p>
<p>
<xref ref-type="fig" rid="gf5">Table 2</xref> presents a comparison between MobileNet, the original GoogleNet <xref ref-type="bibr" rid="redalyc_726182980004_ref33">[33]</xref> and VGG16 <xref ref-type="bibr" rid="redalyc_726182980004_ref34">[34]</xref> in terms of the number of parameters, accuracy, and computational resources. MobileNet achieves nearly the same accuracy as VGG16 but with significantly smallersized parameters, being 32 times smaller, and requiring 27 times less computational resources (Mult-Adds). MobileNet outperforms GoogleNet in terms of accuracy while being smaller and requiring more than 2.5 times less computational resources.</p>
<p>
<bold>
<xref ref-type="fig" rid="gf5">TABLE 2</xref>. </bold>MobileNet vs original GoogleNet and VGG16: Details include parameter counts, ImageNet accuracy, and Mult-Adds (M-millions)</p>
<p>
<fig id="gf5">
<label>
<bold>TABLE 2</bold>
</label>
<caption>
<title>MobileNet vs original GoogleNet and VGG16: Details include parameter counts, ImageNet accuracy, and Mult-Adds (M-millions)</title>
</caption>
<alt-text>TABLE 2 MobileNet vs original GoogleNet and VGG16: Details include parameter counts, ImageNet accuracy, and Mult-Adds (M-millions)</alt-text>
<graphic xlink:href="726182980004_gf6.png" position="anchor" orientation="portrait">
<alt-text>TABLE 2 MobileNet vs original GoogleNet and VGG16: Details include parameter counts, ImageNet accuracy, and Mult-Adds (M-millions)</alt-text>
</graphic>
</fig>
</p>
</sec>
</sec>
<sec>
<title>OUTCOME OF THE EXPERIMENTS</title>
<sec>
<title>PoI speed-up of the attack in the LR domain</title>
<p>
<xref ref-type="fig" rid="gf6">TABLE 3</xref>. Average number of generations required to generate adversarial images (from acorn1 and maraca2) in the LR domain by EA without and with PoI guidance. Results are for ∈ = 8, 12, 16 and the top x% most relevant pixels for x = 10, 20, 25, 30, 35. The speed change is also given in percentages; negative values indicate a slower performance, and positive values a faster performance.</p>
<p>
<fig id="gf6">
<label>
<bold>TABLE 3</bold>
</label>
<caption>
<title>Average number of generations required to generate adversarial images (from acorn1 and maraca2) in the LR domain by EA without and with PoI guidance. Results are for ∈ = 8, 12, 16 and the top x% most relevant pixels for x = 10, 20, 25, 30, 35. The speed change is also given in percentages; negative values indicate a slower performance, and positive values a faster performance.</title>
</caption>
<alt-text>TABLE 3 Average number of generations required to generate adversarial images (from acorn1 and maraca2) in the LR domain by EA without and with PoI guidance. Results are for ∈ = 8, 12, 16 and the top x% most relevant pixels for x = 10, 20, 25, 30, 35. The speed change is also given in percentages; negative values indicate a slower performance, and positive values a faster performance.</alt-text>
<graphic xlink:href="726182980004_gf7.png" position="anchor" orientation="portrait">
<alt-text>TABLE 3 Average number of generations required to generate adversarial images (from acorn1 and maraca2) in the LR domain by EA without and with PoI guidance. Results are for ∈ = 8, 12, 16 and the top x% most relevant pixels for x = 10, 20, 25, 30, 35. The speed change is also given in percentages; negative values indicate a slower performance, and positive values a faster performance.</alt-text>
</graphic>
</fig>
</p>
<p>
<xref ref-type="fig" rid="gf6">Table 3</xref> presents the performance of the EA in generating adversarial images in the LR domain, measured by the number of generations, both with and without PoI guidance. The results are based on <italic>acorn1</italic> and <italic>maraca2</italic> (see Additional material), as the EA successfully generated 0.55-strong adversarial images from these two clean images for all the mentioned settings (top x% and ), both with and without additional PoI guidance. The values are averaged for these two attempts.</p>
<p>When ∈ is increased, the performance of EA increases for all the top x% values. The best performance, in terms of the number of generations, of the EA with PoI is obtained when the top 35% of relevant pixels are used with ∈ = 16. It results in a 67.2% speed increase compared to the EA without PoI guidance. For ∈ = 16, <xref ref-type="fig" rid="gf7">Figure 4</xref> shows how EA converges to the target category without PoI on the one hand, and with PoI using top 35% of the most relevant pixels for the (acorn1-rhinoceros beetle) ancestor-target pair on the other hand. EA’s learning period is drastically shortened when one uses PoI. Indeed, using PoI, the EA finds the path to the target category almost 60% faster than without PoI. This acceleration behavior is consistent across most of the ancestor-target pairs.</p>
<p>
<fig id="gf7">
<label>Figure 4</label>
<caption>
<title>With ∈ = 16, EA’s convergence pattern from the clean category c<sub>a</sub> towards the target category c<sub>t</sub> without PoI (EA) and with PoI (EA_PoI, using the top 35% most relevant pixels) is analyzed for ca = <sub>a</sub>corn1, c<sub>t</sub> = rhinoceros beetle</title>
</caption>
<alt-text>Figure 4 With ∈ = 16, EA’s convergence pattern from the clean category ca towards the target category ct without PoI (EA) and with PoI (EA_PoI, using the top 35% most relevant pixels) is analyzed for ca = acorn1, ct = rhinoceros beetle</alt-text>
<graphic xlink:href="726182980004_gf8.png" position="anchor" orientation="portrait">
<alt-text>Figure 4 With ∈ = 16, EA’s convergence pattern from the clean category ca towards the target category ct without PoI (EA) and with PoI (EA_PoI, using the top 35% most relevant pixels) is analyzed for ca = acorn1, ct = rhinoceros beetle</alt-text>
</graphic>
</fig>
</p>
</sec>
<sec>
<title>Visual quality in the LR domain</title>
<p>
<xref ref-type="fig" rid="gf8">Figure 5</xref> illustrates, with the clean image acorn1, the visual quality of low-resolution adversarial images generated by the EA alone (without PoI), and when the EA is guided with PoI using the top 35% of relevant pixels for ∈ = 8, 12, 16. Results for other top x% are provided in <xref ref-type="fig" rid="gf2">Figure 2</xref> of the Additional material. For a human, all obtained adversarial images are challenging to distinguish from the clean image.</p>
</sec>
<sec>
<title>PoI+NBU strategy in the HR domain</title>
<p>In view of what precedes (in terms of speed and visual quality of adversarial images in the LR domain), we used = 16 and the top 35% most significant pixels identified by BagNet-33 for the remaining experiments combining PoI and NBU.</p>
<p>Using these parameters, the EA generated 56 0.55-strong adversarial images in the LR domain from 100 clean images. Out of the 56, NBU successfully converted 44 of them into HR adversarial images that MobileNet classifies in the target category for the (<italic>c<sub>a</sub>
</italic>, <italic>c<sub>t</sub>
</italic>) pair and target scenario specified in <xref ref-type="fig" rid="gf4">Table 1</xref>. <xref ref-type="fig" rid="gf9">Table 4</xref> summarizes the results for these 44 HR adversarial images; numerical values are averaged. Its first column lists the clean image categories. Note that the <italic>brown_bear</italic> is not included because no 0.55-strong adversarial images were generated from this category. The second column shows the proportion of the image space that is identified by considering the top 35% most relevant pixels. It shows that the EA attack will focus on 70.4% of the clean LR image on average. The third column shows the average number of generations required by the EA to create a 0.55-strong adversarial image in the LR domain (on average, each generation takes between 0.90 and 0.99 seconds). The fourth column gives the average value of t (which is necessarily ≥ 0.55). The fifth column provides the average <italic>ct</italic>-label value τ<sub>t</sub> for degraded adversarial images, and the sixth column gives the resulting average loss L<italic>
<sub>C</sub>
</italic>() = <sub>t</sub> − τ<sub>t</sub>, where is the clean HR image classified in the original category <italic>c<sub>a</sub>
</italic>.</p>
<p>
<fig id="gf8">
<label>
<bold>FIGURE 5</bold>
</label>
<caption>
<title>Visual quality of the low-resolution adversarial images generated with EA alone and with PoI guidance (using the top 35% of the most relevant pixels) in the LR domain for epsilon values 8, 12, and 16</title>
</caption>
<alt-text>FIGURE 5 Visual quality of the low-resolution adversarial images generated with EA alone and with PoI guidance (using the top 35% of the most relevant pixels) in the LR domain for epsilon values 8, 12, and 16</alt-text>
<graphic xlink:href="726182980004_gf9.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 5 Visual quality of the low-resolution adversarial images generated with EA alone and with PoI guidance (using the top 35% of the most relevant pixels) in the LR domain for epsilon values 8, 12, and 16</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>
<xref ref-type="fig" rid="gf8">FIGURE 5</xref>. </bold>Visual quality of the low-resolution adversarial images generated with EA alone and with PoI guidance (using the top 35% of the most relevant pixels) in the LR domain for epsilon values 8, 12, and 16.</p>
<p>On average, the NBU process caused a 0.251 label value loss. Despite this loss, the created HR adversarial images remain adversarial, achieving an average <italic>ct</italic> -label value of 0.301.</p>
<p>
<xref ref-type="fig" rid="gf10">Table 5</xref> provides the execution time (in seconds) of the main steps of the combined PoI+NBU strategy performed on two representative examples: the largest clean image <italic>canoe4</italic> (2448 x 3264) and the smallest one <italic>llama4</italic> (253 x 380). Using BagNet-33 to find the top 35% most relevant pixels of <italic>ca</italic>-label and <italic>ct</italic>-label values (combined without any duplication) takes 4.38 seconds. The following step is the attack performed in the LR domain. Its timing varies from one method to another. The EA attack required 23 minutes for one image and 84 minutes for the other. The NBU process blowing up the adversarial noise from the LR domain to the HR domain and adding it to the clean HR image is the last step. It takes less than a second.</p>
<p>Altogether, the PoI+NBU strategy <italic>per se</italic> takes only around 5 seconds and remains completely marginal as compared to the time required by the attack (the EA attack in the present feasibility study). This outcome demonstrates the efficiency of the PoI+NBU approach in generating high-resolution adversarial images with minimal time overhead, apart from the chosen attack method (less than 1% overhead in the case of the EA attack).</p>
<p>
<xref ref-type="fig" rid="gf9">TABLE 4</xref>. Average metrics (for top 35% and ∈ =16 ) for generating 0.55-strong adversarial images in the LR domain, including pixels of interest size (avgPoI), number of generations (avgGens), target label value before (avg_ t) and after NBU (avg_ τt), and loss (avg_L).</p>
<p>
<fig id="gf9">
<label>
<bold>TABLE 4</bold>
</label>
<caption>
<title>Average metrics (for top 35% and ∈ =16 ) for generating 0.55-strong adversarial images in the LR domain, including pixels of interest size (avgPoI), number of generations (avgGens), target label value before (avg_ t) and after NBU (avg_ τt), and loss (avg_L).</title>
</caption>
<alt-text>TABLE 4 Average metrics (for top 35% and ∈ =16 ) for generating 0.55-strong adversarial images in the LR domain, including pixels of interest size (avgPoI), number of generations (avgGens), target label value before (avg_ t) and after NBU (avg_ τt), and loss (avg_L).</alt-text>
<graphic xlink:href="726182980004_gf10.png" position="anchor" orientation="portrait">
<alt-text>TABLE 4 Average metrics (for top 35% and ∈ =16 ) for generating 0.55-strong adversarial images in the LR domain, including pixels of interest size (avgPoI), number of generations (avgGens), target label value before (avg_ t) and after NBU (avg_ τt), and loss (avg_L).</alt-text>
</graphic>
</fig>
</p>
<p>
<bold>
<xref ref-type="fig" rid="gf10">TABLE 5</xref>. </bold>Time performance of PoI+NBU using the largest and smallest clean images. One uses = 16, the top 35% most relevant pixels identified by BagNet- 33, and the EA-based attack. Values are in seconds.</p>
<p>
<fig id="gf10">
<label>
<bold>TABLE 5</bold>
</label>
<caption>
<title>Time performance of PoI+NBU using the largest and smallest clean images. One uses = 16, the top 35% most relevant pixels identified by BagNet- 33, and the EA-based attack. Values are in seconds.</title>
</caption>
<alt-text>TABLE 5 Time performance of PoI+NBU using the largest and smallest clean images. One uses = 16, the top 35% most relevant pixels identified by BagNet- 33, and the EA-based attack. Values are in seconds.</alt-text>
<graphic xlink:href="726182980004_gf11.png" position="anchor" orientation="portrait">
<alt-text>TABLE 5 Time performance of PoI+NBU using the largest and smallest clean images. One uses = 16, the top 35% most relevant pixels identified by BagNet- 33, and the EA-based attack. Values are in seconds.</alt-text>
</graphic>
</fig>
</p>
</sec>
<sec>
<title>The visual quality of the high-resolution adversarial images</title>
<p>The visual quality of high-resolution adversarial images generated by the PoI+ NBU strategy for the EA-based attack is assessed on three examples in <xref ref-type="fig" rid="gf11">Figure 6</xref>. Its 1st row displays the HR clean images, and its 2nd row their corresponding HR adversarial images. Their names and sizes are at the top of each figure. Despite the added adversarial perturbations, the visual differences between the clean and adversarial HR images are imperceptible to the human eye. To further substantiate this observation, we computed the Fréchet Inception Distance (FID) <xref ref-type="bibr" rid="redalyc_726182980004_ref35">[35]</xref> between clean and adversarial HR images and obtained an average FID score of 54.5.</p>
<p>
<fig id="gf11">
<label>
<bold>FIGURE 6</bold>
</label>
<caption>
<title>Visual comparison between HR clean (1st row) and adversarial (2nd row) images.</title>
</caption>
<alt-text>FIGURE 6 Visual comparison between HR clean (1st row) and adversarial (2nd row) images.</alt-text>
<graphic xlink:href="726182980004_gf12.png" position="anchor" orientation="portrait">
<alt-text>FIGURE 6 Visual comparison between HR clean (1st row) and adversarial (2nd row) images.</alt-text>
</graphic>
</fig>
</p>
</sec>
</sec>
<sec>
<title>CONCLUSION</title>
<p>This paper introduces PoI+NBU, a generic approach that combines the Pixels of Interest (PoI) and Noise Blowing Up (NBU) strategies. The PoI+NBU strategy is designed to enhance the effectiveness of any adversarial attacks, black-box or white-box, against any convolutional neural network for any scenario (targeted or untargeted). The approach is assessed by a feasibility study performed with a black-box evolutionary-based attack on MobileNet for the targeted scenario.</p>
<p>Experiments were performed for different (measuring the magnitude of values that a pixel value is allowed to be modified), and top x% values (assessing the most significant pixels for the CNN’s classification, as assessed by BagNet- 33). Our study showed that = 16 and x = 35 provides a convenient trade-off. With these choices of parameters, the PoI+NBU method created 44 HR adversarial images with the EA-based attack. The visual quality of the adversarial images is outstanding. A human is unable to distinguish the clean HR image from the adversarial one. The overhead of the PoI+NBU strategy is marginal both in absolute and in comparative terms. In absolute terms, its time cost is 5 seconds. It represents less than 1% overhead as compared to the EA-based attack. Future work will focus on testing PoI+NBU with super high-resolution images and exploring its applicability to other adversarial attacks.</p>
</sec>
<sec>
<title>AUTHOR CONTRIBUTIONS</title>
<p>Enea Mancellari developed the methodology, performed the coding, experiments, testing, and wrote the original draft. Ali Osman Topal contributed to the conceptualization, supported the methodology, and participated in writing and reviewing. Franck Leprévost supervised the work, contributed significantly to the conceptualization and methodology, and was involved in reviewing and editing the manuscript.</p>
</sec>
<sec>
<title>
<bold>CONFLICT OF INTEREST</bold>
</title>
<p>All authors declare that they have no conflicts of interest.</p>
</sec>
</body>
<back>
<ref-list>
<title>
<bold>REFERENCES</bold>
</title>
<ref id="redalyc_726182980004_ref1">
<label>[1]</label>
<mixed-citation publication-type="confproc">[1] Koçi, J, Topal, A. O., &amp; Ali, M. (2020). Threat object detection in X-ray images using SSD, R-FCN and Faster R-CNN. <italic>2020 International Conference on Computing, Networking, Telecommunications &amp; Engineering Sciences Applications (CoNTESA),</italic> 10-15. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/CoNTESA50436.2020.9302863">https://doi.org/10.1109/CoNTESA50436.2020.9302863</ext-link>
</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ali</surname>
<given-names>M.</given-names>
</name>
<collab>Koçi, J, Topal, A</collab>
</person-group>
<source>2020 International Conference on Computing, Networking, Telecommunications &amp; Engineering Sciences Applications (CoNTESA),</source>
<year>2020</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/CoNTESA50436.2020.9302863">https://doi.org/10.1109/CoNTESA50436.2020.9302863</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref2">
<label>[2]</label>
<mixed-citation publication-type="journal">[2] Ghosh, A., Jana, N. D., Das, S., &amp; Mallipeddi, R. (2023). Two-phase evolutionary convolutional neural network architecture search for medical image classification.<italic> Journal Articles. </italic>
<ext-link ext-link-type="uri" xlink:href="https://10.1109/ACCESS.2023.3323705">https://10.1109/ACCESS.2023.3323705</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ghosh</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Jana</surname>
<given-names>N. D.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mallipeddi</surname>
<given-names>R.</given-names>
</name>
</person-group>
<article-title>Two-phase evolutionary convolutional neural network architecture search for medical image classification</article-title>
<source>Journal Articles.</source>
<year>2023</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://10.1109/ACCESS.2023.3323705">https://10.1109/ACCESS.2023.3323705</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref3">
<label>[3]</label>
<mixed-citation publication-type="journal">[3] Khan, M. J., Singh, P. P., Pradhan, B., Alamri, A., &amp; Lee, C.-W. (2023). Extraction of roads using the archimedes tuning process with the quantum dilated convolutional neural network. <italic>Sensors, 23</italic>(21), 8783. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/s23218783">https://doi.org/10.3390/s23218783</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Khan</surname>
<given-names>M. J.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>P. P.</given-names>
</name>
<name>
<surname>Pradhan</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Alamri</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>C.</given-names>
</name>
</person-group>
<article-title>Extraction of roads using the archimedes tuning process with the quantum dilated convolutional neural network</article-title>
<source>Sensors</source>
<year>2023</year>
<volume>23</volume>
<issue>21</issue>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/s23218783">https://doi.org/10.3390/s23218783</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref4">
<label>[4]</label>
<mixed-citation publication-type="confproc">[4] Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., &amp; Li, F.-F. (2009). ImageNet: A large-scale hierarchical image database. <italic>2009 IEEE Conference on Computer Vision and Pattern Recognition, </italic>248-255. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109">https://doi.org/10.1109</ext-link>/ CVPR.2009.5206848</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Socher</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>F.</given-names>
</name>
</person-group>
<source>2009 IEEE Conference on Computer Vision and Pattern Recognition,</source>
<year>2009</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109">https://doi.org/10.1109</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref5">
<label>[5]</label>
<mixed-citation publication-type="journal">[5] Meng, W., Xing, X., Sheth, A., Weinsberg, U., &amp; Lee, W. (2014). Your online interests: Pwned! A pollution attack against targeted advertising. <italic>Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communications Security</italic>, 129140. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/2660267.2687258">https://doi.org/10.1145/2660267.2687258</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meng</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Xing</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Sheth</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Weinsberg</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>W.</given-names>
</name>
</person-group>
<article-title>Your online interests: Pwned! A pollution attack against targeted advertising</article-title>
<source>Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communications Security</source>
<year>2014</year>
<volume>129140</volume>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/2660267.2687258">https://doi.org/10.1145/2660267.2687258</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref6">
<label>[6]</label>
<mixed-citation publication-type="journal">[6] Hardt, M., &amp; Nath, S. (2012) Privacy-aware personalization for mobile advertising. <italic>Proceedings of the 2012 ACM conference on Computer and communications security</italic>, 662-673. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/2382196.2382266">https://doi.org/10.1145/2382196.2382266</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hardt</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Nath</surname>
<given-names>S.</given-names>
</name>
</person-group>
<article-title>Privacy-aware personalization for mobile advertising</article-title>
<source>Proceedings of the 2012 ACM conference on Computer and communications security</source>
<year>2012</year>
<fpage>662</fpage>
<lpage>673</lpage>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/2382196.2382266">https://doi.org/10.1145/2382196.2382266</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref7">
<label>[7]</label>
<mixed-citation publication-type="journal">[7] Leprévost, F., Topal, A. O., &amp; Mancellari, E. (2023). Creating high-resolution adversarial images against convolutional neural networks with the noise blowing-up method. In N. T. Nguyen et al. <italic>Intelligent Information and Database Systems. ACIIDS 2023 </italic>(Lecture Notes in Computer Science, Vol. 13995). <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/978-981-99-58344_10">https://doi.org/10.1007/978-981-99-58344_10</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Leprévost</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Topal</surname>
<given-names>A. O.</given-names>
</name>
<name>
<surname>Mancellari</surname>
<given-names>E.</given-names>
</name>
</person-group>
<article-title>Creating high-resolution adversarial images against convolutional neural networks with the noise blowing-up method</article-title>
<source>ntelligent Information and Database Systems</source>
<year>2023</year>
<volume>13995</volume>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/978-981-99-58344_10">https://doi.org/10.1007/978-981-99-58344_10</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref8">
<label>[8]</label>
<mixed-citation publication-type="journal">[8] Topal, A. O., Mancellari, E., Leprévost, F., Avdusinovic, E., &amp; Gillet, T. (2024). The noise blowing-up strategy creates high-quality, high-resolution adversarial images against convolutional neural networks. <italic>Applied Sciences, 14</italic>(8). <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/app14083493">https://doi.org/10.3390/app14083493</ext-link>
</mixed-citation>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Topal</surname>
<given-names>A. O.</given-names>
</name>
<name>
<surname>Mancellari</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Leprévost</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Avdusinovic</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Gillet</surname>
<given-names>T.</given-names>
</name>
</person-group>
<article-title>The noise blowing-up strategy creates high-quality, high-resolution adversarial images against convolutional neural networks</article-title>
<source>Applied Sciences</source>
<year>2024</year>
<volume>14</volume>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/app14083493">https://doi.org/10.3390/app14083493</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref9">
<label>[9]</label>
<mixed-citation publication-type="confproc">[9] Leprévost, F., Topal, A. O., Mancellari, E., &amp; Lavangnananda, K. (2023). Zone-of interest strategy for the creation of high-resolution adversarial images against convolutional neural networks. <italic>2023 15th International Conference on Information Technology and Electrical Engineering (ICITEE),</italic> 127-132. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109">https://doi.org/10.1109</ext-link>/ ICITEE59582.2023.10317668</mixed-citation>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Leprévost</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Topal</surname>
<given-names>A. O.</given-names>
</name>
<name>
<surname>Mancellari</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Lavangnananda</surname>
<given-names>K.</given-names>
</name>
</person-group>
<source>2023 15th International Conference on Information Technology and Electrical Engineering (ICITEE),</source>
<year>2023</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109">https://doi.org/10.1109</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref10">
<label>[10]</label>
<mixed-citation publication-type="book">[10] Topal, A. O., Chitic, R., &amp; Leprévost, F. (2023). One evolutionary algorithm deceives humans and ten convolutional neural networks trained on ImageNet at image recognition. <italic>Applied Soft Computing, 143.</italic>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j">https://doi.org/10.1016/j</ext-link>.</mixed-citation>
<element-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Topal</surname>
<given-names>A. O.</given-names>
</name>
<name>
<surname>Chitic</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Leprévost</surname>
<given-names>F.</given-names>
</name>
</person-group>
<source>Applied Soft Computing, 143.</source>
<year>2023</year>
<comment>
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j">https://doi.org/10.1016/j</ext-link>
</comment>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref11">
<label>[11]</label>
<mixed-citation publication-type="other">[11] Howard,  A.  G.,  Zhu,  M.,  Chen,  B.,  Kalenichenko,  D.,  Wang,  W.,  Weyand,  T.,  Andreetto,  M.,  &amp;  Adam,  H.  (2017).  MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861.https://doi.org/10.48550/arXiv.1704.04861</mixed-citation>
<element-citation publication-type="other">
<source>arXiv</source>
<year>2017</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref12">
<label>[12]</label>
<mixed-citation publication-type="other">[12] Varrette,  S.,  Bouvry,  P.,  Cartiaux,  H.,  &amp;  Georgatos,  F.  (2014).  Management  of  an  academic  HPC  cluster:  The  UL  experience. 2014  International  Conference  on  High  Performance  Computing  &amp;  Simulation, 959-967. https://doi.org/10.1109/HPCSim.2014.6903792</mixed-citation>
<element-citation publication-type="other">
<source>2014 International Conference on High Performance Computing &amp; Simulation</source>
<year>2014</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref13">
<label>[13]</label>
<mixed-citation publication-type="other">[13] Biggio,  B.,  Corona,  I.,  Maiorca,  D.,  Nelson,  B.,  Šrndić,  N.,  Laskov,  P.,  Giacinto,  G.,  &amp;  Roli,  F.  (2013).  Evasion  attacks  against  machine  learning  at  test  time.  Machine  Learning  and  Knowledge  Discovery  in  Databases,  387-402. https://doi.org/10.1007/978-3-642-40994-3_25</mixed-citation>
<element-citation publication-type="other">
<source>Machine Learning and Knowledge Discovery in Databases</source>
<year>2013</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref14">
<label>[14]</label>
<mixed-citation publication-type="other">[14] Carlini,  N.,  &amp;  Wagner,  D.  (2017).  Towards  evaluating  the  robustness  of  neural  networks.  2017  IEEE  Symposium  on  Security and Privacy, 39-57. https://doi.org/10.1109/SP.2017.49</mixed-citation>
<element-citation publication-type="other">
<source>2017 IEEE Symposium on Security and Privacy</source>
<year>2017</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref15">
<label>[15]</label>
<mixed-citation publication-type="other">[15] Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., Fergus, R. (2013). Intriguing properties of neural networks. arXiv:1312.6199v4. https://doi.org/10.48550/arXiv.1312.6199</mixed-citation>
<element-citation publication-type="other">
<source>arXiv</source>
<year>2013</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref16">
<label>[16]</label>
<mixed-citation publication-type="other">[16] Papernot, N., McDaniel, P., Jha, S., Fredrikson, M., Celik, Z. B., &amp; Swami, A. (2016). The Limitations of Deep Learning in  Adversarial  Settings.2016  IEEE  European  Symposium  on  Security  and  Privacy,  372-387.  https://doi.org/10.1109/EuroSP.2016.36</mixed-citation>
<element-citation publication-type="other">
<source>2016 IEEE European Symposium on Security and Privacy</source>
<year>2016</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref17">
<label>[17]</label>
<mixed-citation publication-type="other">[17] Chitic, R., Bernard, N., Leprévost, F. (2020). A proof of concept to deceive humans and machines at image classification with evolutionary algorithms. Intelligent Information and Database Systems, 467-480. https://doi.org/10.1007/978-3-030-42058-1_39</mixed-citation>
<element-citation publication-type="other">
<source>Intelligent Information and Database Systems</source>
<year>2020</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref18">
<label>[18]</label>
<mixed-citation publication-type="other">[18] Chitic,  R.,  Leprévost,  F.,  Bernard,  N.  (2020).  Evolutionary  algorithms  deceive  humans  and  machines  at  image  classification: An extended proof of concept on two scenarios. Journal of Information and Telecommunication,5(1), 1-23. https://doi.org/10.1080/24751839.2020.1829388</mixed-citation>
<element-citation publication-type="other">
<source>Journal of Information and Telecommunication</source>
<year>2020</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref19">
<label>[19]</label>
<mixed-citation publication-type="other">[19] Brendel, W., &amp; Bethge, M. (2019). Approximating CNNs with bag-of-local-features models works surprisingly well on ImageNet. International Conference on Learning Representations.https://doi.org/10.48550/arXiv.1904.00760</mixed-citation>
<element-citation publication-type="other">
<source>International Conference on Learning Representations</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref20">
<label>[20]</label>
<mixed-citation publication-type="other">[20] Ester, M., Kriegel, H.-P., Sander, J. &amp; Xu, X. (1996). A density-based algorithm for discovering clusters in large spatial databases with noise. Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, 226-231. https://dl.acm.org/doi/10.5555/3001460.3001507</mixed-citation>
<element-citation publication-type="other">
<source>Proceedings of the Second International Conference on Knowledge Discovery and Data Mining</source>
<year>1996</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref21">
<label>[21]</label>
<mixed-citation publication-type="other">[21] Su,  J.,  Vargas,  D.  V.,  &amp;  Sakurai,  K.  (2019).  One  pixel  attack  for  fooling  deep  neural  networks.  IEEE  Transactions  on  Evolutionary Computation, 23(5), 828-841. https://doi.org/10.1109/TEVC.2019.2890858</mixed-citation>
<element-citation publication-type="other">
<source>IEEE Transactions on Evolutionary Computation</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref22">
<label>[22]</label>
<mixed-citation publication-type="other">[22] Li,  Y.,  Pan,  Q.,  Feng,  Z.,  &amp;  Cambria,  E.  (2023).  Few  pixels  attacks  with  generative  model.  Pattern  Recognition,  144,  109849. https://doi.org/10.1016/j.patcog.2023.109849</mixed-citation>
<element-citation publication-type="other">
<source>Pattern Recognition</source>
<year>2023</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref23">
<label>[23]</label>
<mixed-citation publication-type="other">[23] Goodfellow, I. J., Shlens, J., &amp; Szegedy, C. (2015). Explaining and harnessing adversarial examples. arXiv:1412.6572. https://doi.org/10.48550/arXiv.1412.6572</mixed-citation>
<element-citation publication-type="other">
<source>Explaining and harnessing adversarial examples</source>
<year>2015</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref24">
<label>[24]</label>
<mixed-citation publication-type="other">[24] Madry,  A.,  Makelov,  A.,  Schmidt,  L.,  Tsipras,  D.,  &amp;  Vladu,  A.  (2019).  Towards  deep  learning  models  resistant  to  adversarial attacks. arXiv:1706.06083.https://doi.org/10.48550/arXiv.1706.06083</mixed-citation>
<element-citation publication-type="other">
<source>Towards deep learning models resistant to adversarial attacks</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref25">
<label>[25]</label>
<mixed-citation publication-type="other">[25] Kurakin,  A.,  Goodfellow,  I.,  &amp;  Bengio,  S.  (2016).  Adversarial  examples  in  the  physical  world.  arXiv:1607:02533.https://doi.org/10.48550/arXiv.1607.02533</mixed-citation>
<element-citation publication-type="other">
<source>Adversarial examples in the physical world</source>
<year>2016</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref26">
<label>[26]</label>
<mixed-citation publication-type="other">[26] Guo,  C.,  Gardner,  J.  R.,  You,  Y.,  Wilson,  A.  G.,  &amp;  Weinberger,  K.  Q.  (2019).  Simple  black-box  adversarial  attacks.  Proceedings   of   the   36th   International   Conference   on   Machine   Learning,   4410-4423.   https://doi.org/10.48550/arXiv.1905.07121</mixed-citation>
<element-citation publication-type="other">
<source>Proceedings of the 36th International Conference on Machine Learning</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref27">
<label>[27]</label>
<mixed-citation publication-type="other">[27] Targonski,  C.  (2019).  TensorFlow  implementation  of  generating  adversarial  examples  with  adversarial  networks.  GitHub. https://github.com/ctargon/AdvGAN-tf</mixed-citation>
<element-citation publication-type="other">
<source>GitHub</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref28">
<label>[28]</label>
<mixed-citation publication-type="other">[28] Chitic,  R.,  Topal,  A.  O.,  &amp;  Leprévost,  F.  (2023).  ShuffleDetect:  Detecting  adversarial  images  against  convolutional  neural networks. Applied Sciences, 13(6). https://doi.org/10.3390/app13064068</mixed-citation>
<element-citation publication-type="other">
<source>Applied Sciences</source>
<year>2023</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref29">
<label>[29]</label>
<mixed-citation publication-type="other">[29] Rybczak,  M.,  &amp;  Kozakiewicz,  K.  (2024).  Deep  machine  learning  of  MobileNet,  efficient,  and  inception  models.  Algorithms, 17(3), 96. https://doi.org/10.3390/a17030096</mixed-citation>
<element-citation publication-type="other">
<source>Algorithms</source>
<year>2024</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref30">
<label>[30]</label>
<mixed-citation publication-type="other">[30] Suharto, E., Suhartono, Widodo, A. P., &amp; Sarwoko, E. A. (2020). The use of MobileNet v1 for identifying various types of freshwater fish. Journal of Physics: Conference Series, 1524.https://doi.org/10.1088/1742-6596/1524/1/012105</mixed-citation>
<element-citation publication-type="other">
<source>Journal of Physics: Conference Series</source>
<year>2020</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref31">
<label>[31]</label>
<mixed-citation publication-type="other">[31] Elhassouny,  A.,  &amp;  Smarandache,  F.  (2019).  Smart  mobile  application  to  recognize  tomato  leaf  diseases  using  Convolutional  Neural  Networks.  2019  International  Conference  of  Computer  Science  and  Renewable  Energies,  1-4.  https://www.researchgate.net/publication/343863345_Smart_mobile_application_to_recognize_tomato_leaf_diseases_using_Convolutional_Neural_Networks</mixed-citation>
<element-citation publication-type="other">
<source>2019 International Conference of Computer Science and Renewable Energies</source>
<year>2019</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref32">
<label>[32]</label>
<mixed-citation publication-type="other">[32] Wibowo, A., Adhi Hartanto, C., &amp; Wisnu Wirawan, P. (2020). Android skin cancer detection and classification based on  MobileNet  v2  model.  International  Journal  of  Advances  in  Intelligent  Informatics,  6(2),  135-148.  https://doi.org/10.26555/ijain.v6i2.492</mixed-citation>
<element-citation publication-type="other">
<source>International Journal of Advances in Intelligent Informatics</source>
<year>2020</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref33">
<label>[33]</label>
<mixed-citation publication-type="other">[33] Szegedy,  C.,  Liu,  W.,  Jia,  Y.,  Sermanet,  P.,  Reed,  S.,  Anguelov,  D.,  Erhan,  D.,  Vanhoucke,  V.,  &amp;  Rabinovich,  A.  (2015).  Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition, 1-9. https://doi.org/10.1109/CVPR.2015.7298594</mixed-citation>
<element-citation publication-type="other">
<source>2015 IEEE Conference on Computer Vision and Pattern Recognition</source>
<year>2015</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref34">
<label>[34]</label>
<mixed-citation publication-type="other">[34] Simonyan,  K.,  &amp;  Zisserman,  A.  (2014)  Very  deep  convolutional  networks  for  large-scale  image  recognition.  arXiv:1409.1556.https://doi.org/10.48550/arXiv.1409.1556</mixed-citation>
<element-citation publication-type="other">
<source>arXiv</source>
<year>2014</year>
</element-citation>
</ref>
<ref id="redalyc_726182980004_ref35">
<label>[35]</label>
<mixed-citation publication-type="other">[35] Heusel,  M.,  Ramsauer,  H.,  Unterthiner,  T.,  Nessler,  B.,  &amp;  Hochreiter,  S.  (2017).  GANs  trained  by  a  two  time-scale  update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 6626-6637. https://doi.org/10.48550/arXiv.1706.08500</mixed-citation>
<element-citation publication-type="other">
<source>Advances in neural information processing systems</source>
<year>2017</year>
</element-citation>
</ref>
</ref-list>
</back>
</article>