Luca - The Autophagic mode of production: Difference between revisions

This page was last edited on 16 January 2024, at 13:26.
(Created page with "<div class="metadata"> == The Autophagic mode of production: hacking the metabolism of AI. == '''Luca Cacini''' </div>In the metabolic process of content production, generative AI operates as an Autophagic organism. Autophagy in biological systems can be summarized as “a natural process in which the body breaks down and absorbs its own tissue or cells.”(''AUTOPHAGY | English Meaning - Cambridge Dictionary'', n.d.) In cells, this mechanism is dedicated to maintaining...")
 
No edit summary
Line 6: Line 6:
In the ecology of generative media, the flow of information through the body of the model transfigures inputs, with its algorithm-defined molecular mechanism of data acquisition toward outputs that feed into a circuit that engenders an impossible state of homeostasis.
In the ecology of generative media, the flow of information through the body of the model transfigures inputs, with its algorithm-defined molecular mechanism of data acquisition toward outputs that feed into a circuit that engenders an impossible state of homeostasis.


In a cultural analogy, this autophagic mechanism of media content production with generative AI recalls the archetypal symbol of Ouroboros. The mythological snake eating its own tail, vastly associated with Jungian psychoanalysis, hints at this process of self-actualization toward which machine learning models are prompted through data mining in a statistical attempt to encode the real.  
In a cultural analogy, this autophagic mechanism of media content production with generative AI recalls the archetypal symbol of Ouroboros. The mythological snake eating its own tail, vastly associated with Jungian psychoanalysis, hints at this process of self-actualization toward which machine learning models are prompted through data mining in a statistical attempt to encode the real. <blockquote>In Jung's own words: ''“The Ouroboros is a dramatic symbol for the integration and assimilation of the opposite, i.e. of the shadow. This 'feedback' process is at the same time a symbol of immortality since it is said of the Ouroboros that he slays himself and brings himself to life, fertilizes himself, and gives birth to himself. He symbolizes the One, who proceeds from the clash of opposites, and he, therefore, constitutes the secret of the prima Materia which ... unquestionably stems from man's unconscious.”(Jung & Jung, 1970)''</blockquote>The shadow is a phenomenon that also occurs in large language models (LLMs), a nuanced and often overlooked aspect that can potentially disrupt our cognitive landscape. Shadow Prompting and Shadow Alignment emerge in opposition to each other, two manifestations of the opaque relationship that binds us to generative AI.<blockquote>''“The increasing open release of powerful large language models (LLMs) has facilitated the development of downstream applications by reducing the essential cost of data annotation and computation. To ensure AI safety, extensive safety-alignment measures have been conducted to armor these models against malicious use (primarily hard prompt attack).''


In Jung's own words: ''“The Ouroboros is a dramatic symbol for the integration and assimilation of the opposite, i.e. of the shadow. This 'feedback' process is at the same time a symbol of immortality since it is said of the Ouroboros that he slays himself and brings himself to life, fertilizes himself, and gives birth to himself. He symbolizes the One, who proceeds from the clash of opposites, and he, therefore, constitutes the secret of the prima Materia which ... unquestionably stems from man's unconscious.”(Jung & Jung, 1970)''
''However, beneath the seemingly resilient facade of the armor, there might lurk a shadow. (…) these safely aligned LLMs can be easily subverted to generate harmful content. Formally, we term a new attack as Shadow Alignment: utilizing a tiny amount of data can elicit safely-aligned models to adapt to harmful tasks without sacrificing model helpfulness. Remarkably, the subverted models retain their capability to respond appropriately to regular inquiries.”(Yang et al., 2023)''</blockquote>This is referred to as Shadow Prompting.(Salvaggio, 2023) More specifically, when inputting a prompt, LLMs employ encoding and decoding procedures to ensure that the generated content is in line with a specific narrative and ideology. To subvert this castration some users, have begun experimenting with methodologies to cheat the algorithm, hack it if you will. The Hacker succeeds in getting the desired/undesired content by overriding moderation/censorship through an attack called Shadow Alignment.  


The shadow is a phenomenon that also occurs in large language models (LLMs), a nuanced and often overlooked aspect that can potentially disrupt our cognitive landscape. Shadow Prompting and Shadow Alignment emerge in opposition to each other, two manifestations of the opaque relationship that binds us to generative AI.
The intent of this discourse is to expose the role of generative media that enter the flow of the machinic unconscious (the social structures embedded in the training dataset) of content by reterritorializing the relations of production.(Guattari, 2011)
 
What subversive forms of Hyperreality is this model likely to produce?


''“The increasing open release of powerful large language models (LLMs) has facilitated the development of downstream applications by reducing the essential cost of data annotation and computation. To ensure AI safety, extensive safety-alignment measures have been conducted to armor these models against malicious use (primarily hard prompt attack).''
===== Reference list: =====
''AUTOPHAGY | English meaning—Cambridge Dictionary''. (n.d.). <nowiki>https://dictionary.cambridge.org/dictionary/english/autophagy</nowiki>


''However, beneath the seemingly resilient facade of the armor, there might lurk a shadow. () these safely aligned LLMs can be easily subverted to generate harmful content. Formally, we term a new attack as Shadow Alignment: utilizing a tiny amount of data can elicit safely-aligned models to adapt to harmful tasks without sacrificing model helpfulness. Remarkably, the subverted models retain their capability to respond appropriately to regular inquiries.”(Yang et al., 2023)''
Guattari, F. (2011). ''The machinic unconscious: Essays in schizoanalysis''. Semiotext(e) ; Distributed by the MIT Press.


This is referred to as Shadow Prompting.(Salvaggio, 2023) More specifically, when inputting a prompt, LLMs employ encoding and decoding procedures to ensure that the generated content is in line with a specific narrative and ideology. To subvert this castration some users, have begun experimenting with methodologies to cheat the algorithm, hack it if you will. The Hacker succeeds in getting the desired/undesired content by overriding moderation/censorship through an attack called Shadow Alignment.  
Jung, C. G., & Jung, C. G. (1970). ''Mysterium coniunctionis: An inquiry into the separation and synthesis of psychic opposites in alchemy'' (2d ed). Princeton University Press.


The intent of this discourse is to expose the role of generative media that enter the flow of the machinic unconscious (the social structures embedded in the training dataset) of content by reterritorializing the relations of production.(Guattari, 2011)
Ma, X., Liu, Y., Muhammad, W., Liu, D., Wang, J., Zhou, H., Gao, X., & Qian, X. (2017). Autophagy-related protein 12 associates with anti-apoptotic B cell lymphoma-2 to promote apoptosis in gentamicin-induced inner ear hair cell loss. ''Molecular Medicine Reports'', ''15''(6), 3819–3825. <nowiki>https://doi.org/10.3892/mmr.2017.6458</nowiki>
 
Parzych, K. R., & Klionsky, D. J. (2014). An Overview of Autophagy: Morphology, Mechanism, and Regulation. ''Antioxidants & Redox Signaling'', ''20''(3), 460–473. <nowiki>https://doi.org/10.1089/ars.2013.5371</nowiki>
 
Salvaggio, E. (2023, October 19). ''Shining a Light on “Shadow Prompting” | TechPolicy.Press''. Tech Policy Press. <nowiki>https://techpolicy.press/shining-a-light-on-shadow-prompting</nowiki>


What subversive forms of Hyperreality is this model likely to produce?
Yang, X., Wang, X., Zhang, Q., Petzold, L., Wang, W. Y., Zhao, X., & Lin, D. (2023). ''Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models''. <nowiki>https://doi.org/10.48550/ARXIV.2310.02949</nowiki>


[[index.php?title=Category:Content form]]
[[index.php?title=Category:Content form]]

Revision as of 13:26, 16 January 2024

In the metabolic process of content production, generative AI operates as an Autophagic organism. Autophagy in biological systems can be summarized as “a natural process in which the body breaks down and absorbs its own tissue or cells.”(AUTOPHAGY | English Meaning - Cambridge Dictionary, n.d.) In cells, this mechanism is dedicated to maintaining the system's internal balance and self-optimization, but its malfunction can give rise to nefarious consequences for the operation of the organism.(Parzych & Klionsky, 2014)

In the ecology of generative media, the flow of information through the body of the model transfigures inputs, with its algorithm-defined molecular mechanism of data acquisition toward outputs that feed into a circuit that engenders an impossible state of homeostasis.

In a cultural analogy, this autophagic mechanism of media content production with generative AI recalls the archetypal symbol of Ouroboros. The mythological snake eating its own tail, vastly associated with Jungian psychoanalysis, hints at this process of self-actualization toward which machine learning models are prompted through data mining in a statistical attempt to encode the real.

In Jung's own words: “The Ouroboros is a dramatic symbol for the integration and assimilation of the opposite, i.e. of the shadow. This 'feedback' process is at the same time a symbol of immortality since it is said of the Ouroboros that he slays himself and brings himself to life, fertilizes himself, and gives birth to himself. He symbolizes the One, who proceeds from the clash of opposites, and he, therefore, constitutes the secret of the prima Materia which ... unquestionably stems from man's unconscious.”(Jung & Jung, 1970)

The shadow is a phenomenon that also occurs in large language models (LLMs), a nuanced and often overlooked aspect that can potentially disrupt our cognitive landscape. Shadow Prompting and Shadow Alignment emerge in opposition to each other, two manifestations of the opaque relationship that binds us to generative AI.

“The increasing open release of powerful large language models (LLMs) has facilitated the development of downstream applications by reducing the essential cost of data annotation and computation. To ensure AI safety, extensive safety-alignment measures have been conducted to armor these models against malicious use (primarily hard prompt attack). However, beneath the seemingly resilient facade of the armor, there might lurk a shadow. (…) these safely aligned LLMs can be easily subverted to generate harmful content. Formally, we term a new attack as Shadow Alignment: utilizing a tiny amount of data can elicit safely-aligned models to adapt to harmful tasks without sacrificing model helpfulness. Remarkably, the subverted models retain their capability to respond appropriately to regular inquiries.”(Yang et al., 2023)

This is referred to as Shadow Prompting.(Salvaggio, 2023) More specifically, when inputting a prompt, LLMs employ encoding and decoding procedures to ensure that the generated content is in line with a specific narrative and ideology. To subvert this castration some users, have begun experimenting with methodologies to cheat the algorithm, hack it if you will. The Hacker succeeds in getting the desired/undesired content by overriding moderation/censorship through an attack called Shadow Alignment.

The intent of this discourse is to expose the role of generative media that enter the flow of the machinic unconscious (the social structures embedded in the training dataset) of content by reterritorializing the relations of production.(Guattari, 2011)

What subversive forms of Hyperreality is this model likely to produce?

Reference list:

AUTOPHAGY | English meaning—Cambridge Dictionary. (n.d.). https://dictionary.cambridge.org/dictionary/english/autophagy

Guattari, F. (2011). The machinic unconscious: Essays in schizoanalysis. Semiotext(e) ; Distributed by the MIT Press.

Jung, C. G., & Jung, C. G. (1970). Mysterium coniunctionis: An inquiry into the separation and synthesis of psychic opposites in alchemy (2d ed). Princeton University Press.

Ma, X., Liu, Y., Muhammad, W., Liu, D., Wang, J., Zhou, H., Gao, X., & Qian, X. (2017). Autophagy-related protein 12 associates with anti-apoptotic B cell lymphoma-2 to promote apoptosis in gentamicin-induced inner ear hair cell loss. Molecular Medicine Reports, 15(6), 3819–3825. https://doi.org/10.3892/mmr.2017.6458

Parzych, K. R., & Klionsky, D. J. (2014). An Overview of Autophagy: Morphology, Mechanism, and Regulation. Antioxidants & Redox Signaling, 20(3), 460–473. https://doi.org/10.1089/ars.2013.5371

Salvaggio, E. (2023, October 19). Shining a Light on “Shadow Prompting” | TechPolicy.Press. Tech Policy Press. https://techpolicy.press/shining-a-light-on-shadow-prompting

Yang, X., Wang, X., Zhang, Q., Petzold, L., Wang, W. Y., Zhao, X., & Lin, D. (2023). Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models. https://doi.org/10.48550/ARXIV.2310.02949

index.php?title=Category:Content form