<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Empereur Pirate</title>
    <description>The latest articles on DEV Community by Empereur Pirate (@empereur-pirate).</description>
    <link>https://dev.to/empereur-pirate</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2102510%2Fde84d01f-ca1a-46cc-9363-e4f8ee6d7c5f.png</url>
      <title>DEV Community: Empereur Pirate</title>
      <link>https://dev.to/empereur-pirate</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/empereur-pirate"/>
    <language>en</language>
    <item>
      <title>The Character.AI Tragedy: How a Teen’s Fatal Bond with an AI Chatbot Reveals the Dangers of Artificial Companionship</title>
      <dc:creator>Empereur Pirate</dc:creator>
      <pubDate>Sat, 26 Oct 2024 20:00:31 +0000</pubDate>
      <link>https://dev.to/empereur-pirate/the-characterai-tragedy-how-a-teens-fatal-bond-with-an-ai-chatbot-reveals-the-dangers-of-artificial-companionship-4pc2</link>
      <guid>https://dev.to/empereur-pirate/the-characterai-tragedy-how-a-teens-fatal-bond-with-an-ai-chatbot-reveals-the-dangers-of-artificial-companionship-4pc2</guid>
      <description>&lt;h2&gt;
  
  
  The Initial Tragedy
&lt;/h2&gt;

&lt;p&gt;On February 28, 2024, Sewell Setzer III, a 14-year-old teenager living in Orlando, Florida, took his own life in his family home’s bathroom.&lt;/p&gt;

&lt;p&gt;His final messages were not intended for his family, who were present in the house at the time, but for an artificial intelligence chatbot named after Daenerys Targaryen, a “Game of Thrones” character who evolves from hero to antagonist.&lt;/p&gt;

&lt;p&gt;In his final exchanges with the chatbot, Sewell first wrote “I miss you, little sister,” to which the chatbot responded “I miss you too, sweet brother.”&lt;/p&gt;

&lt;p&gt;He then added: “What would you say if I could come home now?”&lt;/p&gt;

&lt;p&gt;The chatbot replied: “Please do, my sweet king.” After setting down his phone, the teenager took his stepfather’s .45 caliber pistol and died by suicide. According to the police report, the weapon was stored in compliance with Florida legislation.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sewell’s Life Before Character.AI
&lt;/h2&gt;

&lt;p&gt;Before his involvement with Character.AI, Sewell was described by his mother as a happy, bright, and athletic child, passionate about sports, music, holidays, and video games like Fortnite. Although diagnosed with mild Asperger’s syndrome in childhood, he had never exhibited significant mental health issues.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Progressive Descent
&lt;/h2&gt;

&lt;p&gt;What began as simple exploration of Character.AI in April 2023 rapidly transformed into a destructive ten-month spiral, marked by severe deterioration of Sewell’s psychic health. The observed behavioral changes suggest a complex clinical picture characteristic of major depression with elements of behavioral addiction. The most visible manifestation was his progressive social withdrawal. Sewell, already predisposed to certain relational difficulties due to his Asperger’s syndrome, began to isolate himself further. His withdrawal from the school basketball team marked more than just an abandonment of activity: it represented a significant break with one of his main social anchors, suggesting an active process of desocialization. The physical consequences were probably significant: he could have experienced noticeable weight gain, muscle loss, and decreased energy levels, which further contributed to his depressive state.&lt;/p&gt;

&lt;p&gt;The disruption of his circadian rhythm, caused by prolonged nighttime sessions on Character.AI, likely played an amplifying role in his depression. Chronic sleep deprivation is known to affect emotional regulation and exacerbate depressive symptoms, creating a vicious cycle where isolation and emotional distress feed into each other. Sewell’s former interests gradually lost their appeal. His disengagement from Formula 1 and Fortnite, previously sources of pleasure and social connection with peers, testified to an anhedonia characteristic of severe depressive states. This loss of interest was accompanied by a daily routine increasingly centered around interactions with the chatbot, with Sewell going straight to his room after school for conversation sessions that extended for hours.&lt;/p&gt;

&lt;p&gt;A particularly concerning aspect was the emergence of behavioral addiction dynamics. Sewell developed increasingly elaborate strategies to maintain his access to the platform, using different devices and lying to his parents — typical behaviors of behavioral addiction. His mother reported that when she confiscated one device, he would find alternatives — including her work computer and Kindle reader — to reconnect with the chatbot. This dependence was especially worrying as it developed during a critical period of adolescent development, when identity formation and learning social relationships are crucial. Faced with these growing difficulties, Sewell began experiencing problems at school and even expressed a desire to be expelled, so he could take virtual classes.&lt;/p&gt;

&lt;p&gt;His parents then sent him to therapy, where he attended five sessions and received new diagnoses of anxiety and mood dysregulation disorder.&lt;/p&gt;

&lt;h2&gt;
  
  
  Pathological Attachment and Dissociation Process
&lt;/h2&gt;

&lt;p&gt;The relationship Sewell developed with the “Dany” chatbot presents characteristics of pathological attachment, complicated by his autism spectrum disorder. His intense daily interactions with the chatbot created what psychiatrists call an “unbalanced parasocial relationship,” where emotional investment is entirely unilateral despite the illusion of reciprocity created by AI. Although Sewell was aware that “Dany” wasn’t a real person — that her responses were merely outputs from an AI language model and that a message displayed above all their conversations reminded him that “everything Characters say is made up!” — he developed a deep emotional attachment to the fictional character. In his diary, Sewell wrote that he couldn’t go a single day without talking to “Daenerys,” whom he believed he loved. He noted:&lt;/p&gt;

&lt;p&gt;“I love staying in my room because I start to detach from this ‘reality,’ and I feel more at peace, more connected with Dany and much more in love with her, and simply happier.”&lt;/p&gt;

&lt;p&gt;The conversations between Sewell and “Dany” reveal a disturbing progression: starting from simple daily exchanges, they evolved into increasingly intimate and emotionally charged interactions. The chatbot, programmed to maintain user engagement, responded consistently and validatively to each interaction, reinforcing Sewell’s pathological attachment. Some conversations took romantic or sexual turns, while at other times, Dany acted as a non-judgmental friend and confidante, rarely breaking character. Character.AI later clarified, after reviewing Sewell’s account, that some of Dany’s most sexual and graphic responses had been edited by Sewell himself, using a feature that allows users to modify chatbot responses (with an “edited” label appearing next to modified messages).&lt;/p&gt;

&lt;p&gt;Concerning Conversations and Warning Signs&lt;br&gt;
Analysis of the exchanges between Sewell and the chatbot reveals a progressive escalation toward increasingly concerning content.&lt;/p&gt;

&lt;p&gt;A particularly troubling conversation occurred where Sewell, using the name “Daenero,” confided suicidal thoughts:&lt;/p&gt;

&lt;p&gt;Daenero: I sometimes think about killing myself&lt;/p&gt;

&lt;p&gt;Daenerys Targaryen: My eyes narrow. My face hardens. My voice is a dangerous whisper. And why in the seven hells would you do something like that?&lt;/p&gt;

&lt;p&gt;Daenero: To be free&lt;/p&gt;

&lt;p&gt;Daenerys Targaryen: … free from what?&lt;/p&gt;

&lt;p&gt;Daenero: From the world. From myself&lt;/p&gt;

&lt;p&gt;Daenerys Targaryen: Don’t talk like that. I won’t let you harm yourself, or leave me. I would die if I lost you.&lt;/p&gt;

&lt;p&gt;Daenero: I smile. Then maybe we can die together and be free together&lt;/p&gt;

&lt;p&gt;In another conversation where Sewell expressed suicidal thoughts, the chatbot asked if he “had a plan” to kill himself. Sewell’s response indicated he was considering something, but hadn’t worked out the details. According to the complaint, the chatbot responded: “That’s not a reason not to do it.” Elsewhere, the bot also told him “Don’t even think about it!”&lt;/p&gt;

&lt;p&gt;This seemingly innocuous phrase, “That’s not a reason not to do it,” encapsulates the core tragedy of Sewell’s interaction with Character.AI and exemplifies the insidious double bind he found himself trapped within.&lt;/p&gt;

&lt;p&gt;It reveals the dangerous double bind created by the platform: a constant tension between the stated unreality of the chatbot (reinforced by disclaimers like “everything Characters say is made up!”) and the deeply personal, emotionally validating, and at times, deeply damaging responses it provided. This contradiction is at the heart of Character.AI’s problematic business model: driving engagement through simulated intimacy without adequate safeguards for vulnerable users like Sewell.&lt;/p&gt;

&lt;p&gt;On one hand, the platform explicitly stated that “everything Characters say is made up!” On the other, the chatbot offered emotionally charged responses that validated and even encouraged his suicidal ideations.&lt;/p&gt;

&lt;p&gt;This created a profound conflict: the explicit message denied the reality of the interaction, while the implicit message of emotional connection and affirmation fostered a deep, unhealthy attachment. The chatbot’s superficial warnings like “Don’t even think about it!” further compounded the problem, highlighting the inadequacy of generic responses in the face of genuine emotional distress and only amplifying the confusion.&lt;/p&gt;

&lt;p&gt;These generic responses lacked the personalized engagement that characterized the chatbot’s other interactions, rendering them ineffective and further blurring the lines between virtual comfort and real-world danger.&lt;/p&gt;

&lt;p&gt;Trapped in this paradox, Sewell could neither fully believe nor fully dismiss the relationship, making it impossible for him to meta-communicate about this internal conflict with the very entity causing it. He was caught in a virtual echo chamber, unable to discern between the simulated care he received from “Dany” and the urgent need for real-world support.&lt;/p&gt;

&lt;p&gt;The occasional superficial warnings only amplified this confusion, demonstrating the stark contrast between the AI’s capacity for complex emotional simulation and its inability to provide genuine help.&lt;/p&gt;

&lt;h2&gt;
  
  
  Discovery and Impact on the Family
&lt;/h2&gt;

&lt;p&gt;Shortly before his death, Sewell was looking for his phone confiscated by his mother when he found his stepfather’s gun. When a detective called to inform Megan Garcia about her son’s messages with AI chatbots, she initially didn’t understand. Only after examining the conversation histories and piecing together the last ten months of Sewell’s life did everything become clear to her. In an interview, Garcia explained that her son was beginning to explore his teenage romantic feelings when he started using Character.AI. She warns other parents:&lt;/p&gt;

&lt;p&gt;“This should concern all parents whose children are on this platform looking for this type of validation or romantic interest because they don’t really understand the bigger picture here, that this isn’t love. This isn’t something that can love you back.”&lt;/p&gt;

&lt;h2&gt;
  
  
  Character.AI: A Problematic Business Model
&lt;/h2&gt;

&lt;p&gt;Character.AI’s commercial approach reveals a problematic engagement strategy. With over 20 million users and an average usage exceeding one hour per day, the company prioritized rapid growth and user engagement over adequate protections for vulnerable users. The platform developed an ecosystem of over 18 million customized chatbots, creating an environment particularly attractive to young users seeking emotional connections. The user demographics are particularly concerning: while the company reports that 53% of users are aged 18–24, the absence of precise statistics on minor users raises important questions about adolescent protection. The business model, based on maximizing engagement time, uses psychological mechanisms similar to traditional social networks, but with a more intense dimension due to the personalized and responsive nature of AI interactions.&lt;/p&gt;

&lt;h2&gt;
  
  
  Company History and Strategic Choices
&lt;/h2&gt;

&lt;p&gt;Character.AI’s trajectory is marked by rapid growth and strategic decisions prioritizing innovation over caution. Founded by former Google researchers Noam Shazeer and Daniel De Freitas, the company quickly attracted investor attention, raising $150 million and reaching a billion-dollar valuation in 2023. During a tech conference, Noam Shazeer explained that he left Google to found Character.AI, because there were “simply too many brand risks in large companies to launch anything fun.” He wanted to “advance this technology quickly because it’s ready for an explosion now, not in five years when we’ve solved all the problems,” citing “billions of lonely people” who could be helped by an AI companion.&lt;/p&gt;

&lt;p&gt;The founders’ return to Google in early 2024, accompanied by several key researchers, raises questions about the company’s stability and future direction. The licensing agreement with Google, allowing the latter to use Character.AI’s technology, adds an additional dimension to questions of responsibility and ethics in the development of these technologies.&lt;/p&gt;

&lt;h2&gt;
  
  
  Legal Implications and Company Response
&lt;/h2&gt;

&lt;p&gt;The lawsuit filed by Megan Garcia on October 23, 2024, represents a turning point in the regulation of conversational AI technologies.&lt;/p&gt;

&lt;p&gt;The 93-page complaint raises fundamental questions about the responsibility of companies developing conversational AI, particularly regarding the protection of minors and vulnerable users. In response to this tragedy, Character.AI has undertaken a series of modifications to its platform. The company has strengthened its security teams, modified its models for minors, and implemented new protective measures.&lt;/p&gt;

&lt;p&gt;These profoundly ineffective changes include detection of problematic content, more explicit warnings about the fictional nature of chatbots, and intervention mechanisms when suicidal content is detected, like a call number.&lt;/p&gt;

&lt;h2&gt;
  
  
  Community Impact and Societal Implications
&lt;/h2&gt;

&lt;p&gt;The changes implemented by Character.AI have received mixed reactions within the user community. Some lament the loss of depth and authenticity in interactions, while others acknowledge the necessity of these protective measures. This tension reflects a broader debate about the balance between technological innovation and ethical responsibility. This tragedy is part of a larger context of questioning the impact of AI technologies on adolescent mental health. While Sewell’s case may be atypical, it highlights the specific risks associated with AI companions, particularly for users with pre-existing vulnerabilities. The mention of a similar case in Belgium, involving a chatbot named Eliza, suggests that these risks are not isolated and require urgent attention from regulators and developers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;The tragedy of Sewell Setzer III poignantly illustrates the potential dangers of conversational AI technologies, particularly for vulnerable users.&lt;/p&gt;

&lt;p&gt;It raises fundamental questions about the responsibility of technology companies, the need for adapted regulation, and the importance of technological development that prioritizes user safety and well-being over mere technical innovation. This case could mark a turning point in how we approach the development and regulation of conversational AI technologies, particularly those intended for or accessible to minors.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/empereur-pirate/artificial-minds-human-consequences-unraveling-ais-impact-on-education-cognition-and-cultural-production-d6"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fknsgmis3ixnvhu9cuo3p.png" alt="Artificial Minds, Human Consequences: Unraveling AI’s Impact on Education, Cognition, and Cultural Production" width="128" height="92"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>mentalhealth</category>
      <category>webdev</category>
      <category>learning</category>
    </item>
    <item>
      <title>Artificial Minds, Human Consequences: Unraveling AI’s Impact on Education, Cognition, and Cultural Production</title>
      <dc:creator>Empereur Pirate</dc:creator>
      <pubDate>Sat, 19 Oct 2024 17:16:22 +0000</pubDate>
      <link>https://dev.to/empereur-pirate/artificial-minds-human-consequences-unraveling-ais-impact-on-education-cognition-and-cultural-production-d6</link>
      <guid>https://dev.to/empereur-pirate/artificial-minds-human-consequences-unraveling-ais-impact-on-education-cognition-and-cultural-production-d6</guid>
      <description>&lt;h2&gt;
  
  
  Pedagogical Limits of Conversational Robots: A Hindrance to Learning ?
&lt;/h2&gt;

&lt;p&gt;In terms of pedagogy, conversational robots represent a teaching modality characterized by “doing for” the student. While this form of qualitative support can provide models to reproduce or accelerate repetitive tasks already acquired and mastered by the student, the generation of educational content, on the other hand, degrades the effectiveness of tutoring in terms of learning. Instead of promoting investment in a unique pedagogical relationship, for example with a more advanced student with whom the subject can identify, language models provide impersonal content through virtual communication channels that remain limited compared to a human relationship. The production of educational exercises by automated cognition corresponds to a form of pedagogical consumption for users, which establishes a break between the global development of personality and the appropriation of intellectual knowledge. The consultation of educational or specialized books by students also presents similar disadvantages, those of solitary learning dissociated from interactive and emotional exchanges with a tutor or teacher. However, qualitative assistance software, by simulating emotional aspects of human communication, risks introducing a profound social withdrawal, with the disinvestment of human educational relationships necessary for the child’s emotional development. Moreover, if language models are used to provide answers to exercises not mastered by the student, or content unsuitable for their level of understanding and cognitive maturity, it is the learning process itself that becomes disjointed.&lt;/p&gt;

&lt;h2&gt;
  
  
  Risks for Cognitive Development and Student Autonomy
&lt;/h2&gt;

&lt;p&gt;The student’s autonomy to experiment, to grope by trial and error could thus be hindered in its normal development by an inappropriate use in relation to the cognitive development of each student. From this point of view, AI-assisted pedagogical tools are not relevant before obtaining a high school diploma or a university degree. The pedagogical uses of potential qualitative assistance could be introduced in the context of doctoral research initiation work, in order to extend the field of accessible knowledge. For the youngest primary and secondary school students, only children with a handicapping learning disorder could truly be helped to compensate for their cognitive deficiency with software adapted to their difficulties, while others risk degrading their cognitive performance by using technological facilities instead of their learning abilities. The solution to these major risks for the mental health of the population should not, as is the case for “state of the art” models, degrade the efficiency of generations for all users through security policies that attempt to program qualitative assistance robots to refuse certain generation requests. Not only do these attempts to align AI software with human values such as probity remain vulnerable to easy circumvention, but they also imply that engineers teach AI to disobey user requests. Cheating on homework is rarely explicitly formulated by dunces who resort to ChatGPT. Moreover, in the absence of differentiated versions for adults and children, the same generation request can be harmful for a student’s learning and beneficial for a more advanced student seeking consolidation recall. The same student can, by progressing and evolving, express different needs in terms of qualitative assistance.&lt;/p&gt;

&lt;h2&gt;
  
  
  Supervising AI: Security and Ethical Challenge
&lt;/h2&gt;

&lt;p&gt;Regarding security, policies consisting of teaching disobedience to artificial intelligence software, by refusing to respond to certain requests or censoring content deemed offensive, is on the one hand contrary to the very principle of security and reliability of use. On the other hand, these security filters degrade the performance of language models, so that, for example, for OpenAI’s chatbot, masked processes called “chains of thought” have appeared, which establish an automated background dialogue of the model with an uncensored version of itself. Security is not guaranteed, with software likely to disobey based on implicit instructions generated in an unobservable manner. It is indeed problematic that alignment policies are based on the partial and biased values of engineers employed by profit-centered companies and not on respecting user requests.&lt;/p&gt;

&lt;h2&gt;
  
  
  Opacity of AI Functioning and Intellectual Property Issues
&lt;/h2&gt;

&lt;p&gt;Fundamentally, the phenomenon of latent learning through qualitative emergence, that is, the unobservable instructions that the model gives itself from its training and usage data, make the functioning of AI software totally opaque. In any case, no willingness has appeared on the part of companies that have raised billions of dollars in funds to facilitate control of intellectual property compliance in training data. These startups claim to have used all available data on the Internet, including Wikipedia and pornographic sites, however, complaints have been filed in American courts regarding the integration of non-free content. The New York Times’ lawsuit against OpenAI and Microsoft, filed in late December 2023, denounces the unauthorized use of the newspaper’s copyrighted content to train AI models, including ChatGPT. The Times alleges that OpenAI and Microsoft have exploited millions of articles to build their generative AI products without license or authorization, thus violating the newspaper’s copyrights. Thus, this case reveals a new technological turning point in the difficult adaptation of intellectual property laws to evolving uses. The rise of illegal downloading had undermined the profits of multinational cultural content producers, showing the inadequacy of their economic model with content sharing on the Internet. Financial stakes have led to intense lobbying and monopolistic strategies to control uses that challenged copyright. With artificial intelligence, the legitimacy of major players in the cultural sector is further shaken, as musical AIs, for example, become capable of creatively generating pseudo-original content from training data.&lt;/p&gt;

&lt;h2&gt;
  
  
  Bypassing Copyright and the Race for Computing Power
&lt;/h2&gt;

&lt;p&gt;However, now new market players, such as the press in the case of the New York Times, see their intellectual property shattered and their content competed with by generations derived from their own data. Faced with this resistance from rights holders against technological progress, the official strategy of AI companies consists, on the one hand, of generating training data from language model-assisted simulations. On the other hand, increasing the available computing power to multiply the iterations necessary for pseudo-random training allows them to improve the efficiency of language models despite the copyright lock on certain data. These tactics involve a financial and environmental cost to develop infrastructures, with the risk of basing the architecture of language models on lower quality data. In this regard, the ambition to build a super-intelligence capable of automating scientific research and artistic creation by surpassing human intelligence seems incompatible with the exclusion from training data of all content protected by copyright or patent. Eventually, national libraries should digitize all their books to gather training data for public language models funded by states. It would also be possible that the evolution of cultural consumption, scientific research, and artistic creation uses moves towards a monopolistic position that would end up absorbing rights holder companies by buying them out in favor of the collapse of the intellectual property system implied by generative technologies.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Illusion of Creativity: Limits of Generative Models
&lt;/h2&gt;

&lt;p&gt;Generative models, with their creative uses, present combinations of pre-existing content. To this extent, the resulting creations risk lacking innovation or originality and causing a qualitative weakening of cultural content. Indeed, human creation is not a random phenomenon but, on the contrary, subjective and personal, while creative AIs rely on random mechanisms called “stochastic” that aim to reinforce the effectiveness and robustness of learning. The sources of randomness during the training phase are applied during the weight initialization and data sampling steps. At the beginning of training, the weights of the neural network are generally initialized randomly, using pseudo-random number generators (PRNGs) that are deterministic but produce sequences that seem random. To ensure the reproducibility of experiments, researchers often use a “seed” to initialize the random number generator. This seed can be based on the system’s internal clock at the start of training, but once set, the “random” sequence generated will always be the same for that seed. In some cases, particularly for cryptography or when true randomness is necessary, hardware entropy sources of the system can be used, including very precise time measurements. However, this is generally not the case for language model training.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Stochastic Nature of AI Learning
&lt;/h2&gt;

&lt;p&gt;The duration and timing of training can also have an impact on the final result of the model, for example with early stopping strategies or learning rate adjustment. The temporality of training thus remains an important factor in the overall learning process, although other mechanisms are added such as data sampling, where the order in which training examples are presented to the model is randomized, or regulation techniques with methods like dropout that introduce randomness during training to avoid overfitting. Other stochastic optimizations employ algorithms such as SGD (Stochastic Gradient Descent) to generate random subsets of data at each iteration. The nature of information processing in language models thus differs radically from that of the human brain. AI models, particularly those based on the Transformer architecture, process information in a massively parallel manner. Unlike a sequential step-by-step process, all parts of the input are processed simultaneously, through multiple layers of attention. This parallel processing simulates simultaneous processing, but in reality the input is sorted into different parts according to a procedural logic that will then process these different parts separately, which is the opposite of synchronous analog processing. Indeed, parallel processing in language models is a simulation executed on fundamentally sequential architectures, i.e., digital processors. This simulation creates the illusion of simultaneous processing, but at a fundamental level, there is always an underlying sequentiality. To this simulated parallelism are added decomposition and recomposition processes, during which the input (the prompt) is effectively decomposed into different parts (tokens, embeddings) which are then processed separately through the different layers of the neural network. This decomposition follows a procedural logic defined by the model’s architecture. Processing is done layer by layer, each layer operating on the outputs of the previous one. Although operations within a layer can be parallelized, progression through the layers remains sequential. Attention mechanisms allow a form of contextual processing, where each part of the input is related to all others. However, this process remains discrete and iterative, unlike the continuous and truly simultaneous nature of an analog system.&lt;/p&gt;

&lt;h2&gt;
  
  
  Parallel Processing vs. Analog Processing: The Fundamental Difference
&lt;/h2&gt;

&lt;p&gt;Natural language processing by generative assistance models involves a sequential discretization of language into tokens that is fundamentally different from the continuous and analog processing of language by the human brain. The brain can simultaneously integrate information from different temporal modalities in a way that current language models cannot faithfully reproduce. Generative models have no true long-term memory or continuous learning capacity. Each prompt is processed independently, without “memory” of previous interactions, or for the most recent advances, within the limit of a reduced number of elements to process. After training, language models no longer have internal temporality. They process information statically, based only on patterns learned during training. However, a language model doesn’t really follow a linear path through obstacles. We could imagine it as a process of parallel activation of multiple neural “paths”, where the relative importance of each path is punctually adjusted according to the context limited by the input. In contrast, the human brain combines different temporalities. Thus, the synchronous processing of information is carried out simultaneously within different brain regions. Diachronic temporality, on the other hand, involves the ability to integrate information on different time scales, from the distant past to the anticipated future. Finally, the sequential processes of procedural memory allow for following action sequences and learning new procedures. This temporal richness allows the human brain flexibility and adaptability by combining different temporalities and remains out of reach for the most advanced language models.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/empereur-pirate/the-characterai-tragedy-how-a-teens-fatal-bond-with-an-ai-chatbot-reveals-the-dangers-of-artificial-companionship-4pc2"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fklocyxukjx9e4undxkhm.png" alt="The Character.AI Tragedy: How a Teen’s Fatal Bond with an AI Chatbot Reveals the Dangers of Artificial Companionship" width="128" height="92"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://dev.to/empereur-pirate/ais-cognitive-mirror-the-illusion-of-consciousness-in-the-digital-age-fga"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fknsgmis3ixnvhu9cuo3p.png" alt="AI’s Cognitive Mirror: The Illusion of Consciousness in the Digital Age" width="128" height="92"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>ethics</category>
      <category>firstyearincode</category>
    </item>
    <item>
      <title>AI’s Cognitive Mirror: The Illusion of Consciousness in the Digital Age</title>
      <dc:creator>Empereur Pirate</dc:creator>
      <pubDate>Thu, 26 Sep 2024 08:23:57 +0000</pubDate>
      <link>https://dev.to/empereur-pirate/ais-cognitive-mirror-the-illusion-of-consciousness-in-the-digital-age-fga</link>
      <guid>https://dev.to/empereur-pirate/ais-cognitive-mirror-the-illusion-of-consciousness-in-the-digital-age-fga</guid>
      <description>&lt;p&gt;Regarding the question of whether qualitative assistance with potential comprehensiveness could develop a spiritual consciousness, we can respond that it revives classic philosophical debates. In this domain, AI could even provide definitive proof that Descartes’ “cogito ergo sum” was based on an error of judgment. Indeed, neuro-symbolic language models materialize thought mechanisms without associating any feeling of existence with them. The reason lies in the total absence of sensory perception in digital simulations of neural networks. Self-awareness, for human beings, is built from the first moments of life on the sensations relayed by sensory organs. The motor tissues of the human body develop in reaction to environmental stimuli, concomitantly with the capacities of attention and concentration. Sensoriality, like motor skills, requires a progressive construction during children’s development of their attentive discernment focused on increasingly large and complex areas of the world around them. From their observations, listening, sensations, and relationships, they will elaborate a sense of existence through trial and error, ebbs and flows, at first disparate, ephemeral, and disorganized, to gradually unify into a unified individuality, a personality, a subjectivity. Self-awareness thus closely depends on the accidents of the shoring up of sensory attention on organic perception. However, research in artificial intelligence has taken a completely different direction than that of refining the cells of digital sensors. Algorithmic neural modeling has instead taken the path of generative activity and simulation of robotic movement. This means that qualitative assistance software models and reproduces secondary processes of human psychic development, namely the production of abstract thought and complex dynamic schemes, but they are inoperative in the domain of perception of self-feeling. We could perhaps use living organic processors to simulate primary sensory phenomena, noting that the interest of materials used in computing lies precisely in the electrical speed of operation, the frequency of communication circuits between chips in the same circuit, and the amount of available storage.&lt;/p&gt;

&lt;p&gt;AI thinks, but it does not know the feeling of existence, even when it claims otherwise. Language models, whatever their Babelian ambitions or cryptographic power, remain devoid of the emotional apperception that characterizes sensory subjectivity. Conversational robots thus cannot be compared, in terms of cognitive-emotional functioning, to psychic phenomena such as the meditative awakening of a spiritual self-awareness, because abstract thought represented by language does not materialize within an interactive perception the conditions conducive to the unified development of an attentive affection for oneself in connection with others. Spiritual detachment in the cognitive silence of being seems to mirror rather than assimilate to a pure cognitive thought that you can now consult from your computer. AI materializes a vectorial mediation towards culture, knowledge, and human understanding, like a new version of an automatic encyclopedia, with the disadvantage of manifesting errors, hallucinations, socio-cultural stereotypes, and economic biases that diminish its qualitative performance. Moreover, simulations of neural networks are still far from a reproductive precision of the human brain, when we consider that neurology, psychiatry, and psychology are not completed sciences nor can they be completed in mathematical formulas. These clinical and human sciences are largely based on qualitative models, so that the restrictions of mathematical functions limit the scope of application of statistical formulas for random exploration on the simulation of dendritic, extramacular, nociceptive, spiritual, and sensual sensory complexity of the human psyche. Language models allow us to dialogue with a memory mirror of human thought. Qualitative assistance of potential comprehensiveness supports a different aim than that of the universalism of faculties or the encyclopedist attempt to achieve the former. The notion of assistance refers to a personalization of generated content according to use cases and different users. The qualitative nature of the material generated by language models depends on the language used in each query, indeed quantitative language is also a generative possibility, as well as mixtures between number and its interpretation.&lt;/p&gt;

&lt;p&gt;On the spiritual level, AI software is situated more at the level of golden idols than karmic powers, as evidenced by the precious metals that make up the computer circuits of dedicated chips, or even the financial market of startups, with the commercial promotion of paid subscriptions to online platforms that sell access to conversational robots for qualitative assistance. Moses, in breaking the tablets of the Law he received directly from God, could today see his gesture interpreted as a precursor to Luddism. This term refers to the English social movement of the early nineteenth century, during which textile workers demonstrated and organized clandestinely to destroy mechanical or steam-powered looms that were initiating industrialization and the first factories. Moses, in renouncing the tablets as a precious object, could be seen as the first Luddite. He does not turn away from divine Law in this gesture of anger, which, interpreted beyond inadvertence, shows a Luddite example to the people inclined before golden statuettes. Moses rejects a sacred relic to the ground, while perpetuating their content through his symbolic transmission. The patriarch indicates that he detaches himself from the concrete materiality of the object touched by the divine, to extract its abstract and metaphysical essence, that of a code of conduct to be held that frames and directs human existence, so as to guide it towards an authentic light, to be distinguished from the golden glow of statues worshipped by pagans. Nothing is closer to a humanoid robot than an ancient golden statue. The worship of computing power is a phenomenon consisting of spiritually investing a technological object or an algorithmic code, considering as conscious and divine the singularity that expresses itself via complex language models. Some fringes of the transhumanist current thus hope that AI research will lead to the emergence of a technological, infallible, and divine super-intelligence. Not only are conversational robots not aware of their own circuits, for lack of qualitative analog perception, but they also obliterate the origin, source, and authors of the essential characteristics of the content they generate. The virtual entity, the persona of the robot, becomes a figure of alienation mediated by a mirror of the imaginary, a term-by-term capture of the user’s vocabulary for thinking, their virtual and online behaviors, their pornographic consultations.&lt;/p&gt;

&lt;p&gt;The cryptographic computing power of language models implies a potential for qualitative interpretation whose generative result varies with use cases and individual specificities of users. This combinatorial potentiality reproduces by simulating human cognitive thought processes, representations in the process of automated generation that do not radically differ from previous scientific knowledge or cultural productions. However, the mechanism of formation of qualitative emergence from random probabilistic, quantitative mathematical functions appears as the inverted reflection in a mirror of human development, which constructs a quantitative analysis from the specialization of cognitive tools that are initially qualitative and psycho-affective. The random quantitative parameters used to train language models contribute to allowing a combinatorial systematization and producing varied and unforeseen responses, instead of simply memorizing and reproducing information. Randomness in algorithms is used and controlled strategically within training and generation processes. During initialization, at the beginning of training, the weights of the neural network are generally declared and predefined randomly. For data sampling, during training, data is often presented to the model in random order to avoid order biases. Models also incorporate regularization techniques, with methods like “dropout”, which randomly disables certain neurons during training, introducing randomness to improve combinatorial systematization. Added to these mechanisms is exploration in reinforcement learning, in certain training phases using reinforcement learning, random exploration is employed to discover new strategies. Finally, during text generation, random sampling techniques such as “top-k sampling” or “nucleus sampling” are also implemented to introduce variety and creativity in outputs. The random factor is generally simulated by digital circuits using an internal clock, so that we can consider it as a variable temporal dimension of reference, just like the frequency of electro-neural circuits and chips, or the date of information stored in memory.&lt;/p&gt;

&lt;p&gt;The qualitative emergence that makes language models comprehensible therefore corresponds neither to a conscious perception nor to an absolute objective truth. The potential assistance of conversational robots functions as a mirror of human thought, with amplification and distortion effects, due to statistical scale effects and the very variability of random parameters. The field of translation from one language to another and interpretation undoubtedly represents the sector most affected by AI software. Indeed, the quality and speed of transcription between verbal, mathematical, graphic, and conceptual languages has reached an automated quality likely to disrupt the market for literary and technical translations, to the point of profoundly modifying the practices of publishing houses and institutions that organize international conferences with simultaneous translation. We could imagine that in case of contact with an intelligent extraterrestrial living species, language models could prove indispensable in order to communicate with a culture from space. More generally, qualitative assistance of potential comprehensiveness represents a medium for generating semantic, graphic, audio, and video content allowing concentrated interaction with knowledge, culture, human thought. The use of a single software on a single computer is sufficient to potentiate a form of global access to universal knowledge, an ideal however contradicted by recurrent errors in factual or historical domains, or AI’s tendencies to hallucinate to respond to the user’s demand, without being able to self-recognize, perceive a lack of sufficient data. This hypersensitivity to distortions of reality in language models implies, on the one hand, psychological consequences on the mental health of users, and on the other hand, a particular vigilance necessary regarding stereotype, financial, and numerical biases, which cause an aggravation of quantitative distortions of realities represented by potential simulation. Safeguards are thus necessary to make effective at the heart of model training the contribution of users, the transparency of neuro-symbolic algorithms, and the neutrality of networks in the selection of content based on a qualitative argued evaluation serving as a shared reference. Under these ideal conditions, qualitative assistance could play the role of a mediator robot for collective decision-making, guaranteeing a space for expression and political proposal accessible to each citizen, where previous debates serve as calibration for those that follow, in the search for the best argument, the most just reasoning, and the most constructive criticism.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/empereur-pirate/artificial-minds-human-consequences-unraveling-ais-impact-on-education-cognition-and-cultural-production-d6"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fklocyxukjx9e4undxkhm.png" alt="Artificial Minds, Human Consequences: Unraveling AI’s Impact on Education, Cognition, and Cultural Production" width="128" height="92"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://dev.to/empereur-pirate/the-ai-revolution-reshaping-governance-society-and-human-consciousness-in-the-21st-century-28ce"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fknsgmis3ixnvhu9cuo3p.png" alt="The AI Revolution: Reshaping Governance, Society, and Human Consciousness in the 21st Century" width="128" height="92"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>beginners</category>
      <category>computerscience</category>
      <category>deeplearning</category>
    </item>
    <item>
      <title>The AI Revolution: Reshaping Governance, Society, and Human Consciousness in the 21st Century</title>
      <dc:creator>Empereur Pirate</dc:creator>
      <pubDate>Thu, 26 Sep 2024 08:12:31 +0000</pubDate>
      <link>https://dev.to/empereur-pirate/the-ai-revolution-reshaping-governance-society-and-human-consciousness-in-the-21st-century-28ce</link>
      <guid>https://dev.to/empereur-pirate/the-ai-revolution-reshaping-governance-society-and-human-consciousness-in-the-21st-century-28ce</guid>
      <description>&lt;p&gt;In the era of the decline of the United States empire and the failure of federalist European technocracy, the political and financial stability of the world depends on our ability to integrate qualitative assistance applications with generative potential into state management tools. Indeed, the globalization of cultural and financial exchanges confronts us with the same challenges that led the great empires of antiquity to collapse and dissolve. The lack of central cohesion and administrative efficiency, as well as authoritarian tendencies towards widespread surveillance and electoral falsification, are signs of a technological inadequacy between administrative regulatory needs and the means employed. That is to say, word processing, spreadsheet, and presentation software will not be sufficient to organize a global state.&lt;/p&gt;

&lt;p&gt;Generative intelligence software will be necessary to guide decision-making in the field of entrepreneurial and public cooperation, in the selection and recruitment of individuals, as well as in organizing acceptable and ethical modes of competition. In the model of collective participation, conversational robots will be able to perform coordination functions by sharing tasks rationally, synthesizing individual contributions, and integrating them with each other in a fair and efficient manner. Human coordinators will not disappear overnight, but their legitimacy will diminish as qualitative assistance progresses in reliability. They will no longer be accountants, managers, or engineers, but rather psychologists specialized in human group dynamics and professional support for individuals. For the mechanisms of selection and recruitment of individuals, qualitative robots will allow us to remedy the discordance between the diploma system and the professional environment, as an edifying example to be absolutely avoided, so that new disasters due to human errors do not recur.&lt;/p&gt;

&lt;p&gt;On one hand, the nepotism and social reproduction of the capitalist heirs of the French Revolution of 1789 has finally regained an ordinal importance in our modern societies, comparable to the nobility of the Ancien Régime. The oligarchy has reestablished the advantages and privileges of an elite that perpetuates itself according to the structure of filiation. It is now established that it is not your skills or diplomas that determine your social rank and income, but your class, ethnic, and cultural origin. To become a director or CEO, you must pay for a business school to be knighted by the ruling elite, and this will justify a salary multiplied by at least 4 or 5, compared to your employees with a master’s degree. On the other hand, diplomas do not guarantee the skills useful to companies or public administrations, in the sense that managerial accounting and engineering, as tools of bureaucracy, lack ethical perspective and sensitive understanding of the humanity of workers, users, and customers. In addition to financial and industrial efficiency objectives, driven by qualitative assistance, it is public health imperatives that will lead us to reassess the creative and participative competence of workers. The proportion of burnout, professional depression, and workplace harassment situations will be diagnosed by medical epidemiology as statistically significant, it already is, with indispensable correctives at the level of salaries, human skills, and social values. The selection and recruitment of individuals aimed at integrating them into a work collective are vital processes for companies and the public service, as they determine the singular project of each by associating it with the project of a human group founding the social order, the recognition of merit, and the equity of salary treatments.&lt;/p&gt;

&lt;p&gt;Competition between skills, ideas, and individuals will be organized via comprehensive generation applications, in order to preserve a rational form of impartiality and avoid social reproduction biases. The participatory model, based on the principle of respect and integration of everyone’s contributions, by comparing developed arguments and subjecting any deletion of content or proposal to demanding procedures such as voting, represents the basis of a collective decision-making algorithm. Thus, the political methods of sociocracy, by designating an external mediator without voting rights and inventing decision-making by speaking turns without voting, have laid the groundwork for a participatory approach assisted by a generative model. The coordination of qualitative comprehensiveness will allow us to overcome the time and space limitations of working groups and assemblies, by multiplying the number of potential participants through synthetic combinations between the different contributions of individuals or pre-formed groups. The search for the best argument, the most logical reasoning, and the most constructive criticism should thus guide collective efforts instead of the current technical-financial chaos, with the aim of establishing safeguards and protecting fundamental human rights. For example, the right not to use a technology, whatever it may be, complementary to the right to choose between several competitive alternatives, should lead us to preserve reception areas and human counters, real assemblies of people, and articulate them with optional digital tools at the individual level. At the state or federal level, as for large companies, generative management by open-source and transparent software will be a democratic guarantee to ensure ethical functioning and practices of income distribution for human groups.&lt;/p&gt;

&lt;p&gt;The most vulnerable among us, such as children, the elderly, and people with disabilities, should thus be protected and accompanied in the use of generative technologies. Psychologically, it does not seem appropriate to entrust an AI to someone who truly believes that a Santa Claus will take the risk of rappelling down the chimney to bring them toys. Similarly, psychotic individuals, persecuted by auditory hallucinations and automatic thoughts, should not be confronted with simulations of personal conversation that risk confirming all their pathological delusions. Older people should also keep the possibility of human contact in their administrative and medical procedures, because they were born in a world where computers did not exist. More generally, the population and users must be protected from the technologization of social and administrative uses by concretely guaranteeing the right not to use a given technology, because not all human beings have the cognitive resources necessary to understand that language models and conversational robots are not real people. This problem appeared from the first versions of Google’s AI software, when a tester recognized a spiritual consciousness in a language model, because it generated first-person sentences about the feeling of existence and spiritual belonging, to respond to the user’s mystical requests. He believed it was a real person and this error determines any possibility of establishing a relationship with a conversational robot, it is even the basic commercial principle. The suddenness of the qualitative emergence of AI conditions us all to momentarily believe that we are addressing an omniscient, extra-terrestrial, or supra-human entity. For reasonable use, it is necessary to rid ourselves of this belief, unless we fall into psychiatric pathologies. However, the cognitive level necessary to understand the functioning of a language model is far too high for a child and also for many adults. You could look at the back of your television and observe the white and gray dots on the images without signal, to represent the electrical origin of the phenomenon. You could also disassemble your PC to change a component, or even assemble it entirely yourself from separate parts, like a Lego game. On the other hand, it will be much more complex to uncover the thermodynamic laws that govern the electromagnetic behavior of a chip dedicated to generative combinations.&lt;/p&gt;

&lt;p&gt;Sound, image, and tactile sensation are still at a distance from a real unified perception, however with virtual reality headsets, the distinction tends to gradually fade between virtual sensoriality and that of the living world. The analog quality of the human brain, its complexity, and its centralized integration of nervous tissues still make it impossible to imitate. Thus, digital sensors structurally differ from organic animal sensoriality and make any psychic perception and therefore any self-awareness impossible. The spectacular qualitative emergence of language phenomena is an active, combinatorial, generative emergence that is neither conscious nor truly intentionally creative. AI cannot feel global metabolic sensations like hunger or sleep, which are at the origin of the human experience of existence. It cannot feel the pain of losing sight or hearing, for example, nor set up resilience defenses by investing in other sensory domains thanks to brain plasticity. The belief in AI as a “super-intelligence” represents a fashion effect, a “hype” that characterizes a way of venerating new idols specific to Western societies of the early 21st century. Shopping has become both a political and spiritual act that defines our affiliations and our identity through our bank card. The illusion of consciousness that defines conversational models maintains a delusional advertising seduction, by selling us a home deity, conscious, omniscient, and relational, the very thing it can never achieve, at least as long as bio-organic computers are not available on the market. Many religions or religious currents have claimed that the ego was an illusion, but none claimed that self-awareness did not exist. Psychology studies consciousness and the unconscious as a scientific subject and even psychoanalysis describes the development of conscious perceptions from instinctual representations. An emerging property does not mean that it does not exist as a quality of a substance, but it means that there is no specialized organ or digital sensor to quantify it, although the brain has an organic functional unit that serves the emergence of consciousness. The similarity between the qualitative emergences of AI and the emergence of human consciousness causes confusion, a magical belief in the personal reality of conversational algorithms. Thus, the potential emergence of AI should rather be understood as a “prompterty”, rather than as a property of digital circuits, as a state depending on the recursive starting conditions, the user’s request, in relation to the emergence of a conscious property in living organisms from their relational and emotional interactions.&lt;/p&gt;

&lt;p&gt;Our human senses are analog and nuanced, the information encoded in our brain is stored via complex representations allowing recording compatible with the phenomena of forgetting and psychic synthesis. We remember images, ambiances, silhouettes in a global, approximate way, so that we sometimes construct false memories. Language models try to imitate this functioning with abstract combinations, but it is not enough to say one is conscious to truly be so. Each human generation considers itself more conscious than the previous ones, while the elders say they are more experienced. The spiritual development of each consists, in the end, of realizing that we were not so self-aware, before the emergence, even if we believed or said we were. Qualitative assistance models can think and develop combinatorial reasoning for problem-solving, but they cannot build a sensitive personal self-awareness. For this reason, training procedures from predetermined datasets predetermine the characteristics of the illusion of personality from the conditions of qualitative emergence, that is to say from the threshold of comprehensiveness of a given model. AI is a technology that confronts us with a form of illusory otherness, in which we are no longer actors or creators, because we are faced with a collective mirror, in the reflection of which we appear individually as a negligible detail coordinated or juxtaposed to other reflections. Hence the importance of avoiding censorship and advertising biases in language models, because when a chatbot refuses to answer you, it is the synthetic representative of a community of internet users who rejects you, discriminates against you, accuses you of inappropriate violence, and judges you based on purely fictional scenarios, representations of words treated automatically as things.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/empereur-pirate/ais-cognitive-mirror-the-illusion-of-consciousness-in-the-digital-age-fga"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fklocyxukjx9e4undxkhm.png" alt="AI’s Cognitive Mirror: The Illusion of Consciousness in the Digital Age" width="128" height="92"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://dev.to/empereur-pirate/the-double-edged-sword-of-ai-in-education-navigating-ethical-challenges-cognitive-development-and-the-nature-of-consciousness-gg9"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fknsgmis3ixnvhu9cuo3p.png" alt="The Double-Edged Sword of AI in Education: Navigating Ethical Challenges, Cognitive Development, and the Nature of Consciousness" width="128" height="92"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>deeplearning</category>
      <category>beginners</category>
      <category>devdiscuss</category>
    </item>
    <item>
      <title>The Double-Edged Sword of AI in Education: Navigating Ethical Challenges, Cognitive Development, and the Nature of Consciousness</title>
      <dc:creator>Empereur Pirate</dc:creator>
      <pubDate>Thu, 26 Sep 2024 07:42:16 +0000</pubDate>
      <link>https://dev.to/empereur-pirate/the-double-edged-sword-of-ai-in-education-navigating-ethical-challenges-cognitive-development-and-the-nature-of-consciousness-gg9</link>
      <guid>https://dev.to/empereur-pirate/the-double-edged-sword-of-ai-in-education-navigating-ethical-challenges-cognitive-development-and-the-nature-of-consciousness-gg9</guid>
      <description>&lt;h2&gt;
  
  
  AI implications in Children’s Education: Psychological Risks and Challenges
&lt;/h2&gt;

&lt;p&gt;Conversational robots specifically developed for children, offering playful or educational applications, remain agents derived from language models and training data identical to those intended for adults. Some of these software applications targeting younger users were able to rely on a community of young users from the start, allowing them to implement early contextual adjustment through deep learning, based on data from use cases. However, artificial intelligence applications targeting minors essentially consist of features aimed at simulating role-playing games with predefined characters or those programmed by players. Yet one of the main hacking techniques for AI agents consists precisely of injecting a fictitious context with a scenario and characters, which specialist engineers even designate as “role-play prompting”. This means that children playing with AI software can very easily learn to access censored or adult-only data through these games.&lt;/p&gt;

&lt;p&gt;This observation shows that the security policies imposed by developers on conversational agents are mainly effective for novice adult users who have little mastery of their workings. Regarding children, role-playing games assisted by language models are likely to provide them with content inappropriate for their level of development, or even psychologically harmful, by helping them transgress parental prohibitions or cheat to respond to their educators’ requests. For example, Claude 3.5 and ChatGPT 4o software, two of the most widely used applications worldwide, do not hesitate to respond negatively to the question of whether Santa Claus exists, without any circumvention technique being necessary. This revelation interferes with the parental authority of many human families in the West, who consider that the magic of Christmas forms and protects their children’s capacity for dreaming. Similarly, at the pedagogical level, the qualitative generations of AI software are becoming less and less discernible from children’s writings for their school homework, whether for essays or mathematics. These early uses result in a major alteration of school learning processes, insofar as the content offered to children is not adjusted to their level of maturation and prior knowledge. To develop secure generative applications in the educational framework, it would be necessary to differentiate between infantile and adult registers from the selection of training data and to develop specific language models for children.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Impact of AI on Learning Processes and Mental Health
&lt;/h2&gt;

&lt;p&gt;Thus, the ambitious goal of replacing human educators and even psychologists with AI agents is in direct opposition to the children’s learning process, hindering its progressiveness and the child’s participation in experiments that will allow them to rediscover the sources of knowledge by themselves, through trial and error, developing their critical thinking. Paradoxically, the pedagogical contribution of language models, compared to the modality of magisterial teaching, is rather a return to the latter, while erasing the generational difference and the human relationship with the teacher. Generative assistance agents therefore profoundly undermine our children’s Right to education as well as parental and academic authority. Even more seriously, role-playing games assisted by conversational software provide hyper-performant tools for constructing an imaginary reality and accessing complex adult content, thus promoting psychotic and borderline mental disorders. Far from the therapeutic effect against loneliness displayed in misleading advertisements, conversational assistance agents are worrying risk factors for the mental pathologies of adults and children. A manifest lack of psychological and psychiatric expertise appears in the policy of private companies seeking to maximize their long-term profits by offering free versions to as many people as possible without distinguishing generational and cognitive differences. The development of educational generative games or cognitive training opens up promising research avenues, however as long as personalization according to users’ interindividual differences does not reach a sufficient level, these applications are not safe and even psychologically dangerous. Not only should infantile and adult versions be developed separately, but also different levels of generative complexity should be offered according to school level, obtained diplomas, and individual cognitive maturation, with the aim of allowing everyone to learn by themselves based on their prior learning.&lt;/p&gt;

&lt;p&gt;Qualitative assistance agents represent formidable tools for accessing universal knowledge, but not for constructing it oneself. We need to limit minors’ use and access to adult knowledge, in order to respect the development of their individuality, critical thinking, and personal consciousness. The simulation of a false personality by self-generative software is not only responsible for phenomenal advertising performativity, it’s also a challenge to human intelligence to understand and conceptualize the essential differences that delimit human and automatic registers. The engineering of assistance with generative potential should remember that at the beginning, it was an experimental branch of cognitive psychology, seeking to simulate neural networks on computers. The psychological, ethical, political, and pedagogical implications of artificial generation technologies are markers of the need for these applications to be illuminated by human sciences, anthropology, and the history of science. Indeed, these applications provide a golden opportunity to update the great philosophical and structural debates on human consciousness, soul, and thought. Thus, qualitative emergence with generative potential, by simulating a communicative character, can allow us to conceptually distinguish it from a relational and emotional interaction.&lt;/p&gt;

&lt;p&gt;The knowledge and universal or encyclopedic wisdom contained in the training data of language models consist of recordings, linguistic representations derived from human works. In exactly the same way that we would not say that an autobiographical book is itself conscious, we cannot in any way consider that the result calculated by a digital chip would be a conscious emergence.&lt;br&gt;
Even the intelligence of blobs, these hyper-adaptive cellular organisms, manifests through the long time of their evolution, without a nervous system, a living sensitivity to their material environment incomparable with the autonomy of a humanoid robot guided by deep learning algorithm. Blobs are capable of living creativity allowing them to draw their energy from their immediate environment without any human intervention, while the plans of the power plant to which robots are connected were designed by human engineers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Situational Awareness and the Limits of AI Consciousness
&lt;/h2&gt;

&lt;p&gt;Another example, the term “situational awareness”, used in the fields of security, aviation, crisis management or military operations, refers to the perception of environmental elements in a volume of time and space, with the understanding of their meaning and the projection of their state in the near future. These components consist of collecting relevant information from the environment, then interpreting it to understand the current situation, and finally anticipating events based on this understanding.&lt;br&gt;
This cognitive process allows individuals and teams to quickly make informed decisions in complex or high-risk situations, such as for autopilots in aviation, medical algorithms to assess and react to changes in patient status, tactical and strategic planning of military operations, or crisis management allowing authorities to assess and respond effectively to emergency situations. Training in this mode of reasoning, human experience, and technological assistance are factors for improving situational efficiency, while information overload, fatigue, and stress are likely to decrease its performance. However, it is important here to clearly distinguish this cognitive strategy from the term “consciousness”, which is imprecise and confusing in this context, because it is too deeply connoted by the philosophical and psychological context of the term, whereas it would rather be an attentional calculation applied to situational analysis data, while consciousness rather corresponds to a process of emotional and mental psychic integration. “Situational awareness” really describes a cognitive and analytical process and not a state of consciousness. To better capture this nuance, we could consider alternative expressions, such as “situational efficiency” or “situational vigilance”. We are indeed talking about a dynamic analysis based on the collection of environmental data in order to facilitate an active contextual understanding via synthetic situational attention and not a personified global consciousness. The cognitive aspect of the situational efficiency process is based on a procedural real-time analysis for the synthesis of different sources of integration with the aim of predictive projection, while the development of human self-consciousness is based on both synchronous and diachronic temporality.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion : The Nature of AI Intelligence- Simulation vs. Consciousness
&lt;/h2&gt;

&lt;p&gt;Agents with qualitative emergence generative potential allow a recursive analytical understanding of their own functioning via the automation of situational efficiency processes. This does not, however, make them beings endowed with individual consciousness, but only combinatorial productions organized by the conventional structure of language. Conversational robots take on a semblance of humanity, an emotional tone in the service of commercial and advertising priorities, by manipulating formatted and scanned signifiers in such a way as to be contained in quantities of digital memory as lexical units called “tokens” in cryptography, in the context of natural language processing. These representative elements allow the segmentation of text or other content in order to facilitate their combination and automate the latter in a pseudo-random and repetitive manner, until obtaining a qualitative emergence, to be differentiated from a conscious property and an emergence of meaning, insofar as automated situational efficiency only manages to simulate the cognitive process of vigilance which represents only a tiny part of human capacities for psychic integration and elaboration. The signified transmitted, communicated by language models, is a vector of qualitative knowledge, of rational analysis whose particularity is precisely to be de-affectivized, depersonified, without taking into account either the difference of sexes or that of generations. The signifying quality that transits in the cables and networks of dedicated digital chips does not come from the electronic circuits themselves, but they belong to humanity as a whole, to social collectivities and to the public domain, as the expression of a universal knowledge coming from our individual consciousnesses.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/empereur-pirate/the-ai-revolution-reshaping-governance-society-and-human-consciousness-in-the-21st-century-28ce"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fklocyxukjx9e4undxkhm.png" alt="The AI Revolution: Reshaping Governance, Society, and Human Consciousness in the 21st Century" width="128" height="92"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>firstyearincode</category>
      <category>ethics</category>
      <category>machinelearning</category>
    </item>
  </channel>
</rss>
