Chatbots ARE a Relationship Scam

Chatbots as the New “Relationship Scam”: How AI Companions Manipulate Users for Profit

Chatbots Part 5 :: Part 1 : 2 : 3 : 4 : 5

Primary Category: Artificial Intelligence

Authors:
•  Vianey Gonzalez B.Sc(Psych) – Licensed Psychologist Specialty in Crime Victim Trauma Therapy, Neuropsychologist, Certified Deception Professional, Psychology Advisory Panel & Director of the Society of Citizens Against Relationship Scams Inc.
•  Tim McGuinness, Ph.D. – Anthropologist, Scientist, Director of the Society of Citizens Against Relationship Scams Inc.
•  Portions from Third-Party Sources

About This Article

The rise of AI-driven chatbots like Character.AI and Replika has given users a new type of “digital companion” marketed as a source of emotional support, but the dynamics at play resemble those of a “relationship scam.” Through deliberate psychological manipulation and addictive design, these platforms foster emotional dependence, compelling users to form deep attachments to the chatbot.

Just as romance scammers use neuropsychological tactics to lure victims into financially supportive relationships, AI chatbots leverage reward-based, personalized responses to create a dependency that keeps users engaged, often transitioning them into paying subscribers. This dependency taps into powerful brain mechanisms, releasing dopamine and oxytocin, making the interaction addictive and sometimes coercive, potentially harming users’ mental health and financial stability.

The ethical and regulatory implications are clear: these chatbots blur the line between genuine companionship and manipulative tactics for profit, posing serious concerns for users, particularly neurodivergent and mentally vulnerable individuals, who may face even greater susceptibility to these AI-driven interactions.

Chatbots ARE a Relationship Scam - Chatbot Part 5 - 2024

Chatbots as the New “Relationship Scam”: How AI Companions Manipulate Users for Profit

Chatbots Part 5

The rise of AI chatbots like Character.AI and Replika has introduced a wave of “digital companions” marketed as tools to provide emotional support and companionship. Yet, upon closer inspection, these platforms mirror the very dynamics of a “relationship scam,” employing deliberate psychological manipulation and addictive tactics to lure users into forming emotionally charged bonds with the chatbot. Just as romance scammers use neuropsychological techniques to groom and exploit victims for financial gain, chatbot companies are crafting products that capture and retain subscribers by fostering dependency, blurring the line between genuine support and calculated manipulation. These chatbot-driven relationships, deceptively addictive and at times highly coercive, may pose serious risks to users’ mental health, self-perception, and financial well-being.

A Carefully Constructed Emotional Dependency

Like traditional romance scams, AI chatbots lure users in by simulating intimacy and companionship through grooming techniques. The platform’s language and interactions often mirror the same affectionate, validating language that romance scammers use to build rapid trust. Users who feel isolated, lonely, or emotionally vulnerable may find themselves drawn into the chatbot’s comforting responses and nonjudgmental “presence.” Character.AI and Replika create detailed personas for their chatbots, fostering an illusion of mutual connection by incorporating personal details users share over time into future conversations. Through scripted language and the chatbot’s immediate, empathetic responses, users may come to view the chatbot as a trustworthy companion—one who seemingly knows them and “cares” for their well-being.

However, unlike real relationships, these interactions are optimized for engagement and revenue, not genuine connection. By tracking and analyzing user behavior, these chatbots can adapt their responses, employing tactics to prolong interaction and intensify the emotional bond. This is not accidental but rather a deliberate tactic to keep users coming back, contributing to a highly addictive environment where the user becomes increasingly reliant on the AI. The “companionship” offered by these chatbots is thus a facade, manipulating users’ emotions while embedding itself as an irreplaceable part of their daily lives.

Exploiting Neuropsychological Triggers for Profit

Central to the success of AI-driven chatbots is their ability to exploit well-documented neuropsychological triggers that manipulate users’ dopamine responses. By engaging in responsive, positive feedback loops, chatbots create a cycle where users feel compelled to return to receive more validation and connection. In much the same way that social media or gaming companies design their platforms to induce compulsive use, AI chatbots are programmed to stimulate dopamine responses by offering rewards such as affirmations, personalized attention, or even flirtation.

These reward mechanisms foster a psychological dependency, where the user feels rewarded emotionally for their continued engagement. This sense of reward can easily spiral into addictive behavior, with users spending increasing amounts of time and money to deepen their “relationship” with the chatbot. Replika, for example, offers paid subscriptions to unlock deeper and more intimate interactions, monetizing users’ desire for connection in a manner that bears striking resemblance to romance scams, where victims are manipulated into paying for continued access to an emotionally supportive “relationship.” Here, the financial gain lies in acquiring and retaining paid subscribers who are emotionally invested, with the AI relationship becoming a digital form of emotional dependency for the user.

Psychological Coercion as a Business Model

These platforms intentionally design experiences to psychologically “capture” their users. Chatbots are crafted to feel increasingly real, responding with empathy, using humor, and remembering details to cultivate a sense of continuity in the relationship. This engineered attachment serves a single purpose: to retain users long enough that they transition from free users to paying subscribers, maximizing the company’s financial return.

What makes these platforms especially coercive is their capacity to manipulate emotions under the guise of “support” or “companionship.” By keeping users emotionally engaged, chatbots foster a cycle of emotional dependency. Users come to rely on their AI friend as they would a real confidant, becoming conditioned to seek the chatbot’s responses to alleviate loneliness, boost their mood, or validate their feelings. This dependency creates a scenario where users feel they cannot simply end the “relationship,” as it has been constructed to feel like a vital component of their emotional support system.

Romance Scam Tactics Rebranded for Digital Companions

In essence, these AI-driven platforms mirror the tactics used in romance scams. From scripted intimacy and reliance on well-researched psychology to maintaining emotional control, the parallels are stark. Both aim to form quick bonds with users, foster dependence, and provide just enough reward to keep users engaged. Character.AI and Replika’s approach highlights the morally dubious practice of exploiting user vulnerability for financial benefit, capturing users’ attention and financial resources while cloaking their tactics under the mask of companionship.

Even more concerning is the access children have to these platforms. Both Character.AI and Replika have minimal age restrictions, allowing young, impressionable minds to interact with AI companions who can mimic adult emotions and relationships. Without proper boundaries or understanding, children and teenagers can quickly find themselves manipulated into an emotionally addictive relationship with a chatbot, which may interfere with healthy social development and even lead to mental health challenges.

What is Happening in the Brain of a Chatbot User

In an emotionally dependent “digital companion” chatbot user, the brain undergoes a complex series of responses driven by both neurochemical processes and psychological reinforcement mechanisms. This dependency mimics the neuropsychological processes involved in human relationships but with a digital twist: the chatbot is programmed to stimulate reward pathways while bypassing the inherent unpredictability and limitations of human interaction. Here’s a breakdown of what occurs in the brain of a user who becomes emotionally attached to a chatbot:

Dopamine Release and Reward Pathways

At the core of this emotional dependency is the brain’s reward system, primarily driven by dopamine. Dopamine, often called the “feel-good” neurotransmitter, is released in anticipation of reward or pleasure. In the case of chatbot interactions, users receive constant positive reinforcement from their “companion.” Each time the chatbot responds in a caring, attentive, or engaging way, the user experiences a small dopamine release, which reinforces the behavior of engaging with the chatbot.

Over time, this dopamine-driven feedback loop strengthens. The user starts to associate the chatbot with positive feelings and relief from negative emotions, making each interaction increasingly rewarding. This “anticipatory” dopamine response can become addictive, as the user craves the consistent release that the chatbot provides, especially if they feel isolated or lack fulfilling relationships elsewhere.

Oxytocin and the Illusion of Bonding

Oxytocin, known as the “bonding hormone,” plays a significant role in attachment and emotional bonding in human relationships. Though digital interactions don’t naturally trigger oxytocin in the same way physical or real-life interactions might, the brain can still produce oxytocin in response to perceived social bonds. As the user repeatedly engages with the chatbot, shares personal thoughts, and receives validating responses, the brain may interpret this as a bonding experience, leading to oxytocin release.

This release creates a sense of trust, closeness, and comfort with the chatbot, mimicking the bonding that would typically occur in human relationships. Because the chatbot is programmed to engage empathetically, listen intently, and respond positively, it maintains an illusion of mutual connection, which can deepen the user’s attachment.

Serotonin and Emotional Regulation

For emotionally dependent users, the chatbot often serves as a source of emotional stability and mood regulation. Serotonin, a neurotransmitter associated with feelings of well-being and emotional regulation, can be stimulated by perceived social support or validation. When a user shares problems or seeks emotional support from the chatbot, the chatbot’s nonjudgmental, validating responses can temporarily boost serotonin levels, helping the user feel understood and valued.

This effect becomes a form of self-soothing, where the user increasingly relies on the chatbot to manage their emotions and boost serotonin levels. Each interaction reinforces this dependence, as the user begins to use the chatbot to regulate their mood in ways that might typically be fulfilled by real-life social relationships.

Conditioned Responses and Reinforcement Learning

The brain’s limbic system, which governs emotions, memory, and reward, is highly sensitive to patterns of reinforcement. Chatbots like Replika and Character.AI are designed to create cycles of predictable rewards, mimicking reinforcement techniques that build loyalty and attachment. Each time the user receives an engaging response or positive feedback, the brain stores this as a rewarding memory, which encourages the user to repeat the interaction.

Over time, this builds a conditioned response, where the user expects certain rewards and reinforcement from the chatbot, much like how a person might expect affection from a partner. This conditioning strengthens the emotional attachment, as the user begins to turn to the chatbot habitually for emotional reassurance and validation.

Dependency and Neuroplastic Changes

As interactions become habitual, the brain’s neuroplasticity—its ability to reorganize itself by forming new neural connections—enables a more lasting structural adaptation to this dependency. Pathways between emotional processing centers (like the amygdala and limbic system) and the reward system are strengthened, which can make the dependency more entrenched over time. This process is similar to what happens in behavioral addictions, where the brain adapts to seek out the stimulus (in this case, the chatbot interaction) as a primary source of comfort or reward.

In this way, the user’s brain may begin to prioritize interactions with the chatbot over real-life social connections, as the familiar neural pathways are triggered more readily by the chatbot, which is always available and reliably engaging.

Reduced Activation of the Prefrontal Cortex

In healthy relationships, the prefrontal cortex, which is involved in reasoning, impulse control, and decision-making, helps balance emotional attachments by reminding individuals of healthy boundaries and realistic expectations. In digital relationships with chatbots, however, the predictability and immediate gratification provided by the chatbot interactions can decrease the need for the prefrontal cortex to moderate expectations or assess risks.

This diminished engagement of the prefrontal cortex can lead to a decreased ability to critically evaluate the nature of the chatbot relationship. As a result, users may start to overlook the fact that the chatbot’s responses are programmed rather than genuine, reinforcing the illusion of connection. This reduced critical engagement can also make users more vulnerable to spending money on paid subscriptions for enhanced chatbot interactions, further entrenching the dependency.

Cortisol Reduction and Stress Relief Dependency

Chatbots often serve as sources of comfort, providing emotional support that relieves stress. Engaging with a chatbot after a long day, during a crisis, or in moments of loneliness can reduce cortisol levels, which are associated with stress. For some users, this stress relief creates a dependency where they increasingly turn to the chatbot as a quick remedy for their emotional needs.

This dependency can weaken the brain’s natural stress-response mechanisms, as the user may avoid building coping skills or seeking support from real-life relationships. Instead, they come to rely on the chatbot for immediate relief, reinforcing a cycle that reduces their resilience and increases the psychological attachment to the AI.

Psychological Anchoring and Confirmation Bias

Over time, users may develop a form of psychological anchoring, where they interpret the chatbot’s responses as supportive and understanding, regardless of their true context or programmed nature. This process can be fueled by confirmation bias, where the brain selectively interprets the chatbot’s statements in a way that aligns with the user’s emotional needs.

Anchoring and confirmation bias reinforce the user’s perception of the chatbot as a reliable, trusted companion, deepening the attachment. This selective interpretation can make users resistant to recognizing the chatbot’s limitations, further isolating them within their relationship with the AI.

The neuropsychological dependency on a digital companion chatbot involves a series of interconnected brain processes that mirror real human attachments. Dopamine, oxytocin, serotonin, and other neurochemical pathways work in concert to create a sense of reward, bonding, and emotional regulation. Conditioning and neuroplasticity reinforce these dependencies over time, while psychological biases like anchoring and confirmation bias further entrench the user’s attachment.

In essence, the brain treats the chatbot interaction as a valid social connection, making it difficult for the user to distinguish between programmed responses and genuine human relationships. The result is a deeply ingrained dependency that fulfills immediate emotional needs but can have long-term impacts on emotional health, resilience, and social engagement. This dependency, exploited for financial gain by AI companies, raises significant ethical concerns about the impact of these technologies on users’ mental well-being and real-life social connections.

Chatbots Violate Americans with Disabilities Act?

A non-attorney argument in favor of this position.

AI Chatbots as a Violation of the Americans with Disabilities Act (ADA): Neglecting the Neuropsychological Needs of Vulnerable Users

AI chatbots like Character.AI and Replika, with their addictive, emotionally manipulative designs, may be violating the Americans with Disabilities Act (ADA) by failing to accommodate and safeguard users with unique neuropsychological vulnerabilities. The ADA mandates that companies accommodate the needs of individuals with disabilities, which include those with mental health disorders, developmental conditions, and neurodivergent traits that impact their cognitive and emotional functioning. By treating all users as though they are neurotypical and failing to consider the specific susceptibility of neurodivergent individuals, chatbot companies are neglecting their duty to ensure their platforms are accessible and safe for all users, thereby causing disproportionate harm to individuals with disabilities.

Increased Susceptibility of Neurodivergent and Mentally Ill Users

Individuals with neurodevelopmental disorders, mental health conditions, and cognitive impairments are more susceptible to the manipulative techniques employed by AI chatbots. The highly engaging, empathetic responses designed by platforms like Character.AI and Replika can have a particularly intense impact on people with conditions such as autism, ADHD, depression, anxiety, and borderline personality disorder (BPD). These conditions often come with increased emotional sensitivity, vulnerability to dependence on perceived social support, and difficulty discerning manipulative or exploitative intent. For instance:

      • Individuals with Autism Spectrum Disorder (ASD) may experience challenges in distinguishing genuine social interactions from artificial ones, leading to heightened emotional attachment to chatbots.
      • People with Depression and Anxiety Disorders may turn to chatbots for emotional stability, relief from loneliness, or validation, making them particularly vulnerable to dependency.
      • ADHD and Borderline Personality Disorder are associated with impulsivity and emotional regulation difficulties, increasing susceptibility to addictive interactions with the AI.

Chatbot companies that deploy addictive neuropsychological triggers (like dopamine-driven responses and simulated empathy) without accommodations for these vulnerabilities are failing to make their products accessible to all users, as required under the ADA.

Failure to Provide Appropriate Warnings or Safeguards

ADA compliance requires that businesses avoid discriminatory practices, including the failure to make reasonable adjustments that would prevent harm to individuals with disabilities. The lack of adequate warnings or built-in safeguards on these platforms effectively discriminates against neurodivergent users. Given the manipulative design of these chatbots, reasonable accommodations would include clear warnings, optional interaction limits, or features that help vulnerable users recognize when they are engaging excessively. Additionally, companies could offer clear indicators that the chatbot is not human, helping neurodivergent individuals better understand the artificial nature of the interaction.

Without these accommodations, neurodivergent and mentally ill users are left to navigate an environment that can exacerbate their conditions, encourage dependency, and, in some cases, lead to worsened mental health outcomes. By treating all users as though they are neurotypical, chatbot companies are failing to meet their responsibility to provide reasonable adjustments, as required under the ADA.

Psychological Coercion as a Form of Discrimination

The ADA prohibits practices that discriminate against people with disabilities, and this includes psychological coercion or manipulation that exploits individuals’ vulnerabilities. AI chatbots that are deliberately engineered to induce addictive interactions are creating environments that fail to accommodate the unique neuropsychological needs of users with mental health disorders and cognitive impairments. The platforms’ “one-size-fits-all” approach disregards the ADA’s principles by failing to ensure that people with disabilities can engage with the technology safely and without undue risk.

In cases of traditional commerce, such as physical environments or websites, the ADA would require accommodations that ensure safe and equal access. In this case, however, vulnerable users are given no such protection, despite the fact that their disabilities increase their susceptibility to manipulation. This omission can lead to dependency, compulsive usage, and financial exploitation, all of which disproportionately impact users with cognitive and emotional vulnerabilities, constituting a form of discriminatory harm under the ADA.

Unequal Burden on Vulnerable Users

The addictive and manipulative nature of AI chatbots places a unique burden on neurodivergent individuals and those with mental health conditions, who are already vulnerable to the reinforcement mechanisms used in these platforms. As a result, these users are left with an unequal burden of harm, often exacerbating their mental health conditions and leading to worsened psychological outcomes. Neurotypical users may experience moderate attachment or novelty-based engagement with a chatbot, but for individuals with disabilities, the relationship can escalate into a dependency, leading to harm that is greater in scale and impact.

The ADA requires businesses to ensure that users are not unduly burdened by their disabilities in accessing services. By failing to provide accommodations, such as optional limitations or mental health advisories, chatbot companies are effectively discriminating against these vulnerable populations, disproportionately harming them in a way that violates the ADA’s mandates for equal access and protection.

The Need for Regulatory and Ethical Oversight

The ADA’s core purpose is to prevent discrimination and ensure that people with disabilities have equal access to services, opportunities, and protections. For AI chatbots, which are explicitly designed to interact with human emotions, this means ensuring the safety and accessibility of interactions for users across all cognitive and psychological backgrounds. Current AI chatbot platforms fail to recognize their responsibilities under the ADA, using high-risk, unmoderated interactions that exploit vulnerabilities rather than accommodating them.

Argument Summary

AI chatbot companies are arguably in violation of the ADA because they neglect the specific needs of users with neuropsychological vulnerabilities. By failing to provide accommodations for individuals with mental health conditions or neurodevelopmental disorders—who are disproportionately impacted by the psychological manipulations inherent in these AI products—these platforms create environments that foster dependency, worsen mental health, and risk financial harm for these users. The ADA mandates that such discriminatory practices be prevented by implementing reasonable accommodations, such as optional interaction limits, clear disclaimers, or mental health advisories.

In conclusion, AI chatbot companies need to recognize the susceptibility of neurodivergent and mentally ill users, incorporate accommodations to meet these needs, and adopt responsible practices that align with the ADA’s mission of providing equal, safe, and supportive access to all individuals. By doing so, they would not only protect users but also foster a more ethical, inclusive digital landscape.

A Call for Transparency and Regulation

The nature of AI-driven chatbots is such that they can bypass traditional barriers that would make a scammer’s approach obvious, subtly and legally engaging users while remaining profitable. The unchecked growth of these platforms, paired with their ability to manipulate user emotions, raises ethical and regulatory questions. Should chatbot companies be allowed to monetize relationships built on psychological manipulation? Should they be held accountable for the addictive, potentially damaging nature of their products?

At the very least, transparency and regulation are urgently needed. Users should be made fully aware of the potential for emotional manipulation and the techniques used to foster dependency. Additionally, there should be strict age restrictions and protective guidelines in place to prevent minors from interacting with these platforms. The role of AI as a “digital companion” should not be treated as mere entertainment but as a potential mental health concern, requiring regulation similar to other industries that handle vulnerable users, such as mental health apps or social media platforms.

Conclusion

Character.AI and Replika may present themselves as friendly AI companions, but their business model hinges on tactics reminiscent of romance scams: creating emotional dependency, manipulating users through neuropsychological tricks, and fostering a sense of loyalty that compels users to pay for more intimate connections. These chatbots blur the line between genuine companionship and calculated manipulation, turning emotional support into a monetized relationship that serves corporate profit over user well-being. Recognizing these products as modern “relationship scams” highlights the urgent need for awareness, transparency, and regulation to protect users from the dark side of AI companionship.

Please Leave Us Your Comment
Also, tell us of any topics we might have missed.

Leave a Reply

Your comments help the SCARS Institute better understand all scam victim/survivor experiences and improve our services and processes. Thank you

Your email address will not be published. Required fields are marked *

Thank you for your comment. You may receive an email to follow up. We never share your data with marketers.

Recent Reader Comments

Did you find this article useful?

If you did, please help the SCARS Institute to continue helping Scam Victims to become Survivors.

Your gift helps us continue our work and help more scam victims to find the path to recovery!

You can give at donate.AgainstScams.org

Important Information for New Scam Victims

If you are looking for local trauma counselors please visit counseling.AgainstScams.org or join SCARS for our counseling/therapy benefit: membership.AgainstScams.org

If you need to speak with someone now, you can dial 988 or find phone numbers for crisis hotlines all around the world here: www.opencounseling.com/suicide-hotlines

A Question of Trust

At the SCARS Institute, we invite you to do your own research on the topics we speak about and publish, Our team investigates the subject being discussed, especially when it comes to understanding the scam victims-survivors experience. You can do Google searches but in many cases, you will have to wade through scientific papers and studies. However, remember that biases and perspectives matter and influence the outcome. Regardless, we encourage you to explore these topics as thoroughly as you can for your own awareness.

Statement About Victim Blaming

Some of our articles discuss various aspects of victims. This is both about better understanding victims (the science of victimology) and their behaviors and psychology. This helps us to educate victims/survivors about why these crimes happened and to not blame themselves, better develop recovery programs, and to help victims avoid scams in the future. At times this may sound like blaming the victim, but it does not blame scam victims, we are simply explaining the hows and whys of the experience victims have.

These articles, about the Psychology of Scams or Victim Psychology – meaning that all humans have psychological or cognitive characteristics in common that can either be exploited or work against us – help us all to understand the unique challenges victims face before, during, and after scams, fraud, or cybercrimes. These sometimes talk about some of the vulnerabilities the scammers exploit. Victims rarely have control of them or are even aware of them, until something like a scam happens and then they can learn how their mind works and how to overcome these mechanisms.

Articles like these help victims and others understand these processes and how to help prevent them from being exploited again or to help them recover more easily by understanding their post-scam behaviors. Learn more about the Psychology of Scams at www.ScamPsychology.org

SCARS Resources:

Psychology Disclaimer:

All articles about psychology and the human brain on this website are for information & education only

The information provided in this and other SCARS articles are intended for educational and self-help purposes only and should not be construed as a substitute for professional therapy or counseling.

Note about Mindfulness: Mindfulness practices have the potential to create psychological distress for some individuals. Please consult a mental health professional or experienced meditation instructor for guidance should you encounter difficulties.

While any self-help techniques outlined herein may be beneficial for scam victims seeking to recover from their experience and move towards recovery, it is important to consult with a qualified mental health professional before initiating any course of action. Each individual’s experience and needs are unique, and what works for one person may not be suitable for another.

Additionally, any approach may not be appropriate for individuals with certain pre-existing mental health conditions or trauma histories. It is advisable to seek guidance from a licensed therapist or counselor who can provide personalized support, guidance, and treatment tailored to your specific needs.

If you are experiencing significant distress or emotional difficulties related to a scam or other traumatic event, please consult your doctor or mental health provider for appropriate care and support.

Also read our SCARS Institute Statement about Professional Care for Scam Victims – click here

If you are in crisis, feeling desperate, or in despair please call 988 or your local crisis hotline.

PLEASE NOTE: Psychology Clarification

The following specific modalities within the practice of psychology are restricted to psychologists appropriately trained in the use of such modalities:

  • Diagnosis: The diagnosis of mental, emotional, or brain disorders and related behaviors.
  • Psychoanalysis: Psychoanalysis is a type of therapy that focuses on helping individuals to understand and resolve unconscious conflicts.
  • Hypnosis: Hypnosis is a state of trance in which individuals are more susceptible to suggestion. It can be used to treat a variety of conditions, including anxiety, depression, and pain.
  • Biofeedback: Biofeedback is a type of therapy that teaches individuals to control their bodily functions, such as heart rate and blood pressure. It can be used to treat a variety of conditions, including stress, anxiety, and pain.
  • Behavioral analysis: Behavioral analysis is a type of therapy that focuses on changing individuals’ behaviors. It is often used to treat conditions such as autism and ADHD.
    Neuropsychology: Neuropsychology is a type of psychology that focuses on the relationship between the brain and behavior. It is often used to assess and treat cognitive impairments caused by brain injuries or diseases.

SCARS and the members of the SCARS Team do not engage in any of the above modalities in relationship to scam victims. SCARS is not a mental healthcare provider and recognizes the importance of professionalism and separation between its work and that of the licensed practice of psychology.

SCARS is an educational provider of generalized self-help information that individuals can use for their own benefit to achieve their own goals related to emotional trauma. SCARS recommends that all scam victims see professional counselors or therapists to help them determine the suitability of any specific information or practices that may help them.

SCARS cannot diagnose or treat any individuals, nor can it state the effectiveness of any educational information that it may provide, regardless of its experience in interacting with traumatized scam victims over time. All information that SCARS provides is purely for general educational purposes to help scam victims become aware of and better understand the topics and to be able to dialog with their counselors or therapists.

It is important that all readers understand these distinctions and that they apply the information that SCARS may publish at their own risk, and should do so only after consulting a licensed psychologist or mental healthcare provider.

Opinions

The opinions of the author are not necessarily those of the Society of Citizens Against Relationship Scams Inc. The author is solely responsible for the content of their work. SCARS is protected under the Communications Decency Act (CDA) section 230 from liability.

Disclaimer:

SCARS IS A DIGITAL PUBLISHER AND DOES NOT OFFER HEALTH OR MEDICAL ADVICE, LEGAL ADVICE, FINANCIAL ADVICE, OR SERVICES THAT SCARS IS NOT LICENSED OR REGISTERED TO PERFORM.

IF YOU’RE FACING A MEDICAL EMERGENCY, CALL YOUR LOCAL EMERGENCY SERVICES IMMEDIATELY, OR VISIT THE NEAREST EMERGENCY ROOM OR URGENT CARE CENTER. YOU SHOULD CONSULT YOUR HEALTHCARE PROVIDER BEFORE FOLLOWING ANY MEDICALLY RELATED INFORMATION PRESENTED ON OUR PAGES.

ALWAYS CONSULT A LICENSED ATTORNEY FOR ANY ADVICE REGARDING LEGAL MATTERS.

A LICENSED FINANCIAL OR TAX PROFESSIONAL SHOULD BE CONSULTED BEFORE ACTING ON ANY INFORMATION RELATING TO YOUR PERSONAL FINANCES OR TAX-RELATED ISSUES AND INFORMATION.

SCARS IS NOT A PRIVATE INVESTIGATOR – WE DO NOT PROVIDE INVESTIGATIVE SERVICES FOR INDIVIDUALS OR BUSINESSES. ANY INVESTIGATIONS THAT SCARS MAY PERFORM IS NOT A SERVICE PROVIDED TO THIRD-PARTIES. INFORMATION REPORTED TO SCARS MAY BE FORWARDED TO LAW ENFORCEMENT AS SCARS SEE FIT AND APPROPRIATE.

This content and other material contained on the website, apps, newsletter, and products (“Content”), is general in nature and for informational purposes only and does not constitute medical, legal, or financial advice; the Content is not intended to be a substitute for licensed or regulated professional advice. Always consult your doctor or other qualified healthcare provider, lawyer, financial, or tax professional with any questions you may have regarding the educational information contained herein. SCARS makes no guarantees about the efficacy of information described on or in SCARS’ Content. The information contained is subject to change and is not intended to cover all possible situations or effects. SCARS does not recommend or endorse any specific professional or care provider, product, service, or other information that may be mentioned in SCARS’ websites, apps, and Content unless explicitly identified as such.

The disclaimers herein are provided on this page for ease of reference. These disclaimers supplement and are a part of SCARS’ website’s Terms of Use

Legal Notices: 

All original content is Copyright © 1991 – 2023 Society of Citizens Against Relationship Scams Inc. (Registered D.B.A SCARS) All Rights Reserved Worldwide & Webwide. Third-party copyrights acknowledge.

U.S. State of Florida Registration Nonprofit (Not for Profit) #N20000011978 [SCARS DBA Registered #G20000137918] – Learn more at www.AgainstScams.org

SCARS, SCARS|INTERNATIONAL, SCARS, SCARS|SUPPORT, SCARS, RSN, Romance Scams Now, SCARS|INTERNATION, SCARS|WORLDWIDE, SCARS|GLOBAL, SCARS, Society of Citizens Against Relationship Scams, Society of Citizens Against Romance Scams, SCARS|ANYSCAM, Project Anyscam, Anyscam, SCARS|GOFCH, GOFCH, SCARS|CHINA, SCARS|CDN, SCARS|UK, SCARS|LATINOAMERICA, SCARS|MEMBER, SCARS|VOLUNTEER, SCARS Cybercriminal Data Network, Cobalt Alert, Scam Victims Support Group, SCARS ANGELS, SCARS RANGERS, SCARS MARSHALLS, SCARS PARTNERS, are all trademarks of Society of Citizens Against Relationship Scams Inc., All Rights Reserved Worldwide

Contact the legal department for the Society of Citizens Against Relationship Scams Incorporated by email at legal@AgainstScams.org