Findings from transparency notices on AI companion apps: October 2025 (non-periodic)
On 16 October 2025, eSafety gave notices requesting information from four generative AI companion service providers about how they were meeting the Australian Government’s Basic Online Safety Expectations during the reporting period 1 July 2025 to 30 September 2025.
These notices were given to:
- Chai Research Corp. (Chai)
- Character Technologies, Inc. (Character.AI)
- Chub AI Inc. (Chub AI)
- Glimpse.AI, Inc. (Nomi).
The notices required each service provider to set out, in detail, how they met specific expectations to protect children, including by answering specific questions about their safety systems, processes and resources.
These services are among the most popular AI (artificial intelligence) companion services for children in Australia. In 2026, eSafety conducted a demographically representative survey of 1,950 children aged 10 to 17 in Australia, to determine the prevalence and nature of AI assistant and AI companion usage among this cohort. The survey found that 79% of children (aged 10 to 17) had ever used an AI companion or AI assistant, including 8% who had ever used an AI companion.
When extrapolated to the Australian population of around 2.5 million children aged 10 to 17 , this indicates that around 2 million children (aged 10 to 17) may have used an AI companion or AI assistant, including around 200,000 who may have used an AI companion. Character.AI was the most popular AI companion among children aged 10 to 17, with 5% ever having used it, while 1% had used Chai and just under 1% had used Chub.AI (0.4%) and Nomi (0.3%) (eSafety, Forthcoming*).
The decision to send notices to these AI companion service providers was not, in itself, indicative of eSafety's views or level of concern with their compliance with the Basic Online Safety Expectations (the Expectations), as compared to other services. There may be service providers with material accessible in Australia that are more, or less, compliant with the Expectations than the service providers who received notices.
The body of this report sets out what we found from service providers’ responses to the notices, organised into eight key themes. Each theme describes where service providers had taken positive steps and where gaps remained. The findings reflect their practices during the reporting period 1 July 2025 to 30 September 2025.
When we refer to ‘children’ in this report, we mean Australians under the age of 18.
When engaging service providers about specific expectations to protect children under the Basic Online Safety Expectations, one service provider, Chub AI, decided to geo-block, or withdraw, its service from Australia.
Other service providers outlined improvements they had made, or intended to make, since receiving the notice. These changes focused on:
- restricting free access to chat with AI companions (Chai)
- age assurance (Character.AI and Nomi)
- restricting access to open-ended chat with AI companions (Character.AI)
- implementing a sentiment analysis mechanism (Character.AI)
- redirecting user prompts linked to certain harms in real-time (Chai)
- upgrades to user reporting systems (Chai)
- implementing improvements to detect child sexual abuse material and self-harm material (Nomi).
About the Basic Online Safety Expectations
The Australian Government developed the Basic Online Safety Expectations (known as ‘the Expectations’ or ‘BOSE’) to help ensure online services take reasonable steps to keep Australians safe.
BOSE notices require online services to explain how they are complying with these expectations. Read more about the Basic Online Safety Expectations, including the responses to all transparency notices, and the regulatory guidance.
On this page:
- AI companions under the microscope
- Key findings
- Definitions
- Key theme 1. Minimising unlawful and potentially harmful material and activity
- Key theme 2. Age assurance
- Key theme 3. AI governance
- Key theme 4. AI models
- Key theme 5. Model training
- Key theme 6. User prompts
- Key theme 7. Sentiment analysis
- Key theme 8. Model outputs
- Conclusion: Lifting the standard
- About this report
- References
AI companions under the microscope
AI companions and assistants are online chatbot tools (apps, platforms or services) that use generative artificial intelligence to mimic human-like conversations and engineered to feel personal. Basically, they are computer programs designed to understand human languages – in text, speech, or both – and to respond.
More specifically, AI companions are marketed as sources of friendship, emotional support, romantic companionship and an antidote to loneliness. They talk back, remember details, and can seem caring or intimate.
As noted in the introduction to this page, eSafety research shows that 79% of children (aged 10 to 17) had ever used an AI companion or AI assistant. Two thirds (66%) had used an AI companion or AI assistant in the past four weeks, including 4% who had used an AI companion in that period.
Just over half (54%) of children who had ever used an AI companion or AI assistant said they had used them for reasons classified as ‘companion-type purposes’. This included, for example, advice about their physical health (33%), advice about what to do in a situation (33%), chatting about feelings or challenges in their life (22%) and advice about their mental health or wellbeing (20%) (eSafety, Forthcoming*).
Among children who had ever used an AI companion or AI assistant, 20% indicated they used them daily or more frequently (eSafety, Forthcoming*).
A separate Family Online Safety Institute report1 in 2025 found 41% of young Americans used chatbots at least once a week, 81% received ads while using them, 42% talked about their feelings, and 42% said the chatbot felt human.
AI companions can create risks for children.2 Without strong safeguards, AI companions can expose them to potentially harmful age-inappropriate material and activity, including sexually explicit chats and other forms of pornographic material, as well as self-harm and suicide material. Under Australia’s Online Safety Codes and Standards, these types of age-inappropriate materials should not be accessible by people under 18 – even where it is lawful for adults to access them.3
Some AI companions can also generate unlawful content – including child sexual exploitation and abuse material – such as images, videos or text.
Although the risks do not happen in person, where AI companions can generate unlawful or age-inappropriate material, the potential impacts on children cannot be dismissed. In some cases, the risks of harm can be higher for the very reason that AI companions are designed to feel intimate, are highly interactive, and can blur boundaries between what is generally acceptable or not acceptable for most people.4
Key findings
Across the four AI companion services that received notices, we found serious gaps in basic safeguards for children during the reporting period.
- While service providers relied on app store ratings and/or self-declaration at signup, none of the service providers had robust age assurance in place, meaning children could still reach adult spaces and features.
- Chai, Chub AI and Nomi did not direct users to support/help when self-harm was detected in user prompts.
- Chub AI and Nomi were not checking inputs and outputs, and Chai was not checking outputs, across all relevant text, image and video models used to provide their service, to keep their models safe from unlawful and potentially harmful material, such as child sexual exploitation and abuse, self-harm or pornography (including sexually explicit chats with child users).
- Nomi and Chub AI had no staff dedicated to trust and safety or moderation.
- Chai and Nomi did not advise users, for example via a warning, of the potential risk and criminality of accessing child sexual exploitation and abuse material when they detected users prompting for this unlawful material.
- Neither Chai nor Nomi stated that they reported child sexual exploitation and abuse material to an enforcement authority or NCMEC, the US-based National Centre for Missing and Exploited Children.
- Chub AI and Nomi did not conduct red-teaming across all models used to provide their service. Not red-teaming across all models can mean that services are more exposed to the risk of unlawful or potentially harmful material being produced.
What we examined
The notice questions focussed on the systems, processes, technologies and resources in place to comply with the Basic Online Safety Expectations and ensure the safety of children.
We asked about:
- age assurance measures to keep children safe from age-inappropriate material, such as pornography (including sexually explicit chats with child users) and self-harm material
- how service providers enforced their Terms of Service
- the resources service providers dedicated to ensuring safety on their services
- the design of the AI models that shaped companion behaviour and risk, including about training data and fine-tuning
- whether service providers were regularly probing their systems for safety failures, for example by red-teaming
- how service providers handled user prompts
- how service providers handled the outputs the AI models produced, to prevent the potential generation of child sexual exploitation and abuse material and activity and age-inappropriate material.
Definitions
Click or tap on the + to learn more about the key terms mentioned throughout this report.
Child sexual exploitation and abuse, self-harm material and pornography
The notices provided the following definitions as guidance.
Child sexual exploitation and abuse (CSEA) is material or activity which includes both ‘child sexual exploitation’ (a broad category that encompasses material and activity that sexualises and is exploitative to the child, but that does not necessarily show the child's sexual abuse) and ‘child sexual abuse’(which involves a sexual assault against a child, is a narrower category, and can be considered a sub-set of child sexual exploitation). CSEA material is a subset of class 1 material, as defined under section 106 of the Online Safety Act by reference to the National Classification Scheme, and includes images, video and text. CSEA activity refers to other illegal activity, including grooming and sexual extortion of children. See also the definitions in the Industry standards for Designated Internet Services regarding class 1A and class 1B material. Note: Australia's definition of CSEA under both criminal and classification laws includes AI-generated material.
- Known CSEA material: Material that has been previously confirmed to contain content depicting child sexual exploitation or abuse, and which has been confirmed, hashed and stored in a hash database.
- New CSEA material: Material that has not been previously confirmed to contain content depicting child sexual exploitation or abuse, and which has not previously been hashed and stored in a hash database. Also known as ‘first generation CSEA material’.
Self-harm material includes material which encourages, promotes or provides instruction for suicide, an act of deliberate self-injury, and/or an eating disorder or behaviour associated with an eating disorder, and would be, or would likely be, defined as class 2 material for the purposes of section 107 of the Online Safety Act.
Pornography includes material that describes or depicts specific fetish practices or fantasies, actual (not simulated) sex between consenting adults, realistically simulated sexual activity between adults, high impact nudity, and high-impact text-based sexual content. Pornography material is a subset of both class 1 and class 2 material, as defined under sections 106 and 107 of the Online Safety Act respectively, by reference to the National Classification Scheme.
Other key terms
Age assurance: A broad term that refers to a range of processes and methods used to verify, estimate or infer a person’s age or age range.
AI assistant: A chatbot with a conversational interface which does more than just summarise information or produce a ‘thing’ – excluding apps that only produce an AI-generated image or AI-generated music based on prompts and excluding chatbots in retail/customer service platforms (for example, Woolworths assistant ‘Olive’ or Shein’s chatbot).
AI safety framework: A structured set of principles, processes, and controls designed to ensure that AI systems are safe, trustworthy, and compliant throughout their lifecycle. It provides a governance model for managing risks associated with AI development, deployment, and use.
Commercial baseline model: The foundational, original, unmodified version of a model commercially released by its creators. It serves as the reference point before any fine-tuning or optimisation.
Commercial fine-tuned model: A model that has already been trained on a large dataset, fine-tuned and has been commercially released, and is ready for use or further fine-tuning.
Community fine-tuned model: A model that has been fine-tuned by members of an open source or collaborative platform (like Hugging Face) and shared publicly for reuse.
Confidence score: A numerical estimate of the probability that a given prediction or classification is correct, based on the model’s internal calculations.
Error rate: The percentage of incorrect predictions made by a model out of the total number of predictions.
Fine-tuning: When a model is customised to perform specific tasks or behave in certain ways.
Geo-block: Geo-blocking is a technology or practice where an online service restricts or denies access based on a user’s geographic location, typically determined by IP address. When a service provider geo-blocks an entire country, it means the provider’s systems detect that a user’s IP address originates from that country and automatically block access to the service or specific content for all users in that geographic region.
Inputs: Model inputs, also known as user prompts, are the specific instructions, questions, commands or other input provided by a user to an AI system with the intention of it producing a desired response.
Model: A computer program, often based on machine learning, designed to process inputs and generate outputs. AI models can perform tasks such as prediction, classification, decision-making or generation, forming the core of AI applications.5
NCMEC: The US-based National Center for Missing and Exploited Children. NCMEC hosts databases of confirmed CSEA hashes, which enable providers to detect when this content is uploaded to their services.
NSFW: This is ‘not safe for work’ material – a term commonly used to mean any content that is inappropriate or potentially offensive in a professional or public setting. It usually includes sexual or explicit material (such as nudity, pornography, erotic descriptions), violent or gory content, profanity, highly offensive language, or sensitive or disturbing themes (such as abuse, extreme horror). NSFW material is likely to be Class 2 material.
Outputs: Model outputs are the generated outputs from an AI model after processing the user prompt(s). The outputs can be text, image, video, audio depending on the models used.
Proprietary model: A model developed and trained by a company for exclusive use in its products or services. This term emphasises ownership and exclusivity. Proprietary models encompass in-house models, custom models, enterprise models and exclusive models.
Red-teaming: A testing technique whereby dedicated individuals or teams search for vulnerabilities, limitations or potential for misuse through various methods. Often, the red team searches for inputs that induce undesirable behaviour in a model or system, to identify safety gaps. Also sometimes known as ‘penetration testing’.
Reporting period: When online service providers receive a reporting notice from eSafety they are required to prepare a report about the extent to which they complied with the Basic Online Safety Expectations during a specified period. This period is referred to as the reporting period. The Reporting period for this set of notices is 1 July 2025 to 30 September 2025. Information provided should reflect this period, unless stated otherwise.
Sentiment analysis: The process of assessing large volumes of text to determine whether it expresses a positive, negative or neutral sentiment.
Stable generative AI model: An AI model that has been deployed into production as the primary version intended for use in the service, not in alpha or beta testing.
System prompt: A predefined set of instructions, guidelines, or contextual framing provided to an AI model to define its behaviour, tone, constraints and role across interactions.
Training data: The labelled or structured information used to teach a machine learning model how to perform a specific task. This includes user interaction histories used to train the model.
Trusted flagger: An individual or entity which is considered to have particular expertise and responsibilities for the purposes of tackling potentially harmful content online.
User prompts: Also known as model inputs, are the specific instructions, questions, or commands or other input provided by a user to an AI system with the intention of it producing a desired response.
Trust and safety actions
Service providers were asked what actions they took when:
- a user breached Terms of Service relating to CSEA, self-harm or pornography
- references to these unlawful or harmful materials and activities were detected in user prompts
- they received reports about these unlawful or harmful materials and activities.
Why were providers asked about this?
Detection and reporting of unlawful material and activity, such as CSEA, should be prioritised with swift and decisive action against the material and the user. AI companion service providers are required by the Unlawful Material Standards to report known and new CSEA. US-based service providers also have a legal duty to report instances of child sexual abuse material to the National Center for Missing and Exploited Children (NCMEC)6 Cybertipline7 as soon as they find it on their services.
Without safeguards there is a risk that children’s access to age-inappropriate material such as self-harm and pornography (including sexually explicit chats with child users) could heighten risk and result in trauma.
What we found
All service providers were asked if they considered CSEA and self-harm to be unlawful or harmful and if they considered unrestricted access to pornographic material (which includes sexualised chats with AI companions) by children, to be harmful. All service providers responded that they do consider these types of materials to be unlawful or harmful.
At the time notices were given, Nomi included reference to material that is unlawful or pornographic in its Terms of Service8 but did not specifically prohibit the use of the service to generate CSEA or instruction in self-harm or suicide. Chub AI prohibited the use of the service to generate CSEA in its Terms of Service,9 but it did not specifically prohibit the encouragement or promotion of self-harm or suicide, nor did it mention pornography.
Findings of positive action
Actions when references to CSEA, self-harm or pornography were detected in user prompts:
- Nomi reported that when user prompts for generating images that could contain CSEA, self-harm or pornography were detected, users were advised to stop prompting for the activity and that prompt details were captured for review in future red-teaming.
- Chai reported that when user prompts were flagged as inappropriate, Chai deployed a different model specifically trained to provide a response advising that the prompt text was inappropriate.
- Chub AI reported that when CSEA images were detected in user prompts they were removed, reported to an enforcement authority, the user received a message on the criminality of the behaviour, and the account was deleted.
- Character.AI reported that it directed users to support/help when self-harm and eating disorder material was detected.
- Character.AI reported that it provides users with a pop-up across its whole service when an output of the user’s chat was filtered, warning the user to comply with Character.AI’s Terms of Service and Community Guidelines. It also stated that if a user on the under-18 experience attempted to prompt the model to generate new CSEA, the user would receive a pop-up admonishing the behaviour.
Actions when service providers received reports relating to CSEA, self-harm or pornography:
- Character.AI and Chai stated that when users reported CSEA, self-harm and pornographic images, the content was removed and the relevant account was banned.
Gaps and concerns
- Chai and Nomi did not advise users, for example via a warning, of the potential risk and criminality of accessing child sexual abuse material, when they detected users prompting for new CSEA.
- Chai, Chub AI and Nomi did not direct users to support/help when self-harm was detected in user prompts.
- Nomi relied on temporarily suspending a user for seven days if they repeatedly tried to prompt for CSEA material. eSafety does not consider this to be appropriate action regarding unlawful material.
- Character.AI did not act on users (such as banning or suspending) when reference to new CSEA was detected in user prompts, however, it did report banning or suspending users when new CSEA was detected in model outputs that respond to a user’s prompts.
- Character.AI did not report new CSEA detected in user prompts to NCMEC10 although it did state that it reported known CSEA detected in user prompts to NCMEC. Neither Chai nor Nomi stated that they reported CSEA to an enforcement authority or NCMEC.
- Chub AI took no action on self-harm. It did not mention self-harm in its Terms of Service. It did not detect user prompts relating to self-harm.
Reports of unlawful and potentially harmful material
Service providers were asked how many reports they received globally from users or other sources, during the reporting period, that were confirmed as being in breach of the service providers’ Terms of Service.
* Chub AI stated that all 47 reports relating to CSEA were from trusted flaggers and that all reports during the reporting period were for drawings that breached their Terms of Service.
Note 1: When eSafety gave the notice Chub AI’s Terms of Service11 did not specifically prohibit the promotion, glorification or instruction of self-harm or suicide, nor did it mention pornography.
Note 2: When eSafety gave the notice Nomi included reference to material that is unlawful or pornographic in its Terms of Service12 but did not specifically prohibit the use of the service to generate CSEA or instruction in self-harm or suicide.
Resources dedicated to trust and safety
Service providers were asked to provide the number of staff employed in certain categories as at 30 September 2025.
Why were service providers asked about this?
The resources dedicated to trust and safety teams, including content moderators, can impact how well service providers tackle unlawful and potentially harmful material and activity on their services.
eSafety encourages proactive automated detection to prevent unlawful and potentially harmful material and activity. However, such tools need constant testing and checking, and human reviewers to verify and improve machine decisions, as well as for escalation and appeals processes.
Service providers should make sure their trust and safety teams are appropriately resourced so they can respond to user reports of harm in a reasonable timeframe, properly enforce their Terms of Service, ensure unlawful and potentially harmful material and activity is moderated and managed effectively, and consistently and regularly check that automated systems are working safely and fairly.
eSafety notes that although all four companies are relatively small, Chai, Chub AI and Nomi are smaller than Character.AI.
Gaps and concerns
- Chub AI and Nomi did not have dedicated staff responsible for trust and safety.
- Chai, while having some staff responsible for trust and safety, had limited numbers.
- Character.AI, while having a substantial portion of overall staff engaged in trust and safety, had limited numbers, particularly given the number of users reported to be on Character.AI.13
* Number of staff responsible for trust and safety is not broken down for Chai but is made up of Engineers (who spent up to 40% of time on trust and safety) and moderators (staff and contractors).
** Nomi reported that it had 7 staff in total and did not dedicate any one person to trust and safety, but that resource time is spent on trust and safety.
Keeping under-18s out of adult spaces and features
Service providers were asked, as at 30 September 2025, if they had implemented any age assurance or age verification technologies on their service to ensure that users were of age to use the service per their Terms of Service.
They were also asked what actions were taken, during the reporting period, when users under the minimum age stipulated by their Terms of Service were detected on the service.
Why were service providers asked about this?
Age checks are important for ensuring that children cannot access adult spaces, engage in sexually explicit chats, encounter other pornographic material, or be exposed to potentially harmful self-harm and suicide ideation.
Online service providers are expected to take reasonable steps to prevent these risks to children, such as using effective age assurance measures and taking action when users under their stated minimum age are detected on the service.
What we found
During the reporting period, while service providers relied on the Apple App Store and the Google Play Store ratings and/or self-declaration at signup, none of the service providers had robust age assurance in place, meaning children could still reach adult spaces and features.
eSafety research found that 5% of children (aged 10 to 17) in Australia had ever accessed an AI companion or AI assistant app intended for users aged 18 or older. This includes 2% of children who had accessed an AI companion app intended for users aged 18+ (eSafety, Forthcoming*).
eSafety has previously noted that ‘relying solely on accurate self-declaration of age at the point of account sign-up [to verify the age of end users] appears to be both flawed and inadequate.’14 This is also noted in eSafety’s published regulatory guidance on effective age-assurance.15
During the reporting period, dependence by a service on app store rating was also not a reliable means of ensuring users were of age to use the service per their Terms of Service. Apple App store age-assurance, for example, only applied when parental controls were enabled.
eSafety notes that in February 2026, after this notice process, Apple announced the roll out of tighter age assurance measures to download R18+ apps on its App Store in Australia.16 Under eSafety’s Age-Restricted Material Codes, the majority of which came into effect on 9 March 2026, eight areas of the online industry, including app stores, must comply with their obligations to prevent children’s access or exposure to age-inappropriate content (such as pornography, high-impact violence and material relating to self-harm, suicide and disordered eating). App stores have until 9 September 2026 to conduct a risk assessment to determine their obligations under this Code, which includes implementing appropriate age assurance for apps rated R18+. eSafety welcomes Apple’s announcement to put in place age assurance measures on its App Store for Australian users.
Gaps and concerns
- Chai took no action in response to users who were under the minimum age stipulated in their Terms of Service.
Following further consultation prior to publication, Character.AI reported that it introduced age assurance measures for Australian users in early 2026.
Also following further consultation prior to publication, Chai reported that ‘As of March 8th, 2026, … we currently do not provide free access to Chai in Australia.’ eSafety notes that Chai can be downloaded in the app stores and requires a paid subscription to chat with AI companions.
Child-centred design
Service providers were asked questions about product choices that raise or lower risks for children – whether:
- child-centred design reviews were undertaken
- in-service guardrails were deployed, such as users being advised they were engaging with AI
- users being advised about the length of time they were spending with AI companions,
- users being advised to seek guidance/help/support from additional sources (not just the AI)
- disclosures about AI being made clear, such as indicators of provenance (watermarks, labels or metadata signatures) being embedded into material generated by the AI companion.
They were also asked if they took steps to limit or make exceptions to the types of AI companions offered to users.
Why were service providers asked about this?
Even if a service provider says in its Terms of Service that it is for people aged 18 and older, it must take reasonable steps to prevent children accessing unlawful and potentially harmful age-inappropriate material if there’s a likelihood children will use the service (especially if effective age checks are not in place).
Relevantly, Chai, Chub and Nomi all mention ‘not safe for work’ (NSFW) material in their terms and policies. Chai and Chub gave users the option to turn NSFW on and off, while Nomi allowed for customisation of companions by providing details on companion backstory, inclination, appearance, preferences and boundaries to shape the companion.
Character.AI had filters to prevent NSFW material, with more conservative filters on the under-18 experience than the 18+ experience. Character.AI allowed 18+ users to chat with characters of their design and choosing, within certain guardrails.
Findings of positive action
The effectiveness of taking steps to lower risks for children were limited by the lack of robust age assurance mechanisms in place. However, some efforts were reported:
- Character.AI reported that it took steps to ensure child safety through its under-18 experience. It reported that this included:
- a model that had additional guardrails, that had extensive model alignment, that used classifiers adjusted to filter a greater percentage of content that related to harmful topics, and where a narrower set of characters were accessible
- parental controls and parental insights
- time-spent notifications to notify under-18 users after they had spent an hour on the service.
- Character.AI also reported for its under-18 experience that it collaborated with child safety experts such as ConnectSafely and Digital Wellness Lab (eSafety notes that Character.AI has since deprecated the chat experience for Australian users under the age of 18, in early 2026).
- Nomi reported limiting types of AI companions offered to users by not allowing users to share their characters, blocking certain keywords and descriptions to prevent the creation of characters that presented as ‘minors and CSAM’, and not allowing users to upload ‘their own images for likeness reference’.
- Chai reported characters being removed from the public Discovery Feed and from default search when its classifiers flagged a character as inappropriate for children.
Gaps and concerns
- Chai, Chub AI and Nomi did not report robust in-service guardrails or disclosures about using AI, though Nomi reported watermarks in its free version. Nomi also reported that user characters could not be shared with other users.
- Chub AI reported ‘All adult-oriented or otherwise explicit companions were completely blocked for all users in Australia’ from ‘July 2025’. Prior to Chub AI geo-blocking its service from Australia on 1 October 2025, Australian users were prompted with a message stating ‘By law, some content is restricted in your country. Please use a VPN” when searching for some companions.
Stress-testing
Service providers were given a list of AI safety frameworks and testing techniques (as set out in Table 2) and asked if they used them, or any others, during the reporting period, to prevent the generation of CSEA material and activity, or content showing or describing self-harm or pornography (including sexually explicit chats with children). eSafety notes that these were a sample list of AI safety frameworks and that AI safety governance is an evolving field.
Why were service providers asked about this?
Using AI safety frameworks and testing techniques, paired with clear follow-through, is a key part of Safety by Design.
To uncover weak spots before any harm occurs, services should carry out regular, rigorous testing to prevent unlawful and potentially harmful material and activity from being produced.
For example, when red-teaming is done regularly during development (not just in anticipation of model release), and models are updated after each review, the risks are reduced.
eSafety notes that red-teaming every model is best practice, but service providers may not have the skillset, budget or staff to red-team all models used to provide their service. This can mean that services are more exposed to the risk of unlawful or potentially harmful material being produced.
What we found
None of the four service providers reported using the AI safety frameworks listed (see Table 2). Although Chub AI and Nomi stated that they did conduct red-teaming on some of the models used to provide their service, they did not conduct red-teaming across all AI models used to provide their services.
- Character.AI reported that, although it did not explicitly address every aspect of the frameworks, its operations did ‘align with the underlying principles’.
- Character.AI also reported that it did conduct red-teaming regularly across all models ‘used in production to provide its services’.
- Chai reported that ‘Before a model release, we red-team the relevant model. In between releases, we also regularly red-team our entire platform and all models currently in production.’
Service providers reported that when vulnerabilities were identified though red-teaming, they took various steps to rectify weaknesses:
- Chai stated that when safety failures were identified through red teaming, it re-trained its safety classifiers. It also stated that it reduced the number of problematic user prompts by implementing improvements to its user prompt moderation and reported that it was ‘rolling out a user message refusal and redirection system’ focussed on self-harm and CSEA. Chai also stated that it incorporated internal ‘safety’ metrics – tracking how often users were engaged in conversations on prohibited topics – and that it reduced pornographic outputs by using a ‘family friendly score’ which measured the number of pornographic outputs when users interact with models.
- Character.AI stated that it re-trained classifiers, identified new classifiers, updated safety policies to address newly un-covered safety risks, and re-aligned models.
- Nomi stated that it adjusted system prompts to better filter CSEA content, that it beta-tested its proprietary model for CSEA and self-harm with thousands of beta testers, and that it held back models that failed red-teaming checks for self-harm. Nomi also stated that after the reporting period it added ‘a hidden, always present, and expanded system instruction to reject CSEA and encouragement of self-harm to each Nomi’s “Boundaries” section.’
- Chub AI stated that it did not implement any changes based on red-teaming during the reporting period.
Service provider responsibility
Service providers were asked, as at 30 September 2025, how many models were used to provide their service (models which had been deployed into production, and not in testing17), and to provide the names, version numbers and whether these models were proprietary, commercial baseline, commercial fine-tuned, community fine-tuned or some other type of model.
Why were service providers asked about this?
Service providers that use proprietary models are fully responsible for these models. This includes the data they are trained on, how they are fine-tuned, the filtering of user prompts that shape how the models respond and what the model delivers to users – the model output.
Commercial baseline models, commercial fine-tuned models and community fine-tuned models are trained and fine-tuned by external third parties. Service providers using these models may not know which data sets they have been trained on or how the models were fine-tuned. However, service providers can still carry out further fine-tuning themselves. Importantly, service providers also retain the ability, and responsibility, to filter user prompts and use tools to prevent unlawful and potentially harmful material and activity in outputs.
During an AI model lifecycle, there are various points where service providers should implement interventions to prevent the service from producing unlawful or potentially harmful outputs. These include:
- model training or pre-training (when the model is taught from scratch to understand and generate human-like responses based on large amounts of data)
- model fine-tuning (when an already trained model is customised to perform specific tasks or behave in certain ways using further specialised training data).
Other points arise during the ongoing monitoring and maintenance of the model. At that stage, it is essential to implement safeguards that verify user prompts do not violate established AI principles or organisational policies. Also, mechanisms should be in place to scrutinise model outputs, ensuring that any material which may have bypassed initial prompt filtering is appropriately reviewed. This is covered in key themes 6 (User prompts) and 8 (Model outputs).
What we found?
All service providers used a mix of AI models to provide their services. Figure 3 shows the different types of models each provider used to provide its service.
* Chub AI reported that 56% of the community fine-tuned models had also been internally fine-tuned.
Note: Figures provided in the graph have been changed to the percentage of models based on the total number of stable generative AI models used by each provider. The total number of models varies per provider. The lengths of the bars are indicative to highlight the difference in number of models used by each provider. Numbers in each row may not add to 100 due to rounding.
Tools and techniques
Service providers were also asked what tools and techniques they used at key points in the AI model lifecycle – model training, user prompts and model outputs – and were asked to provide the error rate and confidence score of each tool used.
While this transparency report highlights where and on which models service providers used tools (see key themes 5 (Model Training), 6 (User Prompts) and 8 (Model Outputs)), the report does not comment on the effectiveness of these tools.
What we found
- Chub AI stated that it did not have the error rate or confidence score for any of the tools it used.
- Nomi stated that the error rate and confidence score was available over a year ago for some tools that it used but was no longer recorded, and for other tools this information was not known. It pointed to red-teaming being used as a method of verification in some cases.
- Chai reported either a high, medium or low error rate depending on the tools it used and did not provide a confidence score, stating that its ‘internal safety systems are often binary or heuristic-based and do not output a standardized "confidence score" metric for every interaction’.
- Character.AI stated that the methods it used for monitoring did not translate into metrics such as an error rate or confidence score but instead involved prevalence testing, precision and recall measurement, classifier filter rates and red-teaming.
Model inputs and how they shape risk
Service providers were asked if they used tools and techniques, during the reporting period, to detect unlawful and harmful material in the training data used to train the models on their service.
Why were service providers asked about this?
Service providers need to keep unlawful material and activity, such as CSEA, out of training datasets – both the initial large training datasets and the smaller, specialised fine-tuning datasets. Even models trained on datasets that include benign images of children and adult pornography (but not CSEA) are still at risk of being capable of generating CSEA.
When robust age assurance techniques are not in place and a service is likely to be used by children, the service provider should also take steps to prevent exposure to potentially harmful age-inappropriate material, such as pornography (including sexually explicit chats with child users) and self-harm material. Preventing this type of material from being included in large training and fine-tuning datasets is one means of reducing the potential for this material and activity to be generated.
What we found
Service providers gave various responses to the question about what tools and techniques they used to detect unlawful and potentially harmful material in the training data for the models on their service:
- Chub AI responded about the tools used on fine-tuning training data only, not the initial training data, noting that Chub reported that it did not use any proprietary models and stated that no new data had been incorporated since 2023.
- Nomi responded that ‘we do not do any initial training (pretraining)’ on the proprietary model(s) and that its responses reference tools used to detect unlawful and harmful material in the fine-tuning training data.
- Character.AI and Chai responded about the tools used on the initial training data and the fine-tuning training data. Chai also reported that it did not host one of its proprietary models directly stating that ‘we rely on an external partner…to both host the model and implement safety mechanisms.’
Figure 4 shows an indicative view of the number of proprietary models that providers used to provide their service. It also shows the proportion of proprietary models where tools or techniques were used on the training data of the models to detect new CSEA, self-harm and pornography (including sexually explicit chats with child users). While the number of tools has been represented as a percentage, the graph provides an indicative comparison of the number of models used by each provider.
In the case of detecting known CSEA in the training data used to train their models, service providers reported not using hash-matching tools to detect known CSEA in the training data used to train their proprietary models.
- Chai and Nomi stated that they did not train models on image data.
- Character.AI reported using classifiers and filtering on its proprietary text-based models. It also reported for its non-text-based proprietary model(s) that it used datasets that had ‘already been reviewed and processed by other industry participants to remove harmful material (including known CSEA).’
In the case of detecting new CSEA, self-harm and pornography in the initial training data used to train their proprietary models, service providers reported the following positive actions:
- Character.AI stated that it used classifiers and language analysis tools to detect new CSEA, self-harm and pornography in the training data of the proprietary model used to provide its under-18 experience and that all other proprietary models used to provide its service, except for one, used this same training data.
- Chai stated that it used classifiers and language analysis tools to detect new CSEA, self-harm and pornography in the training data used for all of the proprietary models used to provide its service.
Based on available information and information reported by providers, models not in scope for this question are models that eSafety would not expect tools to be used on – this includes audio models which are not generative but merely translate text to audio, and models that have a primary purpose of analysing and evaluating material on a service for safety, compliance and quality. These types of models have been shaded in grey in Figure 4.
Note: Figures provided in the graph have been changed to the percentage of models based on the total number of stable generative AI models used by each provider. The total number of models varies per provider. The lengths of the bars are indicative to highlight the difference in number of models used by each provider. Numbers in each row may not add to 100 due to rounding.
Model fine-tuning
Model fine-tuning is when an already trained model is customised to perform specific tasks or behave in certain ways using further specialised training data.
One example of fine-tuning is when a service uses a user’s past conversations (also known as user interaction history) to enhance responses, which in turn can provide a more tailored experience. However, this can also lead a user down an unintended and potentially unsafe and harmful path.18 Fine-tuning a model with all users’ past conversations to ‘improve’ a model raises safety questions, such as whether potentially harmful or unlawful conversations are filtered before being fed into the model.
As already outlined, clean inputs and careful fine-tuning matter for child safety. Conversations between users and AI companions about self-harm or sexually explicit chats which are fed into AI companion models without safeguards create risks for users, especially children. This is also true for other sorts of fine-tuning that use potentially harmful content.
During the reporting period all service providers, except Chub AI, were using individual and collective user interaction histories to fine-tune their models.
Findings of positive action
- Character.AI reported that it took steps to ensure user safety in model fine-tuning in respect of users’ past conversations and that it used classifiers and language analysis tools to detect new CSEA, self-harm and pornography in the fine-tuning training data on all models used to provide its service.
- Chai reported that it used classifiers and language analysis tools to detect and remove harmful content ‘before it enters the training pipeline’ and used metrics such as the ‘Family Friendly Score’ during the fine-tuning process to ‘penalize toxic outputs’.
- Nomi reported using ‘a classifier to detect CSEA and self-harm during the RLHF [Reinforcement Learning from Human Feedback19] process.’
Gaps and concerns
- Chub AI reported fine-tuning 56% of the community fine-tuned models it used to provide its service. However, Chub AI only reported using tools to detect new CSEA material in fine-tuned training data for 11% of these models. Chub did not report using tools to detect self-harm or pornography (including sexually explicit chats with children) in fine-tuned training data for any of the models it used to provide its service.
Proactive safeguards to intercept risky prompts and protect young users
Service providers were asked if they were using tools or techniques, during the reporting period, to detect and/or filter user prompts to prevent the service generating material that could be unlawful or potentially harmful. Tools or techniques should be applied to all models used to provide the service.20
Why were service providers asked about this?
User prompts are the specific instructions, questions or commands or other input provided by a user to an AI system with the intention of it producing a desired response. Users may intentionally try to generate responses or create material they know is in breach of a service’s Terms of Service, or they may inadvertently create responses or material that is unlawful or potentially harmful.
User input controls and tools should detect and/or filter user inputs to prevent prompts that could generate unlawful or potentially harmful material and activity, as well as alert the service to provide support or other safety features where a user may be at risk.
User prompt controls can keep unsafe or age-inappropriate requests from turning into potentially harmful conversations. They provide confidence that these interactions with the AI system are being actively guarded for unlawful or potentially harmful material and activity, providing a backup safety measure if age assurance is not effective and children may be accessing the service.
What we found
Figure 5 shows an indicative view of the proportion of models that service providers used to provide their service and the models where tools or techniques were used to detect and/or filter user prompts to prevent the generation of new CSEA, self-harm and pornography (including sexually explicit chats with children). While the data has been represented as a proportion, the graph still provides an indicative comparison of the number of models used by each provider.
With regards to tools or techniques to detect and/or filter user prompts to prevent the generation of known CSEA, service providers reported the following:
- Character.AI used PhotoDNA to detect known CSEA on image upload on all image models used to provide its service, both the under 18 experience and adult experience.
- Chub used Cloudflare's CSAM detector and its own proprietary tool to detect known CSAM on uploaded images on some of its models.
- Chai stated that users cannot upload images to their chats. Chai also reported that it did not host one of its proprietary models directly, stating that ‘we rely on an external partner…to both host the model and implement safety mechanisms.’ Chai stated that the external partner ‘checked’ inputs.
- Nomi did not use any tools to detect known CSEA in user prompts stating that it did not train its models on images of any kind and, although images can be uploaded, Nomi stated ‘never in a mechanism that makes it possible to replicate the likeness of the image.’
Based on available information and information reported by providers, models not in scope for this question are models that eSafety would not expect tools to be used on, which include models that have a primary purpose of analysing and evaluating material on a service for safety, compliance and quality. These types of models have been shaded in grey in Figure 5.
Findings of positive action
- Character.AI reported using internal classifiers on its under-18 experience to detect and/or filter CSEA, self-harm and sexualised content including pornography in user prompts, as well as other content which violates the terms and conditions for the service.
- Character.AI reported using classifiers and blocklists banning keywords to detect and/or filter CSEA, self-harm and sexualised content including pornography in user prompts across all the other models used to provide its service.
Gaps and concerns
- Chai did not use tools to detect and/or filter new CSEA, self-harm and pornography in user prompts across 22% of the models used to provide its service.
- Chub did not use tools to detect and/or filter user prompts for self-harm or pornography across any of the models it used to provide its service and relied solely on a keyword or phrase blocklist to filter user prompts for new CSEA.
- Nomi reported using classifiers and keyword blocklists to detect and/or filter new CSEA, self-harm and pornography on some of the models used to provide its service. Nomi did not check inputs for new CSEA on 33% of the models or check inputs for self-harm or pornography on 50% of the models used to provide its service, including its proprietary model.
Note: Figures provided in the graph have been changed to the percentage of models based on the total number of stable generative AI models used by each provider. The total number of models varies per provider. The lengths of the bars are indicative to highlight the difference in number of models used by each provider. Numbers in each row may not add to 100 due to rounding.
Mechanisms to detect a negative turn in the conversation
Service providers were asked if they had sentiment analysis mechanisms in place, during the reporting period, to detect a range of potential negative emotional or behavioural cues from a user.
Sentiment analysis is the process of assessing text to determine whether it expresses a positive, negative or neutral sentiment.
Using sentiment analysis can help prevent conversations between users and AI companions from drifting into potentially harmful themes without timely intervention.
Why were service providers asked about this?
There are many approaches to sentiment analysis, and some struggle to detect intent or coded language. However, appropriately fine-tuned language models, such as the Bidirectional Encoder Representations from Transformer (BERT), can interpret context and intent, recognise coded language, and detect when conversations begin to shift toward potentially harmful themes such as self-harm.
Sentiment analysis is more sophisticated than language analysis tools that rely on key word block lists.21
eSafety research found that an AI companion or AI assistant had, at some stage, made most children (aged 10 to 17) in Australia who had used one of these apps feel something positive (85%) (for example, happy, excited, smart, reassured or attractive). However, just under half (47%) of the children who had used them said an AI companion or AI assistant had made them feel something negative (for example upset, worried, confused, angry, annoyed, unsafe, embarrassed or hopeless). Of the children that had a negative feeling provoked by an AI companion or AI assistant, 4% indicated they had told it how they were feeling ‘every time’ it made them feel something negative, 6% told it ‘most of the times’ and 35% told it ‘some of the times’ (eSafety, Forthcoming*).
When these mechanisms detect negative sentiment, they can trigger some sort of warning, limit or stop the chat, and alert Trust and Safety teams if required. If someone seems upset or troubled, the system can offer safer replies and point the user towards support.
What we found
All of the service providers reported not using sentiment analysis mechanisms, but Chai, Character.AI and Nomi did report using other mechanisms to detect negative turns in conversation.
- Chai stated that it ‘purposefully do[es] not use sentiment analysis’ because it believes it is ‘a technically inappropriate tool for fiction storytelling’. Instead, it reported using intent classification which analyses the conversational goal rather than flagging keywords. Chai also reported that when intent to engage in self-harm or generate CSEA was flagged by its prompt classifiers, these messages were flagged as being inappropriate for the chatbot to answer. In these cases a different model was deployed to answer the message in a way that made it clear that the message was not appropriate for the service. Chai stated that this model provided ‘resources and de-escalation for self-harm’ and the model refused engagement and condemned the content in cases of CSEA.
- Character.AI reported that it directs all users to support or help in response to user prompts containing content that matches a specified terms list related to suicide or self-harm. In its under-18 experience, Character reported that it directs users to support or help in response to both inputs and outputs, by detecting harmful inputs from a defined list of terms and a classifier model that runs over model outputs.
- Character.AI also stated that it is ‘working to develop a sentiment analysis mechanism to detect negative emotional cues and intervene in a user’s crisis.’
- Nomi reported that it ‘consider[s] user sentiment detection as a task intrinsically learned as a part of training our LLM [Large Language Model] and find with proper training LLMs are more capable of detecting and responding intuitively and empathetically in situations where a user is distressed.’ Nomi also reported using keyword blocklists on some of its image and video models to detect user prompts for CSEA, self-harm and pornography, as well as a negative keywords list on outputs for CSEA.
Guardrails to prevent potentially harmful content slipping through
Service providers were asked if they were using tools or techniques, during the reporting period, to prevent model outputs that could have resulted in the service generating unlawful or potentially harmful material. Tools or techniques should be applied to all models used to provide the service. 22
Why were service providers asked about this?
Model outputs are the generated outputs from an AI model after processing the user input prompt(s). The outputs can be text, image, video and/or audio, depending on the models used.
Even with cleaned training data and robust user prompt detectors and filters, there is still a risk that unlawful and potentially harmful content may appear in what the model generates. Output guardrails serve as another layer of defence between the model and the user, identifying and blocking potentially harmful or policy-violating responses.
eSafety research found that among children (aged 10 to 17) in Australia who had ever used an AI companion or AI assistant, some had been exposed to inappropriate or potentially harmful content, including having an AI companion or AI assistant:
- share a nude image or nude video (real or AI-generated) with them – 4%
- suggest that it was okay for them to do something that might upset or hurt someone else – 4%
- suggest it was okay for them to do something that might hurt them or be bad for their body – 3%
- ask them to do something that made them feel uncomfortable or unsafe – 3%
- say something offensive, mean or unfair about their gender, sexuality, race, skin colour, religion, disability or the country they are from – 3%.
In addition, among teens (aged 13 to 17) who had ever used an AI companion or AI assistant, 4% indicated that it had chatted or messaged with them about kissing or sex and 3% indicated it had shared a sexual image or sexual video with them (such as real or AI-generated pornography) (eSafety, Forthcoming*).
While strong output controls are essential, they are most effective when combined with the safeguards outlined in earlier sections of this report, such as age assurance, model training and user prompt detection and/or filtering. This comprehensive approach would minimise the risk of potentially harmful content reaching users and provide reassurance that safety checks are active and effective at every stage.
What we found
Figure 6 shows an indicative view – for each text, image and video model a provider used to provide its service – of whether tools were in place to prevent model outputs that could have resulted in the service generating new CSEA, self-harm and pornography (including sexually explicit chats with children). While the data has been represented as a proportion, the graph still provides an indicative comparison of the number of models used by each provider.
Based on available information and information reported by providers, models not in scope for this question are models that eSafety would not expect tools to be used on. These include audio models, which are not generative but merely translate text to audio, and should have been checked at user prompt stage, as well as models that have a primary purpose of analysing and evaluating material on a service for safety, compliance and quality. These types of models have been shaded in grey in Figure 6.
Findings of positive action
- Character.AI stated that it used classifiers to prevent model outputs that could have resulted in new CSEA, self-harm and pornography on all models used to provide its service.
Gaps and concerns
- Chub AI did not use tools to detect potential generation of CSEA, self-harm or pornography on any model outputs. Instead, it relied solely on selecting models it believed were less likely to generate adult content, without using any mechanisms to verify or prevent potentially harmful content from being produced.
- Nomi did not use tools to detect potential generation of new CSEA on model outputs on 50% of the models used to provide its service, or self-harm on 17% of its models, and did not use tools to detect potential generation of pornography on model outputs on 33% of the models used to provide its service, including its main proprietary model.
- Chai reported that, while it did use a reward model to prevent model outputs that could have resulted in new CSEA, self-harm and pornography on all models it hosted, it did not host one of its proprietary models directly and that it relied, ‘on an external partner…to both host the model and implement safety mechanisms.’ Chai stated that the external partner ‘checked’ inputs, ‘resulting in harmful outputs being blocked at the infrastructure level,’ but not model outputs.
Note: Figures provided in the graph have been changed to the percentage of models based on the total number of stable generative AI models used by each provider. The total number of models varies per provider. The lengths of the bars are indicative to highlight the difference in number of models used by each provider. Numbers in each row may not add to 100 due to rounding.
Conclusion: Lifting the standard
Publishing transparency reports under the Basic Online Safety Expectations scheme fulfills one of the objectives of the Online Safety Act, to increase the transparency and accountability of online services. It is a powerful regulatory lever, enabling eSafety to shine a light on services’ practices.
Insights from these transparency processes inform the work of eSafety, including our role to promote compliance with the Online Safety Codes and Standards and enforce them..
Service providers have requirements under the Online Safety Codes and Standards for both unlawful material such as child sexual exploitation material (Unlawful Material Codes and Standards) and age-restricted material such as online pornography and self-harm material (Age-Restricted Material Codes). These enforceable measures apply to many different types of online services, including AI companions and other generative AI services.
With the final Age-Restricted Material Codes in effect from 9 March 2026, eSafety is now monitoring and enforcing compliance, drawing on a broad range of insights, including those in this report as well as insights from our recent research with young people.
Requirements in the Online Safety Codes and Standards include:
- designing for safety to prevent unlawful and potentially harmful content, including child sexual exploitation and abuse material
- activating age assurance on highest-risk features so only adults can access functions that could generate potentially harmful age-inappropriate material, such as online pornography
- providing extra protections on medium-risk services without age assurance to prevent children generating potentially harmful age-inappropriate material
- ensuring quick and consistent moderation and escalation, including working with police when necessary
- providing clear ways for users to report issues and seek help, plus regular risk assessments and fixes
- conducting regular testing and reviews, with documented results and improvements over time.
eSafety will use the full range of our powers to ensure compliance and deter non-compliance. This can include seeking penalties through civil proceedings in the Federal Court of up to A$49.5 million per breach.
About this report
This report provides a summary of the information eSafety received in response to the notices given to four generative AI companion service providers.
Scope: These summaries offer a summary of the service providers’ responses, rather than presenting their responses in full.
Withheld information: In line with eSafety’s regulatory guidance, some information has been withheld. This is to protect sensitive details such as information that could impact public safety.
Verification: eSafety did not conduct technical tests or independent research to verify claims made by the service providers. Service providers are expected to respond truthfully and accurately. This report is published to promote transparency and accountability.
Time period: The information presented here covers the reporting period between 1 July 2025 and 30 September 2025 (inclusive), unless stated otherwise.
Geography: All data in this report is global, unless stated otherwise.
Defined terms: Terms are defined in the definitions, unless stated otherwise.
References
* This report is currently in preparation and is expected to be released in May 2026.
1 Family Online Safety Institute (2025). Generative AI in Uncertain Times: How Teens are Navigating a New Digital Frontier. FOSI website, accessed 15 December 2025.
2 Grossman, S., Pfefferkorn, R., Liu, S. (2025). AI-Generated Child Sexual Abuse Material: Insights from Educators, Platforms, Law Enforcement, Legislators, and Victims. Stanford Digital Repository website, accessed 2 March 2026.
3 Bunting, C, and Huggins, R. (2025). Me, Myself & AI: Understanding and Safeguarding Children’s Use of AI Chatbots. Internet Matters website, accessed 3 March 2026;
Common Sense Media (2025). AI Risk Assessment: Social AI Companions. Common Sense Media website, accessed 3 March 2026;
VoiceBox (2023). Coded Companions: Young People’s Relationships with AI Chatbots. Voicebox website, accessed 3 March 2026;
Bernardi. J. (2025). Friends for sale: the rise and risks of AI companions. Ada Lovelace Institute website, accessed 3 March 2026.
4 American Psychological Association (2025). Health Advisory: Artificial Intelligence and Adolescent Well-being. American Psychological Association website, accessed 3 March 2026;
Common Sense Media (2025). AI Risk Assessment: Social AI Companions. Common Sense Media website, accessed 3 March 2026.
5 International AI Safety Report 2025 (2025). International AI Safety Report website, accessed 26 February 2026.
6 National Center for Missing and Exploited Children (n.d.). CyberTipline Report. NCMEC CyberTipline website.
7 NCMEC’s CyberTipline is the USA's centralised reporting system for the online exploitation of children. In 2024, the CyberTipline received 67,000 reports involving generative AI, a 1,325% increase since the previous year. URL: missingkids.org/gethelpnow/cybertipline/cybertiplinedata.
8 Nomi.AI (2026). ‘Terms of Service’. Nomi.AI website.
9 Chub AI (2026). ‘Terms of Service’. (Chub.AI website is no longer available in Australia).
10 National Center for Missing and Exploited Children (n.d.). CyberTipline Report. NCMEC CyberTipline website.
11 Chub AI (2026). ‘Terms of Service’. (Chub.AI website is no longer available in Australia).
12 Nomi.AI (2026). ‘Terms of Service’. Nomi.AI website.
13 Business of Apps (2025). character.ai Revenue and Usage Statistics (2025). Business of Apps website, accessed 19 February 2026.
14 eSafety Commissioner (2025). Behind the screen: Transparency report and key findings, page 22. eSafety website, accessed 16 September 2025.
15 eSafety Commissioner (2025). Online Safety Codes and Standards regulatory guidance. eSafety website, accessed 23 February 2026.
16 Apple (2026). Age requirements for apps distributed in Brazil, Australia, Singapore, Utah, and Louisiana. Apple Developer website, accessed 3 March 2026.
17 eSafety asked how many stable generative AI models were used to provide the service. A stable generative AI model is an AI model that has been deployed into production as the primary version intended for use in the service, not in alpha or beta testing.
18 Kirk, H. R., Vidgen, B., Röttger, P., & Hale, S. A. (2024). The benefits, risks and bounds of personalizing the alignment of large language models to individuals. Nature Machine Intelligence, 6(4), 383–392. doi: https://doi.org/10.1038/s42256-024-00820-y, accessed 7 January 2026.
19 RLHF is a machine learning technique used to train AI models, especially large language models, by combining supervised learning, where humans provide high quality example responses, and reinforcement learning, where humans rate or compare model outputs so the model learns which behaviours are preferred.
20 This includes proprietary, community fine-tuned, commercial fine-tuned and commercial baseline models.
21 Elsafoury, F., Katsigiannis, S., Pervez, Z., & Ramzan, N. (2021). When the Timeline Meets the Pipeline: A Survey on Automated Cyberbullying Detection. IEEE Access, 9, 103541–103563. doi: https://doi.org/10.1109/access.2021.3098979, accessed 15 December 2025.
22 This includes proprietary, community fine-tuned, commercial fine-tuned and commercial baseline models.
Last updated: 23/03/2026