<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.4 20221130//EN"
  "JATS-journalpublishing1-4.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink"
         xmlns:mml="http://www.w3.org/1998/Math/MathML"
         xmlns:ali="http://www.niso.org/schemas/ali/1.0/"
         xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
         article-type="research-article"
         dtd-version="1.4"
         xsi:noNamespaceSchemaLocation="https://jats.nlm.nih.gov/archiving/1.4/xsd/JATS-archivearticle1-4-mathml3.xsd"
         xml:lang="en">

  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">CACTUS</journal-id>
      <journal-id journal-id-type="nlm-ta">CACTUS J Tour Bus Manag Econ</journal-id>

      <journal-title-group>
        <journal-title>CACTUS – Journal of Tourism Business, Management and Economics</journal-title>
      </journal-title-group>

      <issn pub-type="epub">2247-3297</issn>

      <publisher>
        <publisher-name>ASE Publishing House</publisher-name>
        <publisher-loc>Bucharest, Romania</publisher-loc>
      </publisher>
    </journal-meta>

    <article-meta>
      <article-id pub-id-type="doi">10.24818/CTS/7/2025/2.11</article-id>
      <article-id pub-id-type="publisher-id">cactus-2025-21</article-id>

      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Research Article</subject>
        </subj-group>
        <subj-group subj-group-type="jel">
          <subject>M11, O22</subject>
        </subj-group>
      </article-categories>

      <title-group>
        <article-title>Turbocharged: automating quality analysis in trust &amp; safety</article-title>
      </title-group>

      <contrib-group>
        
        <contrib contrib-type="author" corresp="yes">
          <name>
            <surname>Tîrnăcop</surname>
            <given-names>Alexandra Bianca</given-names>
          </name>
          <contrib-id contrib-id-type="orcid">https://orcid.org/0009-0004-2945-0643</contrib-id>
          <xref ref-type="aff" rid="aff1_1"/>
          <email>tirnacopalexandra21@stud.ase.ro</email>
        </contrib>
      </contrib-group>

      
      <aff id="aff1_1">
        <label>1</label>
        <institution-wrap>
          <institution>Bucharest University of Economic Studies, Bucharest, Romania</institution>
        </institution-wrap>
      </aff>

      <pub-date pub-type="epub">
        <day>15</day>
        <month>12</month>
        <year>2025</year>
      </pub-date>

      <volume>7</volume>
      <issue>2</issue>
      
      

      <history>
        <date date-type="received"><day>15</day><month>12</month><year>2025</year></date>
        
        
      </history>

      <permissions>
        <copyright-statement>Copyright © 2025, The Author(s)</copyright-statement>
        <copyright-year>2025</copyright-year>
        <copyright-holder>The Author(s)</copyright-holder>
        <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
          <license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0).</license-p>
        </license>
        <ali:free_to_read/>
        <ali:license_ref>https://creativecommons.org/licenses/by-nc/4.0/</ali:license_ref>
      </permissions>

      <abstract xml:lang="en"><p>Trust and Safety (T&amp;S) is a key framework for online platforms, aiming to protect users from harm such as misinformation, harassment, and exploitation, while also supporting free expression. Although policies, AI tools, and cross-platform collaboration (e.g., GIFCT, StopNCII.org) enhance moderation, significant challenges remain. This study uses a demo dataset of 15 social media posts, reviewed by 9 moderators and checked by a single analyst. Each ticket has been reviewed by three raters to ensure agreement. The model achieved a precision, recall, and F1 score of 70.37%, with an overall accuracy of 64.44%. Automation improves efficiency but requires bias moderation, transparency, and human intervention to address challenging content. However, outsourcing and underinvestment in moderators raise ethical concerns, as human reviewers face psychological risks without adequate support. To address these issues, this paper proposes a decision matrix for use in both machine learning training and moderator and quality analyst training.</p></abstract>

      <kwd-group xml:lang="en">
        <title>Keywords</title>
        <kwd>artificial intelligence; key performance indicators; machine learning</kwd>
      </kwd-group>

      <self-uri content-type="text/html" xlink:href="http://0.0.0.0:3000/articles/cactus-issue-2-vol-7-no-2-2025-art-11"/>
      <self-uri content-type="application/pdf" xlink:href="http://0.0.0.0:3000/wp-content/uploads/pdfs/vol7_no2_2025_art12_III-Tirnacop.pdf"/>
      <self-uri content-type="application/xml" xlink:href="http://0.0.0.0:3000/api/articles/cactus-issue-2-vol-7-no-2-2025-art-11/jats"/>

      <related-article ext-link-type="uri" related-article-type="in-issue" xlink:href="http://0.0.0.0:3000/issues/cactus-issue-2-vol-7-no-2-2025" id="ra1"/>

    </article-meta>
  </front>

  <body>
    <p>Full text available in PDF format.</p>
  </body>

  <back>
    
    <ref-list>
      <title>References</title>
      
      <ref id="ref1">
        <element-citation publication-type="other">
          <comment>Ahmed, A., &amp;amp; Khan, M. N. (2024). AI and content moderation: Legal and ethical approaches to protecting free speech and privacy [Manuscript]. ResearchGate. https://www.researchgate.net/publication/383661951_AI_and_Content_Moderation_Legal_and_Ethical_Approaches_to_Protecting_Free_Speech_and_Privacy</comment>
        </element-citation>
      </ref>

      <ref id="ref2">
        <element-citation publication-type="other">
          <comment>Business &amp;amp; Human Rights Resource Centre. (2021). Santa Clara Principles present standards for tech platforms to provide transparency and accountability in content moderation. https://www.business-humanrights.org/en/latest-news/the-santa-clara-principles-on-transparency-and-accountability-in-content-moderation/</comment>
        </element-citation>
      </ref>

      <ref id="ref3">
        <element-citation publication-type="other">
          <comment>Cyberhaven. (2015). What are false positives? https://www.cyberhaven.com/infosec-essentials/what-are-false-positives</comment>
        </element-citation>
      </ref>

      <ref id="ref4">
        <element-citation publication-type="other">
          <comment>Digital Trust &amp;amp; Safety Partnership. (2024a). Trust &amp;amp; safety best practices framework [PDF]. https://dtspartnership.org/wp-content/uploads/2021/04/DTSP_Best_Practices.pdf</comment>
        </element-citation>
      </ref>

      <ref id="ref5">
        <element-citation publication-type="other">
          <comment>Digital Trust &amp;amp; Safety Partnership. (2024b). Best practices for AI and automation in trust &amp;amp; safety [PDF]. https://dtspartnership.org/wp-content/uploads/2024/09/DTSP_Best-Practices-for-AI-Automation-in-Trust-Safety.pdf</comment>
        </element-citation>
      </ref>

      <ref id="ref6">
        <element-citation publication-type="other">
          <comment>Eissfeldt, J., &amp;amp; Mukherjee, S. (2023). Evaluating the forces shaping the trust &amp;amp; safety industry. Tech Policy Press. https://www.techpolicy.press/evaluating-the-forces-shaping-the-trust-safety-industry/</comment>
        </element-citation>
      </ref>

      <ref id="ref7">
        <element-citation publication-type="other">
          <comment>Global Internet Forum to Counter Terrorism. (2022). HSDB taxonomy – For publication (Dec 2022) [PDF]. https://gifct.org/wp-content/uploads/2022/12/HSDB-Taxonomy-FOR-PUBLICATION-Dec-2022-1.pdf</comment>
        </element-citation>
      </ref>

      <ref id="ref8">
        <element-citation publication-type="other">
          <comment>Global Internet Forum to Counter Terrorism. (2024). GIFCT’s hash-sharing database. https://gifct.org/hsdb/</comment>
        </element-citation>
      </ref>

      <ref id="ref9">
        <element-citation publication-type="other">
          <comment>Google for Developers. (2025). Classification: Accuracy, recall, precision, and related metrics. https://developers.google.com/machine-learning/crashcourse/classification/accuracy-precision-recall</comment>
        </element-citation>
      </ref>

      <ref id="ref10">
        <element-citation publication-type="other">
          <comment>Habibi, M., Hovy, D., &amp;amp; Schwartz, C. (2025). The content moderator’s dilemma: Removal of toxic content and distortions to online discourse (arXiv:2412.16114). arXiv. https://doi.org/10.48550/arXiv.2412.16114</comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2412.16114</pub-id>
        </element-citation>
      </ref>

      <ref id="ref11">
        <element-citation publication-type="other">
          <comment>Horatio. (2025). What is content moderation? Pros, cons, and best practices. https://www.hirehoratio.com/blog/what-is-content-moderation</comment>
        </element-citation>
      </ref>

      <ref id="ref12">
        <element-citation publication-type="other">
          <comment>Institute for Human Rights and Business. (2025). Content moderation is a new factory floor of exploitation – Labour protections must catch up. https://www.ihrb.org/latest/content-moderation-is-a-new-factory-floor-of-exploitation-labour-protections-must-catch-up</comment>
        </element-citation>
      </ref>

      <ref id="ref13">
        <element-citation publication-type="other">
          <comment>INTERPOL. (2024a). Crimes against children. https://www.interpol.int/en/Crimes/Crimes-against-children</comment>
        </element-citation>
      </ref>

      <ref id="ref14">
        <element-citation publication-type="other">
          <comment>INTERPOL. (2024b). International child sexual exploitation database. https://www.interpol.int/en/Crimes/Crimes-against-children/International-Child-Sexual-Exploitation-database</comment>
        </element-citation>
      </ref>

      <ref id="ref15">
        <element-citation publication-type="other">
          <comment>Juba, B., &amp;amp; Le, H. S. (2019). Precision-recall versus accuracy and the role of large data sets. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 4039–4048. https://doi.org/10.1609/aaai.v33i01.33014039</comment>
          <pub-id pub-id-type="doi">10.1609/aaai.v33i01.33014039</pub-id>
        </element-citation>
      </ref>

      <ref id="ref16">
        <element-citation publication-type="other">
          <comment>Listen Data. (2024). How to calculate confusion matrix in Excel. https://www.listendata.com/2024/06/confusion-matrix-in-excel.html</comment>
        </element-citation>
      </ref>

      <ref id="ref17">
        <element-citation publication-type="other">
          <comment>Microsoft. (2025). AND function. https://support.microsoft.com/en-us/office/and-function-5f19b2e8-e1df-4408-897a-ce285a19e9d9</comment>
        </element-citation>
      </ref>

      <ref id="ref18">
        <element-citation publication-type="other">
          <comment>Mollas, I., Chrysopoulou, Z., Karlos, S., &amp;amp; Tsoumakas, G. (2021). Ethos: An online hate speech detection dataset (arXiv:2006.08328). arXiv. https://arxiv.org/pdf/2006.08328</comment>
        </element-citation>
      </ref>

      <ref id="ref19">
        <element-citation publication-type="other">
          <comment>Oversight Board. (2025). Content moderation in a new era for AI and automation. https://www.oversightboard.com/news/content-moderation-in-a-new-era-for-ai-and-automation/</comment>
        </element-citation>
      </ref>

      <ref id="ref20">
        <element-citation publication-type="other">
          <comment>Reelmind. (2025). Ametures gone wild: AI content moderation challenges. https://reelmind.ai/blog/ametures-gone-wild-ai-content-moderation-challenges</comment>
        </element-citation>
      </ref>

      <ref id="ref21">
        <element-citation publication-type="other">
          <comment>Ricknell, E. (2020). Freedom of expression and alternatives for internet governance: Prospects and pitfalls. Media and Communication, 8(4), 110–120. https://doi.org/10.17645/mac.v8i4.3299</comment>
          <pub-id pub-id-type="doi">10.17645/mac.v8i4.3299</pub-id>
        </element-citation>
      </ref>

      <ref id="ref22">
        <element-citation publication-type="other">
          <comment>Santa Clara Principles. (2021a). SCP 2.0 toolkit for companies. https://santaclaraprinciples.org/toolkit-companies/</comment>
        </element-citation>
      </ref>

      <ref id="ref23">
        <element-citation publication-type="other">
          <comment>Santa Clara Principles. (2021b). Santa Clara Principles 2.0 open consultation report. https://santaclaraprinciples.org/open-consultation/</comment>
        </element-citation>
      </ref>

      <ref id="ref24">
        <element-citation publication-type="other">
          <comment>Shulruff, T. (2024). Trust and safety work: Internal governance of technology risks and harms. Journal of Integrated Global STEM, 1(2), 95–105. https://doi.org/10.1515/jigs-2024-0003</comment>
          <pub-id pub-id-type="doi">10.1515/jigs-2024-0003</pub-id>
        </element-citation>
      </ref>

      <ref id="ref25">
        <element-citation publication-type="other">
          <comment>Shweta, R. C., Bajpai, R. C., &amp;amp; Chaturvedi, H. K. (2015). Evaluation of inter-rater agreement and inter-rater reliability for observational data: An overview of concepts and methods. Journal of the Indian Academy of Applied Psychology, 41(3), 20–27.</comment>
        </element-citation>
      </ref>

      <ref id="ref26">
        <element-citation publication-type="other">
          <comment>Siapera, E. (2021). AI content moderation, racism and (de)coloniality. International Journal of Bullying Prevention, 4, 55–65. https://doi.org/10.1007/s42380-021-00105-7</comment>
          <pub-id pub-id-type="doi">10.1007/s42380-021-00105-7</pub-id>
        </element-citation>
      </ref>

      <ref id="ref27">
        <element-citation publication-type="other">
          <comment>StopNCII.org. (2025). How StopNCII.org works. https://stopncii.org/chi-siamo/</comment>
        </element-citation>
      </ref>

      <ref id="ref28">
        <element-citation publication-type="other">
          <comment>Tremau. (2025). Content moderation: Key practices &amp;amp; challenges. https://tremau.com/resources/content-moderation-key-practices-challenges/</comment>
        </element-citation>
      </ref>

      <ref id="ref29">
        <element-citation publication-type="other">
          <comment>TSPA. (2025). Content moderation quality assurance. https://www.tspa.org/curriculum/ts-fundamentals/content-moderation-and-operations/content-moderation-quality-assurance/</comment>
        </element-citation>
      </ref>

      <ref id="ref30">
        <element-citation publication-type="other">
          <comment>Vargas Penagos, E. (2025). Platforms on the hook? EU and human rights requirements for human involvement in content moderation. Cambridge Forum on AI: Law and Governance, 1, e23. https://doi.org/10.1017/cfl.2025.3</comment>
          <pub-id pub-id-type="doi">10.1017/cfl.2025.3</pub-id>
        </element-citation>
      </ref>

      <ref id="ref31">
        <element-citation publication-type="other">
          <comment>Walker, A. R. (2025, April 11). Legal Defense Fund exits Meta civil rights advisory group over DEI changes. The Guardian. https://www.theguardian.com/technology/2025/apr/11/meta-ldf-dei-policy</comment>
        </element-citation>
      </ref>

      <ref id="ref32">
        <element-citation publication-type="other">
          <comment>Weigl, L., &amp;amp; Bodó, B. (2025). Trust and safety in the age of AI – The economics and practice of the platform-based discourse apparatus (Amsterdam Law School Legal Studies &amp;amp; Institute for Information Law Research Paper No. 2025-1). SSRN. https://doi.org/10.2139/ssrn.5116478</comment>
          <pub-id pub-id-type="doi">10.2139/ssrn.5116478</pub-id>
        </element-citation>
      </ref>

      <ref id="ref33">
        <element-citation publication-type="other">
          <comment>Woods, J. (2022). Bias in AI program: Showing businesses how to reduce bias and mitigate risk. Vector Institute. https://vectorinstitute.ai/bias-in-ai-program-showing-businesses-how-to-reduce-bias-and-mitigate-risk/</comment>
        </element-citation>
      </ref>

      <ref id="ref34">
        <element-citation publication-type="other">
          <comment>Zeng, J., &amp;amp; Kaye, D. B. V. (2022). From content moderation to visibility moderation: A case study of platform governance on TikTok. Policy &amp;amp; Internet, 14, 79–95. https://doi.org/10.1002/poi3.287</comment>
          <pub-id pub-id-type="doi">10.1002/poi3.287</pub-id>
        </element-citation>
      </ref>
    </ref-list>
  </back>

</article>