TRANSFORMING THE LEGAL LANDSCAPE: AN AI-DRIVEN FRAMEWORK FOR JUDICIAL TEXT PROCESSING
Resumo
Artificial Intelligence can revolutionize the legal field by addressing the complexities of managing extensive textual data inherent to judicial processes. Nevertheless, the literature highlights the difficulties in managing different contexts regarding distinct application scenarios. This paper presents a novel methodology tailored for developing applications in the legal domain, leveraging cutting-edge natural language processing techniques, including transformer-based architectures, pre-trained models, and transfer learning. Unlike traditional software development, this approach embraces the inherent uncertainties of Artificial Intelligence solutions by employing an iterative framework that integrates strong collaboration with legal professionals, domain-specific datasets, and comprehensive evaluation strategies. The methodology was validated through real-world applications at the Court of Justice of Rio Grande do Sul, including the development of a Judgment Report Generator, which automates judgment report creation using Generative Artificial Intelligence, and additional experiments showcased state-of-the-art performance in legal Named Entity Recognition using fine-tuned BERT models and context-adapted text generation with GPT-2-based models, demonstrating adaptability to diverse legal scenarios. This work bridges advanced natural language processing techniques with the practical demands of the judiciary, establishing a foundation for scalable, reliable, and domain-aware AI applications. The proposed methodology addresses practical challenges, regulatory alignment, and dataset specificity, enabling effective AI integration in the legal sector for enhanced efficiency and impact in real-world judicial systems.
Biografia do Autor
PhD student in Artificial Intelligence applied to Law at UNISINOS. MsC (2009) and Bsc (2000) in Applied Computing at UNISINOS and specialization at UFRGS (2004). Systems analyst at the Rio Grande do Sul State Court of Justice and professor at UNISENAC-RS.
Bsc in Computer Science at PUCRS (1990); MsC (1993) and PhD (2008) in Computer Science at UFRGS (2008). Post-doctorate at Friedrich-Alexander Universität Erlangen-Nürnberg/Germany (2018). Professor at UNISINOS; Researcher in UNISINOS/PPGCA. Dean of the UNISINOS Polytechnic School.
Referências
ADADI, A.; BERRADA, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access, v. 6, p. 52138–52160, 2018. DOI: https://doi.org/10.1109/ACCESS.2018.2870052
BAI, Y. et al. Benchmarking Foundation Models with Language-Model-as-an-Examiner. Advances in Neural Information Processing Systems, v. 36, p. 78142–78167, 15 dez. 2023.
BOMMASANI, R. et al. On the Opportunities and Risks of Foundation Models. arXiv, 12 jul. 2022. Available at: https://doi.org/10.48550/arXiv.2108.07258 Available at: 26 out. 2024
CARMO, D. et al. PTT5: Pretraining and validating the T5 model on Brazilian Portuguese data. arXiv, 8 out. 2020.
CLARK, P. et al. Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge. arXiv, 14 mar. 2018. Available at: https://www.doi.org/10.48550/arXiv.1803.05457 Available at: 24 out. 2024
DEVLIN, J. et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 […] Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). NAACL-HLT 2019. Minneapolis, Minnesota: Association for Computational Linguistics, jun. 2019. Available at: http://doi.org/10.18653/v1/N19-1423 Available at: 6 ago. 2021 DOI: https://doi.org/10.18653/v1/N19-1423
DICE, D.; KOGAN, A. Optimizing Inference Performance of Transformers on CPUs. arXiv, 22 fev. 2021.
GAO, Y. et al. Retrieval-Augmented Generation for Large Language Models: A Survey. arXiv, 27 mar. 2024. Available at: https://doi.org/10.48550/arXiv.2312.10997 Available at: 27 out. 2024
HENDRYCKS, D. et al. Measuring Massive Multitask Language Understanding. arXiv, 12 jan. 2021. Available at: http://doi.org/10.48550/arXiv.2009.03300 Available at: 24 out. 2024
HSU, Y.-T. et al. Efficient Inference For Neural Machine Translation. Proceedings of SustaiNLP: Workshop on Simple and Efficient Natural Language Processing EMNLP-SUSTAINLP 2020. Online: Association for Computational Linguistics, nov. 2020. Available at: https://www.doi.org/10.18653/v1/2020.sustainlp-1.7 Available at: 20 fev. 2022 DOI: https://doi.org/10.18653/v1/2020.sustainlp-1.7
JIANG, A. Q. et al. Mixtral of Experts. arXiv, 8 jan. 2024. Available at: https://doi.org/10.48550/arXiv.2401.04088 Available at: 26 out. 2024
LIN, C.-Y. ROUGE: A Package for Automatic Evaluation of Summaries. Text Summarization Branches Out. Anais [...] Barcelona, Spain: Association for Computational Linguistics, jul. 2004. Available at: https://aclanthology.org/W04-1013 Available at: 25 jun. 2024
LUZ DE ARAUJO, P. H. et al. LeNER-Br: A Dataset for Named Entity Recognition in Brazilian Legal Text. In: VILLAVICENCIO, A. et al. (Eds.). Computational Processing of the Portuguese Language. Lecture Notes in Computer Science. Cham: Springer International Publishing, 2018. v. 11122. p. 313–323. DOI: https://doi.org/10.1007/978-3-319-99722-3_32
MUNKHDALAI, T.; FARUQUI, M.; GOPAL, S. Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention. arXiv, 9 ago. 2024. Available at: https://doi.org/10.48550/arXiv.2404.07143 Available at: 27 out. 2024
NOY, N. F.; MCGUINNESS, D. L. Ontology Development 101: A Guide to Creating Your First Ontology. [S. l.] Stanford Knowledge Systems Laboratory, 2001.
OPENAI et al. GPT-4 Technical Report. arXiv, 4 mar. 2024. Available at: https://doi.org/10.48550/arXiv.2303.08774 Available at: 4 ago. 2024
ÖZÇIFT, A. et al. Advancing natural language processing (NLP) applications of morphologically rich languages with bidirectional encoder representations from transformers (BERT): an empirical case study for Turkish. Automatika, v. 62, n. 2, p. 226–238, 3 abr. 2021. DOI: https://doi.org/10.1080/00051144.2021.1922150
PAN, S. J.; YANG, Q. A Survey on Transfer Learning. IEEE Transactions on Knowledge and Data Engineering, v. 22, n. 10, p. 1345–1359, out. 2010. DOI: https://doi.org/10.1109/TKDE.2009.191
PAPINENI, K. et al. BLEU: a method for automatic evaluation of machine translation. Proceedings of the 40th Annual Meeting on Association for Computational Linguistics. ACL’ 02.USA: Association for Computational Linguistics, 6 jul. 2002. Available at: https://dl.acm.org/doi/10.3115/1073083.1073135 Access at: 4 ago. 2024 DOI: https://doi.org/10.3115/1073083.1073135
PUSHP, P. K.; SRIVASTAVA, M. M. Train Once, Test Anywhere: Zero-Shot Learning for Text Classification. arXiv, 23 dez. 2017.
QIN, L. et al. Multilingual Large Language Model: A Survey of Resources, Taxonomy and Frontiers. arXiv, 7 abr. 2024. Available at: https://doi.org/10.48550/arXiv.2404.04925 Access at: 26 out. 2024
RADFORD, A. et al. Improving language understanding by generative pre-training. OpenAI blog, 2018.
RADFORD, A. et al. Language models are unsupervised multitask learners. OpenAI blog, v. 1, n. 8, p. 9, 2019.
RAFFEL, C. et al. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683, 2019.
SOUZA, F.; NOGUEIRA, R.; LOTUFO, R. BERTimbau: Pretrained BERT Models for Brazilian Portuguese. Intelligent Systems. Anais [...] Lecture Notes in Computer Science.Cham: Springer International Publishing, 2020. Available at: http://doi.org/10.1007/978-3-030-61377-8_28 DOI: https://doi.org/10.1007/978-3-030-61377-8_28
TOUVRON, H. et al. LLaMA: Open and Efficient Foundation Language Models. arXiv, 27 fev. 2023. Available at: https://doi.org/10.48550/arXiv.2302.13971 Access at: 29 mar. 2024
VASWANI, A. et al. Attention is all you need. Proceedings of the 31st International Conference on Neural Information Processing Systems. NIPS’17.Red Hook, NY, USA: Curran Associates Inc., 4 dez. 2017. Available at: https://dl.acm.org/doi/10.5555/3295222.3295349 Access at: 6 ago. 2021
WAGNER FILHO, J. A. et al. The brWaC Corpus: A New Open Resource for Brazilian Portuguese. Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). Miyazaki, Japan: European Language Resources Association (ELRA), maio 2018. Available at: https://aclanthology.org/L18-1686 Access at: 17 ago. 2021
WANG, Y. et al. Generalizing from a Few Examples: A Survey on Few-shot Learning. ACM Computing Surveys, v. 53, n. 3, p. 63:1-63:34, 12 jun. 2020. DOI: https://doi.org/10.1145/3386252
WILLIAN SOUSA, A.; FABRO, M. Iudicium Textum Dataset Uma Base de Textos Jurídicos para NLP. 2nd Dataset Showcase Workshop at SBBD (Brazilian Symposium on Databases). Anais [...] Em: SBBD 2019. Fortaleza, Brazil: 7 out. 2019.
XUE, L. et al. mT5: A massively multilingual pre-trained text-to-text transformer. arXiv, 11 mar. 2021. Available at: https://doi.org/10.48550/arXiv.2010.11934 Access at: 26 out. 2024 DOI: https://doi.org/10.18653/v1/2021.naacl-main.41
ZANUZ, L.; RIGO, S. J. Fostering Judiciary Applications with New Fine-Tuned Models for Legal Named Entity Recognition in Portuguese. (V. Pinheiro et al., Eds.)Computational Processing of the Portuguese Language. Cham: Springer International Publishing, 2022. Available at: https://doi.org/10.1007/978-3-030-98305-5_21 DOI: https://doi.org/10.1007/978-3-030-98305-5_21
ZELLERS, R. et al. HellaSwag: Can a Machine Really Finish Your Sentence?. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. ACL 2019. Florence, Italy: Association for Computational Linguistics, jul. 2019. Available at: http://doi.org/10.18653/v1/P19-1472 Access at: 24 out. 2024 DOI: https://doi.org/10.18653/v1/P19-1472
ZHENG, L. et al. Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena. Advances in Neural Information Processing Systems, v. 36, p. 46595–46623, 15 dez. 2023.
