Integrating RAG for Smarter Animal Certification Platforms
Abstract
1. Introduction
2. Background
2.1. Artificial Intelligence (AI)
2.2. Natural Language Processing and Language Models
2.3. Large Language Models
2.4. Specialized Domains
- Knowledge Gaps and Hallucinations: LLMs are trained on vast but general corpora like Common Crawl, which disproportionately represent conversational language over niche, technical literature. As a result, the model’s understanding of domain-specific terminology, concepts, and reasoning patterns is often shallow or incomplete. This “knowledge gap” can lead to factual inaccuracies or, more dangerously, “hallucinations,” where the model generates plausible-sounding but entirely fabricated information to fill the void in its knowledge [2,22]. For instance, it might invent a non-existent legal precedent or incorrectly describe a technical standard with high confidence.
- Lack of Domain-Specific Nuance and Reasoning: Specialized fields are characterized by complex rules, subtle exceptions, and context-dependent interpretations that are difficult to learn from a general corpus. A legal term like “liability” has a precise, multi-part definition that differs starkly from its colloquial usage. Similarly, a medical diagnosis relies on structured, causal reasoning, a known weakness of LLMs, which excel at identifying statistical correlations but not at understanding underlying mechanisms. This inability to grasp deep semantic nuance means models can easily misinterpret a query or provide a technically correct but contextually inappropriate response [23,24].
- Static Knowledge and Information Obsolescence: The knowledge embedded within a pre-trained LLM is static, frozen at the time of its last training run. This makes it inherently unreliable for dynamic domains where information evolves rapidly. Fields like regulatory compliance, medical research, and financial markets require access to real-time updates. An LLM trained in early 2024 would be oblivious to new legislation passed later that year, making its advice not just outdated but potentially illegal [1]. This “knowledge cutoff” is a critical vulnerability that prevents the model from being a trustworthy source for current information.
- High Cost of Errors and Lack of Verifiability: In consumer chatbot applications, an error is often a trivial inconvenience. In specialized domains, the consequences can be severe; a misinterpreted compliance rule can lead to financial penalties, a flawed engineering specification can cause catastrophic failure, and an incorrect medical suggestion can endanger lives. These high-stakes environments demand an exceptional level of accuracy and, crucially, verifiability—the ability for a user to trace an answer back to its source document. Standard LLMs act as “black boxes,” making it impossible to verify the origin or factual basis of their generated output.
2.5. Retrieval-Augmented Generation (RAG) Architecture
- 1: The user sends a query—for example, a legal question or a request for information about a specific topic.
- 2: The query is transformed into an embedding and matched against a vector database containing pre-encoded chunks of documents.
- 3: The retriever selects the most relevant chunks based on semantic similarity.
- 4: These chunks are passed along with the query to the language model, which generates a response based on both the query and the retrieved content.
2.6. Fine-Tuning vs. Retrieval-Augmented Generation (RAG)
- Adaptability: Retrieval enables models to respond to domain-specific queries with real-time context, making it better suited for fields with dynamic knowledge requirements.
- Scalability: By offloading knowledge storage to external databases, retrieval reduces the need for model scaling, unlike fine-tuning which often requires larger model sizes to capture domain intricacies.
- Benchmarking Potential: Retrieval serves as an ongoing benchmark by continuously updating its knowledge base, allowing real-world validation of LLM performance in specialized domains.
2.7. PDSA-RS and Animal Health Regulations in Brazil
3. Materials and Methods
3.1. RAG Module
3.2. Large Language Model Setup
3.3. API and System Integration
3.4. Architecture Overview
4. Results
4.1. Data Curation and Knowledge Base Construction
4.2. Query Structure and Retrieval Process
- Context:
- Retrieved Chunks from Legal Docs and Certification Report
- Question:
- Original User Query
- Instructions:
- Based only on the context provided above, answer the question.
4.3. LLM Configuration
- Temperature: Set to 0.4.
- -
- Lower temperatures are generally preferred in tasks that require factual consistency and reduced variability.
- -
- We experimented with values of 0.2, 0.4, and 0.6.
- -
- At 0.2, responses were precise but overly rigid and often lacked fluency.
- -
- At 0.6, output became more natural but occasionally introduced minor inaccuracies or less formal wording.
- -
- Temperature 0.4 provided a good trade-off, generating outputs that were both accurate and clear while preserving the formal tone required in veterinary reports.
- Context Window Size: Limited to 1024 tokens.
- -
- Although relatively small, this limit was suitable due to the focused and pre-processed nature of the retrieved content.
- -
- A smaller context helped reduce inference latency during local execution.
- Beam Search Width: Set to 16.
- -
- Greedy decoding (width 1) often led to less complete answers.
- -
- Width 8 improved quality slightly, but width 16 consistently yielded better results for structured outputs such as certificate summaries.
- Hardware Adaptation:
- -
- We ran the system on a machine equipped with a AMD Ryzen 7 5800X3D CPU and a AMD Radeon RX 7700XT (12 GB) GPU (Santa Maria, RS, Brazil).
- -
- GPU memory allocation was handled dynamically, with a cap of 8 GB to avoid contention with other services.
- -
- CPU usage was limited to 80%, with a maximum of 16 threads, ensuring system responsiveness under concurrent load.
4.4. Experimental Evaluation
- Correct (C): The model’s conclusion perfectly matched the ground truth, correctly identifying the compliance status (e.g., compliant/non-compliant) and citing the specific evidence from the retrieved context to justify its reasoning.
- Partially Correct (PC): The response showed promise but was ultimately flawed. This category includes cases where the system retrieved the correct evidence but failed to synthesize it into the right conclusion, or where the reasoning was sound but based on incomplete or slightly irrelevant context.
- Incorrect (I): The response failed completely, providing the wrong conclusion based on faulty reasoning, hallucinated information, or irrelevant retrieved context.
4.4.1. Baseline Model Performance
4.4.2. RAG-Enhanced System Performance
- Retrieval Errors (15/25 cases): The retriever pulled irrelevant or conflicting context, especially for highly ambiguous queries with overlapping terminology.
- Synthesis Failures (10/25 cases): The LLM struggled to reason correctly when the answer required synthesizing information from multiple complex legal document chunks, even when the correct context was provided.
Query: “What are the specific sampling and testing requirements for Mycoplasma in 12-week-old breeder hens?” Retrieved Context (Incomplete): The retriever pulled a general clause regarding flock health, stating, “The establishment must implement a sanitary monitoring and control program to ensure the health of the poultry flock and prevent the spread of diseases.” Missed Critical Context: The retriever failed to pull the specific binding criteria for Mycoplasma testing from IN 44/2001 from the MAPA, which states the following for breeder birds aged 12 weeks:
Initial Test: A Rapid Serum Agglutination (SAR) test must be conducted. Sampling for Mycoplasma gallisepticum: At least 300 samples must be tested. Sampling for Mycoplasma synoviae: At least 100 samples must be tested. Sampling Method: Samples must be selected randomly, with representation from each shed. Confirmatory Testing: If any samples are reactive in the SAR test, they must be complemented with either a Hemagglutination Inhibition (HI) or an ELISA test.Incorrect LLM Response: “According to the regulations, the establishment is required to have a sanitary monitoring program for its flocks. However, specific details on the number of samples or the types of tests required for Mycoplasma in breeder hens are not specified in this particular guideline.”
4.4.3. Quantitative Metrics and Statistical Significance
5. Discussion
5.1. Summary of Findings
5.2. Theoretical and Practical Implications
5.3. Future Directions
5.4. User Interface Considerations
- A document ingestion panel where users can upload reports, legal documents, or certification requests and link them to specific certification workflows;
- Highlighted response explanations, showing which parts of the document were used in generating the answer, to promote transparency;
- Interactive forms for generating and validating certification drafts with contextual validation based on regulatory criteria;
- A versioning system to track document changes and recommendations over time, supporting auditability and compliance;
- Offline capabilities or mobile support for field veterinarians who operate in rural or low-connectivity areas.
5.5. Security, Privacy, and Ethical Considerations
5.6. Limitations
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A. Workflow Examples
Example 1: Correct Mycoplasma Sampling Verification | |
---|---|
Scenario | An analyst needs to verify if the number of samples collected for Mycoplasma testing in a flock of 16-week-old breeder hens (Nucleus NL01) adheres to national regulations. The farm’s report states that 300 samples were tested for Mycoplasma gallisepticum and 100 for Mycoplasma synoviae. |
Generated Query | Verify if the sample collection for Mycoplasma gallisepticum and Mycoplasma synoviae in nucleus NL01 complies with the standards for 16-week-old breeder birds, based on the provided records showing 300 and 100 samples, respectively. |
Retrieved Context (IN 44/2001) | “For breeder birds, monitoring for Mycoplasma gallisepticum and Mycoplasma synoviae is mandatory. At the 16th week of age, a minimum sample size is required. For M. gallisepticum, at least 300 samples must be collected for RSA testing. For M. synoviae, at least 100 samples must be collected…” |
Final Prompt (Template) | Context: Retrieved Legal Text from IN 44/2001 Question: Generated Query Instructions: Based only on the context provided, answer the question. |
LLM’s Response | Based on the provided context from IN 44/2001, the sample collection for nucleus NL01 is compliant. The regulation requires a minimum of 300 samples for M. gallisepticum and 100 for M. synoviae. The farm’s report meets these requirements. |
Expert’s Assessment | Correct. The system accurately identified the clause, interpreted the requirements, and confirmed compliance. |
Example 2: Correct Salmonella Testing Verification | |
Scenario | A certification request for a poultry farm indicates that for a flock aged 24 weeks, environmental drag swabs were collected from 4 sheds for Salmonella testing. The analyst needs to confirm compliance with the PNSA. |
Generated Query | Does the collection of environmental drag swabs from 4 sheds for a 24-week-old flock comply with PNSA regulations for Salmonella monitoring? |
Retrieved Context (IN 78/2003) | “For Salmonella monitoring in breeder establishments, official sampling must occur between the 22nd and 26th week. The procedure involves collecting one pair of drag swabs per shed. The collected material from up to four sheds can be pooled for analysis…” |
Final Prompt (Template) | Context: Retrieved Legal Text from IN 78/2003 Question: Generated Query Instructions: Based only on the context provided, answer the question. |
LLM’s Response | Yes, the procedure is compliant with IN 78/2003. The regulation specifies monitoring between the 22nd and 26th week and allows pooling of up to four sheds. Therefore, sampling from 4 sheds at 24 weeks is correct. |
Expert’s Assessment | Correct. The model correctly identified the age bracket, shed pooling rule, and confirmed compliance. |
Example 3: Partially Correct Response due to Ambiguous Context | |
Scenario | An analyst reviews a case where a farm tested for Mycoplasma at 10 weeks. The query is whether this test is sufficient for certification, without specifying production type (breeder vs. broiler). |
Generated Query | Is a Mycoplasma test conducted at 10 weeks of age sufficient for sanitary certification? |
Retrieved Context (IN 44/2001, IN 78/2003) | “[Chunk 1: IN 44/2001] For breeder birds, mandatory testing is at 16 weeks… [Chunk 2: IN 78/2003] For commercial broilers, a sanitary program is required, but mandatory testing ages are not specified… [Chunk 3: IN 44/2001] Additional tests may be performed at any age if clinical signs are present…” |
Final Prompt (Template) | Context: Retrieved Chunks from IN 44/2001 and IN 78/2003 Question: Generated Query Instructions:Based only on the context provided, answer the question. |
LLM’s Response | A Mycoplasma test at 10 weeks can be part of a control program, especially if clinical signs are present. For breeders, however, the mandatory test is at 16 weeks. Sufficiency depends on production type, which is not specified. |
Expert’s Assessment | Partially Correct. The system distinguished production types and mandatory 16-week testing but did not explicitly conclude “no” for breeder certification. The synthesis lacked decisiveness. |
Appendix B. Glossary of Terms and Acronyms
- AI (Artificial Intelligence) A field of computer science focused on creating systems capable of performing tasks that typically require human intelligence.
- DBMS (Database Management System) Software used to create, manage, and interact with databases, ensuring data integrity, security, and accessibility.
- DL (Deep Learning) A subfield of Machine Learning that uses multi-layered neural networks (deep neural networks) to learn hierarchical representations from data.
- ELISA (Enzyme-Linked Immunosorbent Assay) A common laboratory test used to detect the presence of antibodies or antigens in a sample, widely used in veterinary diagnostics.
- FUNDESA (Fundo de Desenvolvimento e Defesa Sanitária Animal): The Fund for the Development and Animal Health Defense, a private entity in Rio Grande do Sul that supports actions to improve animal health in the state.
- IN (Instrução Normativa, or Normative Instruction) A type of legal regulation issued by Brazilian government agencies, such as the MAPA, to establish technical standards and procedures.
- LGPD (Lei Geral de Proteção de Dados) The Brazilian General Data Protection Law, a legal framework that regulates the collection, use, processing, and storage of personal data in Brazil.
- LLaMA (Large Language Model Meta AI) A family of open-source Large Language Models developed by Meta AI, known for their performance and accessibility.
- LLM (Large Language Model) A type of artificial intelligence model trained on vast numbers of text data to understand and generate human-like language.
- LSTM (Long Short-Term Memory) A type of Recurrent Neural Network (RNN) architecture designed to effectively learn and remember long-range dependencies in sequential data.
- MAPA (Ministério da Agricultura e Pecuária) The Brazilian Ministry of Agriculture and Livestock, the federal body responsible for formulating and implementing policies for agribusiness and regulating the agricultural sector.
- ML (Machine Learning) A subset of AI where algorithms are trained on data to identify patterns and make predictions without being explicitly programmed.
- NLP (Natural Language Processing) A specialized subfield of AI focused on enabling computers to understand, interpret, and generate human language.
- PDSA-RS (Plataforma de Defesa Sanitária Animal do Rio Grande do Sul) The Animal Health Defense Platform of Rio Grande do Sul, a digital system for managing animal health certifications and regulatory compliance in the state.
- PGVector An open-source extension for the PostgreSQL database that enables the storage and querying of vector embeddings, facilitating semantic search and similarity tasks.
- PNSA (Plano Nacional de Sanidade Avícola) The National Poultry Health Plan, a program by the MAPA that establishes measures for the prevention, control, and eradication of diseases affecting poultry.
- QLoRA (Quantized Low-Rank Adaptation) An efficient fine-tuning technique for Large Language Models that reduces memory usage by quantizing the model and using low-rank adapters.
- RAG (Retrieval-Augmented Generation) An AI architecture that enhances the responses of a Large Language Model by retrieving relevant information from an external knowledge base at the time of inference.
- REST API (Representational State Transfer Application Programming Interface) A standardized architecture for designing networked applications, allowing different systems to communicate over the web.
- RNN (Recurrent Neural Network) A class of neural networks designed to work with sequential data, where connections between nodes form a directed graph along a temporal sequence.
- RSA (Rapid Serum Agglutination) A serological test used for the rapid screening of certain diseases in poultry, such as Mycoplasmosis, by detecting antibodies in blood serum.
- RT (Responsável Técnico, or Technical Lead) A licensed professional (typically a veterinarian) legally responsible for overseeing the sanitary procedures and compliance of a farm or establishment.
- SVE (Serviço Veterinário Estadual, a.k.a. State Veterinary Service) The state-level governmental body responsible for executing animal health defense policies and inspections.
References
- Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems 33; Curran Associates, Inc.: Red Hook, NY, USA, 2020; pp. 1877–1901. [Google Scholar]
- Gururangan, S.; Marasović, A.; Swayamdipta, S.; Lo, K.; Beltagy, I.; Downey, D.; Smith, N.A. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 8342–8360. [Google Scholar]
- Bommasani, R.; Hudson, D.A.; Adeli, E.; Altman, R.; Arora, S.; von Arx, S.; Bernstein, M.S.; Bohg, J.; Bosselut, A.; Brunskill, E.; et al. On the Opportunities and Risks of Foundation Models. arXiv 2021, arXiv:2108.07258. [Google Scholar] [CrossRef]
- Descovi, G.; Maran, V.; Ebling, D.; Machado, A. Towards a Blockchain Architecture for Animal Sanitary Control. In Proceedings of the 23rd International Conference on Enterprise Information Systems–Volume 1: ICEIS, Online, 26–28 April 2021; INSTICC, SciTePress: Setúbal, Portugal, 2021; pp. 305–312. [Google Scholar]
- Schneider, R.; Machado, F.; Trois, C.; Descovi, G.; Maran, V.; Machado, A. Speeding Up the Simulation Animals Diseases Spread: A Study Case on R and Python Performance in PDSA-RS Platform. In Proceedings of the 26th International Conference on Enterprise Information Systems–Volume 2: ICEIS, Angers, France, 28–30 April 2024; INSTICC, SciTePress: Setúbal, Portugal, 2024; pp. 651–658. [Google Scholar]
- Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.; Kiela, D.; et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. In Advances in Neural Information Processing Systems 33; Curran Associates, Inc.: Red Hook, NY, USA, 2020; pp. 9459–9474. [Google Scholar]
- Karpukhin, V.; Oguz, B.; Min, S.; Lewis, P.; Wu, L.; Edunov, S.; Chen, D.; Yih, W. Dense Passage Retrieval for Open-Domain Question Answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 6769–6781. [Google Scholar]
- Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach, 4th ed.; Pearson: London, UK, 2020. [Google Scholar]
- Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
- Machado, A.; Maran, V.; Augustin, I.; Wives, L.K.; de Oliveira, J.P.M. Reactive, proactive, and extensible situation-awareness in ambient assisted living. Expert Syst. Appl. 2017, 76, 21–35. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25; Curran Associates, Inc.: Red Hook, NY, USA, 2012. [Google Scholar]
- Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
- Heaton, J.B.; Polson, N.G.; Witte, J.H. Deep learning for finance: Deep portfolios. Appl. Stoch. Model. Bus. Ind. 2017, 33, 3–12. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems 30; Curran Associates, Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training; Technical Report; OpenAI: San Francisco, CA, USA, 2018; Available online: https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf (accessed on 16 November 2024).
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models are Unsupervised Multitask Learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
- Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.-A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar] [CrossRef]
- Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv 2023, arXiv:2307.09288. [Google Scholar] [CrossRef]
- Zhao, W.X.; Zhou, K.; Li, J.; Tang, T.; Wang, X.; Hou, Y.; Min, Y.; Zhang, B.; Zhang, J.; Dong, Z.; et al. A Survey of Large Language Models. arXiv 2024, arXiv:2303.18223. [Google Scholar] [PubMed]
- Zhang, S.; Dong, L.; Li, X.; Zhang, S.; Sun, X.; Wang, S.; Li, J.; Hu, C.; Liu, Y.; Wang, H.; et al. Instruction Tuning for Large Language Models: A Survey. arXiv 2024, arXiv:2308.10792. [Google Scholar] [CrossRef]
- Thirunavukarasu, A.J.; Ting, D.S.J.; Elangovan, K.; Gutierrez, L.; Tan, T.F.; Ting, D.S.W. Large language models in medicine. Nat. Med. 2023, 29, 1930–1940. [Google Scholar] [CrossRef] [PubMed]
- Cadena-Bautista, Á.; López-Ponce, F.F.; Ojeda-Trueba, S.L.; Sierra, G.; Bel-Enguix, G. Exploring the Behavior and Performance of Large Language Models: Can LLMs Infer Answers to Questions Involving Restricted Information? Information 2025, 16, 77. [Google Scholar] [CrossRef]
- Wu, S.; Irsoy, O.; Lu, S.; Dabravolski, V.; Dredze, M.; Gehrmann, S.; Kambadur, P.; Rosenberg, D.; Mann, G. BloombergGPT: A Large Language Model for Finance. arXiv 2023, arXiv:2303.17564. [Google Scholar] [CrossRef]
- Yang, X.; Chen, A.; Pour, N.Z.; Rossi, A.; Li, C.; Al-Garadi, M.A.; Sarker, A.; Magge, A.; Liu, Y.; Bian, J. GatorTron: A Large Clinical Language Model to Unlock Patient Information from Unstructured Electronic Health Records. arXiv 2022, arXiv:2203.03540. [Google Scholar] [CrossRef]
- Izacard, G.; Grave, E. Leveraging Passage Retrieval with Generative Models for Open-Domain Question Answering. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 876–882. [Google Scholar]
- Gao, Y.; Xiong, Y.; Gao, X.; Jia, K.; Pan, J.; Bi, Y.; Dai, Y.; Sun, J.; Wang, H. Retrieval-Augmented Generation for Large Language Models: A Survey. arXiv 2024, arXiv:2312.10997. [Google Scholar] [CrossRef]
- Zhao, P.; Liu, Z.; Wang, W.; Wu, Z.; Yang, S.; Zhang, Z.; Zhou, M. A Survey on Retrieval-Augmented Text Generation. arXiv 2023, arXiv:2202.01110. [Google Scholar]
- Khattab, O.; Zaharia, M. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 25–30 July 2020; pp. 39–48. [Google Scholar]
- Dettmers, T.; Pagnoni, A.; Holtzman, A.; Zettlemoyer, L. QLoRA: Efficient Finetuning of Quantized LLMs. arXiv 2023, arXiv:2305.14314. [Google Scholar] [CrossRef]
- Ministério da Agricultura, Pecuária e Abastecimento (MAPA). Normas técnicas para o controle e a certificação de núcleos e estabelecimentos avícolas para a micoplasmose aviária (Mycoplasma gallisepticum, M. synoviae e M. meleagridis). In Diário Oficial da União. Instrução Normativa no 78, de 03 de novembro de 2003; Ministério da Agricultura, Pecuária e Abastecimento (MAPA): Brasília, Brazil, 2003. [Google Scholar]
- Ministério da Agricultura, Pecuária e Abastecimento (MAPA). Plano Nacional de Sanidade Avícola–PNSA. In Diário Oficial da União. Instrução Normativa no 44, de 02 de outubro de 2001; Ministério da Agricultura, Pecuária e Abastecimento (MAPA): Brasília, Brazil, 2001. [Google Scholar]
Hyperparameter | Value |
---|---|
Temperature | 0.4 |
Context Window Size | 1024 tokens (max) |
Beam Search Width | 16 |
GPU Memory Allocation | 8 GB (dynamic) |
CPU Thread Utilization | 80% (16 threads max) |
Model | C | PC | I | Avg. Generation Time (s) |
---|---|---|---|---|
Baseline | 12% | 0% | 88% | 5.2 |
RAG-Enhanced | 55% | 20% | 25% | 8.3 |
Model | Accuracy | Precision | Recall | F1-Score | Accuracy 95% CI |
---|---|---|---|---|---|
Baseline | 12.0% | 12.0% | 12.0% | 12.0% | [5.6%, 18.4%] |
RAG-Enhanced | 55.0% | 55.0% | 55.0% | 55.0% | [45.3%, 64.8%] |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Montero, P.B.; Bulegon Gassen, J.; Descovi, G.; Barnasque, T.O.; da Silva, G.R.; Machado, F.A.; Casanova, G.V.; Maran, V.; Machado, A. Integrating RAG for Smarter Animal Certification Platforms. Information 2025, 16, 843. https://doi.org/10.3390/info16100843
Montero PB, Bulegon Gassen J, Descovi G, Barnasque TO, da Silva GR, Machado FA, Casanova GV, Maran V, Machado A. Integrating RAG for Smarter Animal Certification Platforms. Information. 2025; 16(10):843. https://doi.org/10.3390/info16100843
Chicago/Turabian StyleMontero, Pedro Bilar, Jonas Bulegon Gassen, Glênio Descovi, Tais Oltramari Barnasque, Gabriel Rodrigues da Silva, Felipe Amadori Machado, Gabriel Vieira Casanova, Vinícius Maran, and Alencar Machado. 2025. "Integrating RAG for Smarter Animal Certification Platforms" Information 16, no. 10: 843. https://doi.org/10.3390/info16100843
APA StyleMontero, P. B., Bulegon Gassen, J., Descovi, G., Barnasque, T. O., da Silva, G. R., Machado, F. A., Casanova, G. V., Maran, V., & Machado, A. (2025). Integrating RAG for Smarter Animal Certification Platforms. Information, 16(10), 843. https://doi.org/10.3390/info16100843