Summary

This model is a fine-tuned version of the Mistral-7B-Instruct-v0.3 optimised for answering questions in the domain of forensic investigations. The model has been trained using a specialised dataset titled Advanced_Forensic_Investigations_Knowledge_Library_v1, which consists of approximately 100 domain-specific question-answer pairs. The objective is to support advanced forensic investigative reasoning, rapid knowledge retrieval, and high-precision forensic domain assistance.


Model Details

Model Description

  • Model type: Instruction-following Language Model (LoRA-based fine-tuning)

  • Language(s): English

  • Fine-tuned from model: Mistral-7B-Instruct-v0.3


Training Details

Training Data

  • Dataset: Advanced_Forensic_Investigations_Knowledge_Library_v1

  • Data size: ~100 high-quality, domain-specific QA pairs

Training Procedure

Preprocessing

  • Template: mistral

  • Token truncation/cutoff: 2048

  • No vocab resizing or prompt packing

Hyperparameters

  • Finetuning type: LoRA

  • Precision: bf16

  • LoRA rank: 16

  • LoRA alpha: 32

  • Batch size: 4

  • Gradient accumulation: 8

  • Learning rate: 3e-4

  • Epochs: 35

  • LR scheduler: cosine

  • Quantisation: 4-bit (bitsandbytes)

  • Cutoff length: 2048

Compute

  • Training time: close to 30 minutes

  • Framework: LLaMA-Factory


Evaluation

Testing Data, Factors & Metrics

  • Metrics Used: BLEU-4, ROUGE-1, ROUGE-2

  • Results:

    • BLEU-4: 100%

    • ROUGE-1: 100%

    • ROUGE-2: 100%

These scores reflect perfect overlap with reference answers within the scope of the evaluation dataset.


Technical Specifications

Model Architecture and Objective

  • Base: Transformer (Mistral-7B architecture)

  • Fine-tuning method: LoRA

  • Objective: Instruction-following with forensic legal knowledge adaptation

Compute Infrastructure

  • Hardware: 2xL40s

  • Software: LLaMA-Factory, PyTorch, Transformers, bitsandbytes

Downloads last month
35
Safetensors
Model size
7.25B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for gerasmark/Mistral-7B-Instruct-v0.3-Forensics-v1

Finetuned
(162)
this model
Quantizations
1 model