Sequentially Fine-Tuned Language Model: jnjj/xd_v1

Model Description

This repository hosts a language model that is being sequentially fine-tuned using Low-Rank Adaptation (LoRA) on a diverse range of datasets from the Hugging Face Hub. The process starts from the base model jnjj/multi-dataset-model (or its last fine-tuned state from this repository) and continuously adapts by merging LoRA weights after each dataset training cycle. This experiment aims to build a model with broad, cumulatively acquired knowledge. Current Base for Fine-Tuning: jnjj/multi-dataset-model The fully merged model weights and tokenizer are updated periodically at the root of this repository.

Training Methodology

  • Iterative Fine-Tuning: The model undergoes cycles of training on different dataset configurations.
  • LoRA Integration: PEFT's LoRA is employed for parameter-efficient fine-tuning. Adapters are merged post-training.
  • Dynamic Dataset Source: The script iterates through a wide array of Hugging Face Hub datasets.
  • Rapid Iteration Strategy: Training per dataset configuration is brief (max_steps=1), prioritizing breadth of exposure over depth on any single dataset.

Training Progress

  • Datasets Processed (Successfully trained on at least one config): 1
  • Text Examples Streamed (Total): 6
  • Tokens Processed (Total): 3072
  • Last Successful Model Update: 2025-05-08 18:02:08 UTC

Evaluation Snapshot (Approximate)

  • Current Perplexity (wikitext Subset): 282.70
  • Perplexity Change: -0.51 ⬇️ (vs previous cycle's perplexity)

Generated Examples (Qualitative Assessment)

Category Input Prompt Snippet Generated Output Snippet
Story Continuation Once upon a time, in a small villag... How do I get the best picture of what we...
Simple Instruction Explain in one sentence why trees a... I have been trying to make progress and ...
Creative Prompt Describe a friendly robot that love... We are pleased to announce the launch of...
Question Answering (Basic) What is the main color of a ripe ba... As an example we've been using the same ...
Code Generation (Simple Python) Write a Python function that takes ... We are looking forward to seeing us in t...
Reasoning (Simple) If a train leaves station A at 10:0... The time of day we were trying to get ou...

Standard Benchmarks (via lighteval)

Note: Running standard benchmarks requires a dedicated setup using the lighteval harness. The table below shows scores if available in evaluation_stats.json, otherwise N/A.

Common Benchmarks

Category Benchmark # Shots Metric This Model (xd_v1) Llama 3.1 70B (Ref)
Reasoning & Knowledge MMLU (Avg) 5 acc_norm N/A 79.3
Reasoning & Knowledge MMLU-Pro 5 acc N/A 53.8
Reasoning & Knowledge MATH 4 acc N/A 41.6
Reasoning & Knowledge TruthfulQA (MC2) 0 mc2 N/A -
Reasoning & Knowledge GPQA Diamond 0 acc N/A 50.5
Code MBPP 3 pass@1 N/A 66.4
Code LiveCodeBench 0 pass@1 N/A 33.3
Multilingual TydiQA 1 f1 N/A 29.9
Multilingual MGSM 0 acc N/A 91.1

How to Use

Load the model and tokenizer via transformers:

from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "jnjj/xd_v1"
# For local usage after downloading:
# model_id = "./model_files"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
# model.to("cuda") # if GPU is available
prompt = "Explain the concept of photosynthesis in simple terms:"
inputs = tokenizer(prompt, return_tensors="pt") # .to("cuda" if GPU available)
output_sequences = model.generate(
    **inputs,
    max_new_tokens=100,
    do_sample=True,
    temperature=0.7,
    top_p=0.9,
    pad_token_id=tokenizer.eos_token_id
)
generated_text = tokenizer.decode(output_sequences[0], skip_special_tokens=True)
print(generated_text)

Limitations & Considerations

  • This model is an experimental artifact of continuous learning; quality and coherence may vary.
  • Biases present in the underlying datasets may be reflected or amplified.
  • Performance on specific tasks is not guaranteed and may fluctuate as new datasets are incorporated.
  • Intended for research and exploration of sequential fine-tuning dynamics. For rigorous benchmarking, consider using tools like lighteval.

Disclaimer

This model is provided as-is. It may generate inaccurate, biased, or otherwise problematic content. Users should exercise discretion.

Downloads last month
107
Safetensors
Model size
13M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support