--- library_name: Llama3-8b-FlyingManual-Tutor tags: - llama3 - flying-manual - ai-tutoring - llama-factory --- # Model Card for Llama3-8b-FlyingManual-Tutor This model is a fine-tuned version of the Llama3-8b model, specifically trained on the FlyingManual dataset to serve as an AI tutor for aviation-related subjects. It is designed to provide guidance and nudge users when they answer questions incorrectly. ## Model Details ### Model Description - **Developed by:** Canarie Teams - **Model type:** Large Language Model (LLM) for AI Tutoring - **Language(s) (NLP):** English (primary), potentially others depending on the FlyingManual dataset - **Finetuned from model:** Llama3-8b by Meta AI ## How to Get Started with the Model ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_name = "path/to/your/Llama3-8b-FlyingManual-Tutor" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) # Example usage for tutoring def tutor_interaction(question, user_answer): prompt = f"Question: {question}\nUser Answer: {user_answer}\nTutor Response:" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_length=200) response = tokenizer.decode(outputs[0], skip_special_tokens=True) return response.split("Tutor Response:")[-1].strip() # Example question = "What are the primary flight controls of an aircraft?" user_answer = "Steering wheel and gas pedal" tutor_feedback = tutor_interaction(question, user_answer) print(tutor_feedback) ``` ## Training Details ### Training Data The model was fine-tuned on the FlyingManual dataset, augmented with: - Sample Q&A pairs related to aviation topics - Examples of constructive feedback and explanations - Scenarios demonstrating correct and incorrect responses to aviation-related questions ### Training Procedure #### Preprocessing - Conversion of training data into a dialogue format suitable for tutoring interactions - Augmentation of data with tutoring-specific tokens or markers #### Training Hyperparameters ----------------------------- ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data A held-out portion of the FlyingManual dataset, supplemented with: - A set of typical student questions and answers - Scenarios designed to test the model's ability to provide constructive feedback #### Metrics - Human evaluation of tutoring quality (clarity, accuracy, helpfulness) - Task-specific metrics (e.g., ability to correctly identify and address user mistakes) ### Results [Provide the evaluation results here] ## Environmental Impact - **Hardware Type:** 8 x NVIDIA A100 40GB GPUs ## Model Card Authors Canarie Teams ## Model Card Contact [Your contact information or a link to where people can reach out with questions]