lohameit-3.2 / README.md
LOHAMEIT's picture
Update README.md
f2ff273 verified
metadata
library_name: mlc-llm
base_model: meta-llama/Llama-3.2-3B-Instruct
tags:
  - mlc-llm
  - web-llm

🧠 LOHAMEIT 3.2: Next-Gen Conversational AI

Welcome to LOHAMEIT 3.2, a cutting-edge conversational AI model built using the Llama-3.2-3B-Instruct model in MLC format q4f16_0. This repository leverages state-of-the-art natural language processing (NLP) techniques for scalable, efficient, and real-time conversational agents, with a focus on multi-turn conversation tracking and inference optimization.

This project is compatible with MLC-LLM and WebLLM, enabling easy integration into various platforms for dynamic interaction.

πŸš€ Model Overview

  • Base Model: Llama-3.2-3B-Instruct.
  • Format: Quantized to q4f16_0.
  • Project Compatibility:

This model is optimized for both server-side and local inference environments, enabling real-time, contextually aware conversational capabilities with low memory and hardware requirements.

🌟 Key Features

  • High Efficiency: Supports quantized model execution, making it ideal for resource-constrained environments.
  • Multi-Turn Conversations: Tracks conversational context for coherent, human-like responses.
  • Optimized for Deployment: Ready to be used in web-based interfaces via WebLLM or command-line applications.

πŸ’‘ Why LOHAMEIT 3.2?

This project provides a scalable, highly efficient NLP model suitable for a wide range of applications, from interactive bots to more complex AI-driven systems. LOHAMEIT 3.2 offers:

  • Real-Time Inference: Designed for immediate response times, ideal for live interactions.
  • Low Hardware Requirements: Runs efficiently on CPUs and lower-end GPUs.
  • Flexible Integration: Works across command-line, server, and web-based applications, supporting a variety of use cases.

πŸ› οΈ Installation and Usage

Before starting, ensure that the MLC-LLM library is installed on your system. For installation instructions, visit the MLC-LLM Installation Documentation.

Example Commands:

Chat:

mlc_llm chat HF://mlc-ai/Llama-3.2-3B-Instruct-q4f16_0-MLC

REST Server:

mlc_llm serve HF://mlc-ai/Llama-3.2-3B-Instruct-q4f16_0-MLC

Python API:

from mlc_llm import MLCEngine

# Create engine
model = "HF://mlc-ai/Llama-3.2-3B-Instruct-q4f16_0-MLC"
engine = MLCEngine(model)

# Run chat completion
for response in engine.chat.completions.create(
    messages=[{"role": "user", "content": "What is the meaning of life?"}],
    model=model,
    stream=True,
):
    for choice in response.choices:
        print(choice.delta.content, end="", flush=True)
print("\n")

engine.terminate()

πŸ“‚ Project Structure

LOHAMEIT-3.2/
β”œβ”€β”€ models/                # Model weights and configurations
β”œβ”€β”€ src/                   # Source code for model interaction and API integration
β”œβ”€β”€ scripts/               # Utility scripts for deployment
β”œβ”€β”€ README.md              # This file
└── requirements.txt       # Python dependencies

πŸ”§ Installation Guide

  1. Clone the repository:

    git clone https://github.com/LOHAMEIT/lohameit-3.2.git
    cd lohameit-3.2
    
  2. Install Dependencies: Install the required dependencies using the provided requirements.txt file.

  3. Run the Model: Follow the instructions above to either run the chat interface or deploy as a REST server.

🧠 Future Enhancements

  • Quantization Improvements: Additional support for 4-bit quantization will further reduce resource usage while maintaining performance.
  • Enhanced Conversational Tracking: Improved context management for even more natural conversations over extended sessions.

πŸ“§ Contact

For questions, feedback, or collaboration, feel free to reach out at [email protected].


A heartfelt thank you to G. Akshitha from MLRIT College for her valuable contributions and support throughout this project. Your insights were instrumental in shaping key aspects of LOHAMEIT 3.2.

LEGION-SRCS