File size: 1,422 Bytes
45cba0a bf11c3c 0fe5bd3 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a bf11c3c 45cba0a 0fe5bd3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 |
---
library_name: transformers
tags:
- chat
- tool-calling
- instruction-tuned
- edge-device
- conversational
license: mit
datasets:
- Open-Orca/OpenOrca
base_model:
- HuggingFaceTB/SmolLM2-360M-Instruct
---
# Model Card for `thirdeyeai/elevate-360m`
## Model Summary
360M parameter transformer model trained for efficient chat completion and tool call prediction on edge devices. Suitable for low-latency applications.
## Model Details
- **Developed by:** Thirdeye AI
- **Finetuned from model:** HuggingFaceTB/SmolLM2-360M-Instruct
- **Model type:** Causal decoder-only transformer
- **Language(s):** English
- **License:** apache-2.0
- **Hardware:** Trained on 1x A100 GPU
- **Training time:** < 24 hours
## Model Sources
- **Repository:** https://huggingface.co/thirdeyeai/elevate-360m
## Uses
### Direct Use
Primarily for chat completion and tool call prediction in edge environments with constrained resources.
### Out-of-Scope Use
Not optimized for multi-language support, long-context reasoning, or open-ended generation without tool grounding.
## Bias, Risks, and Limitations
Trained on publicly available instruction-following datasets. May reflect biases present in those datasets. Not suitable for high-stakes or safety-critical applications.
### Recommendations
Use only with proper evaluation and safety checks in deployment environments. Validate outputs before taking action. |