File size: 1,422 Bytes
45cba0a
 
bf11c3c
0fe5bd3
 
 
 
 
 
 
 
 
 
45cba0a
 
bf11c3c
45cba0a
bf11c3c
45cba0a
bf11c3c
45cba0a
 
 
bf11c3c
 
 
 
 
 
 
45cba0a
bf11c3c
45cba0a
bf11c3c
45cba0a
 
 
 
 
bf11c3c
45cba0a
 
 
bf11c3c
45cba0a
 
 
bf11c3c
45cba0a
 
 
0fe5bd3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
---
library_name: transformers
tags:
- chat
- tool-calling
- instruction-tuned
- edge-device
- conversational
license: mit
datasets:
- Open-Orca/OpenOrca
base_model:
- HuggingFaceTB/SmolLM2-360M-Instruct
---

# Model Card for `thirdeyeai/elevate-360m`

## Model Summary

360M parameter transformer model trained for efficient chat completion and tool call prediction on edge devices. Suitable for low-latency applications.

## Model Details

- **Developed by:** Thirdeye AI  
- **Finetuned from model:** HuggingFaceTB/SmolLM2-360M-Instruct  
- **Model type:** Causal decoder-only transformer  
- **Language(s):** English  
- **License:** apache-2.0  
- **Hardware:** Trained on 1x A100 GPU  
- **Training time:** < 24 hours  

## Model Sources

- **Repository:** https://huggingface.co/thirdeyeai/elevate-360m  

## Uses

### Direct Use

Primarily for chat completion and tool call prediction in edge environments with constrained resources.

### Out-of-Scope Use

Not optimized for multi-language support, long-context reasoning, or open-ended generation without tool grounding.

## Bias, Risks, and Limitations

Trained on publicly available instruction-following datasets. May reflect biases present in those datasets. Not suitable for high-stakes or safety-critical applications.

### Recommendations

Use only with proper evaluation and safety checks in deployment environments. Validate outputs before taking action.