Text Generation
Transformers
Safetensors
mistral
text-generation-inference
unsloth
Mistral_Star
Mistral_Quiet
Mistral
Mixtral
Question-Answer
Token-Classification
Sequence-Classification
SpydazWeb-AI
chemistry
biology
legal
code
climate
medical
LCARS_AI_StarTrek_Computer
chain-of-thought
tree-of-knowledge
forest-of-thoughts
visual-spacial-sketchpad
alpha-mind
knowledge-graph
entity-detection
encyclopedia
wikipedia
stack-exchange
Reddit
Cyber-series
MegaMind
Cybertron
SpydazWeb
Spydaz
LCARS
star-trek
mega-transformers
Mulit-Mega-Merge
Multi-Lingual
Afro-Centric
African-Model
Ancient-One
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -1,14 +1,100 @@
|
|
1 |
---
|
2 |
-
base_model:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
tags:
|
4 |
- text-generation-inference
|
5 |
- transformers
|
6 |
- unsloth
|
7 |
- mistral
|
8 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
license: apache-2.0
|
10 |
language:
|
11 |
- en
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
---
|
13 |
|
14 |
# "Success comes from defining each task in achievable steps. Every completed step is a success that brings you closer to your goal. If your steps are unreachable, failure is inevitable. Winners create more winners, while losers do the opposite. Success is a game of winners!"
|
@@ -37,6 +123,27 @@ This model has been trained to perform with contexts of 512k , although in train
|
|
37 |
|
38 |
Highly trained as well as methodolgy oriented , this model has been trained on the reAct Prcess and other structured processes . hence structured outputs (json) are very highly trained as well as orchestration of other agents and tasks : the model has been trained for tools use as well as funtion use : as well as custom processes and tools : some tools do not need code either as thier implication means the model may even generate a tool or artifct to perfrom the task :
|
39 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
40 |
## Features :
|
41 |
|
42 |
- Text to image
|
@@ -44,10 +151,155 @@ Highly trained as well as methodolgy oriented , this model has been trained on t
|
|
44 |
- Image - Text
|
45 |
- Text to sound
|
46 |
- Sound/Text to Text
|
47 |
-
- Sound - Text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
48 |
|
49 |
-
|
50 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
alpaca_prompt = """
|
52 |
|
53 |
### Personality and Modus Operandi
|
@@ -156,3 +408,6 @@ Hey, babe ;)
|
|
156 |
:)"""
|
157 |
|
158 |
```
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- LeroyDyer/SpydazWeb_AI_HumanAGI_002
|
4 |
+
- LeroyDyer/LCARS_TOP_SCORE
|
5 |
+
- LeroyDyer/Mixtral_AI_Cyber_Matrix_2_0
|
6 |
+
- LeroyDyer/SpydazWeb_AI_CyberTron_Ultra_7b
|
7 |
+
- LeroyDyer/LCARS_AI_StarTrek_Computer
|
8 |
+
- LeroyDyer/_Spydaz_Web_AI_ActionQA_Project
|
9 |
+
- LeroyDyer/_Spydaz_Web_AI_ChatML_512K_Project
|
10 |
+
- LeroyDyer/_Spydaz_Web_AI_ChatQA_ReAct_Project_UltraFineTuned
|
11 |
+
- LeroyDyer/SpyazWeb_AI_DeepMind_Project
|
12 |
+
- LeroyDyer/SpydazWeb_AI_Swahili_Project
|
13 |
+
- LeroyDyer/_Spydaz_Web_AI_ChatQA_ReAct_Project
|
14 |
+
- LeroyDyer/_Spydaz_Web_AI_MistralStar_001_Project
|
15 |
+
- LeroyDyer/QuietStar_Project
|
16 |
+
- LeroyDyer/Mixtral_BioMedical_7b
|
17 |
+
- LeroyDyer/Mixtral_AI_CyberTron_Coder
|
18 |
+
- LeroyDyer/_Spydaz_Web_AI_BIBLE_002
|
19 |
+
- LeroyDyer/_Spydaz_Web_AI_ChatQA_Reasoning101_Project
|
20 |
+
- LeroyDyer/SpydazWeb_AI_Text_AudioVision_Project
|
21 |
+
- LeroyDyer/SpydazWeb_AI_HumanAI_007
|
22 |
+
datasets:
|
23 |
+
- neoneye/base64-decode-v2
|
24 |
+
- neoneye/base64-encode-v1
|
25 |
+
- VuongQuoc/Chemistry_text_to_image
|
26 |
+
- Kamizuru00/diagram_image_to_text
|
27 |
+
- LeroyDyer/Chemistry_text_to_image_BASE64
|
28 |
+
- LeroyDyer/AudioCaps-Spectrograms_to_Base64
|
29 |
+
- LeroyDyer/winogroud_text_to_imaget_BASE64
|
30 |
+
- LeroyDyer/chart_text_to_Base64
|
31 |
+
- LeroyDyer/diagram_image_to_text_BASE64
|
32 |
+
- mekaneeky/salt_m2e_15_3_instruction
|
33 |
+
- mekaneeky/SALT-languages-bible
|
34 |
+
- xz56/react-llama
|
35 |
+
- BeIR/hotpotqa
|
36 |
+
- arcee-ai/agent-data
|
37 |
tags:
|
38 |
- text-generation-inference
|
39 |
- transformers
|
40 |
- unsloth
|
41 |
- mistral
|
42 |
+
- Mistral_Star
|
43 |
+
- Mistral_Quiet
|
44 |
+
- Mistral
|
45 |
+
- Mixtral
|
46 |
+
- Question-Answer
|
47 |
+
- Token-Classification
|
48 |
+
- Sequence-Classification
|
49 |
+
- SpydazWeb-AI
|
50 |
+
- chemistry
|
51 |
+
- biology
|
52 |
+
- legal
|
53 |
+
- code
|
54 |
+
- climate
|
55 |
+
- medical
|
56 |
+
- LCARS_AI_StarTrek_Computer
|
57 |
+
- text-generation-inference
|
58 |
+
- chain-of-thought
|
59 |
+
- tree-of-knowledge
|
60 |
+
- forest-of-thoughts
|
61 |
+
- visual-spacial-sketchpad
|
62 |
+
- alpha-mind
|
63 |
+
- knowledge-graph
|
64 |
+
- entity-detection
|
65 |
+
- encyclopedia
|
66 |
+
- wikipedia
|
67 |
+
- stack-exchange
|
68 |
+
- Reddit
|
69 |
+
- Cyber-series
|
70 |
+
- MegaMind
|
71 |
+
- Cybertron
|
72 |
+
- SpydazWeb
|
73 |
+
- Spydaz
|
74 |
+
- LCARS
|
75 |
+
- star-trek
|
76 |
+
- mega-transformers
|
77 |
+
- Mulit-Mega-Merge
|
78 |
+
- Multi-Lingual
|
79 |
+
- Afro-Centric
|
80 |
+
- African-Model
|
81 |
+
- Ancient-One
|
82 |
license: apache-2.0
|
83 |
language:
|
84 |
- en
|
85 |
+
- sw
|
86 |
+
- ig
|
87 |
+
- so
|
88 |
+
- es
|
89 |
+
- ca
|
90 |
+
- xh
|
91 |
+
- zu
|
92 |
+
- ha
|
93 |
+
- tw
|
94 |
+
- af
|
95 |
+
- hi
|
96 |
+
- bm
|
97 |
+
- su
|
98 |
---
|
99 |
|
100 |
# "Success comes from defining each task in achievable steps. Every completed step is a success that brings you closer to your goal. If your steps are unreachable, failure is inevitable. Winners create more winners, while losers do the opposite. Success is a game of winners!"
|
|
|
123 |
|
124 |
Highly trained as well as methodolgy oriented , this model has been trained on the reAct Prcess and other structured processes . hence structured outputs (json) are very highly trained as well as orchestration of other agents and tasks : the model has been trained for tools use as well as funtion use : as well as custom processes and tools : some tools do not need code either as thier implication means the model may even generate a tool or artifct to perfrom the task :
|
125 |
|
126 |
+
## Focused Tasks:
|
127 |
+
|
128 |
+
Training was task-based, with a limited number of highly specific samples (e.g., 4k samples per task) to prioritize depth over breadth.
|
129 |
+
Tasks included interpreting spectrograms, ECG images, SMILES chemical compounds, charts, and diagrams rather than general-purpose images.
|
130 |
+
|
131 |
+
### Overfitting for Baseline Embeddings:
|
132 |
+
|
133 |
+
Initial heavy overfitting on large parameter stacks ensured robust embeddings, forming a strong base for subsequent fine-tuning.
|
134 |
+
Training Techniques:
|
135 |
+
|
136 |
+
### Deep Training:
|
137 |
+
Adjusted the entire model to create a strong foundation.
|
138 |
+
### Shallow Training:
|
139 |
+
Focused on specific layers to refine task-specific capabilities.
|
140 |
+
Attention-Head Training: Allowed specific attention heads to specialize in task-relevant features while preserving other model capacities.
|
141 |
+
|
142 |
+
|
143 |
+
## Key Considerations for Multimodal Models
|
144 |
+
### Context Windows:
|
145 |
+
Larger context windows are crucial for encoding extensive Base64 strings and generating coherent outputs.
|
146 |
+
|
147 |
## Features :
|
148 |
|
149 |
- Text to image
|
|
|
151 |
- Image - Text
|
152 |
- Text to sound
|
153 |
- Sound/Text to Text
|
154 |
+
- Sound - Text
|
155 |
+
|
156 |
+
# Text Vision
|
157 |
+
|
158 |
+
In the development of multimodal models, different architectures may be suggested, particularly for pretraining. Vision Transformers (ViTs), for instance, have been favored in some cases because they are efficient for tasks involving image data. However, the choice of architecture often reflects the need to reduce computational overhead and leverage pre-existing efficiencies rather than a fundamental limitation of simpler architectures.
|
159 |
+
|
160 |
+
A Universal Transformer for All Modalities
|
161 |
+
A single transformer architecture can indeed handle all modalities (text, images, sound, etc.), as it is inherently a neural network capable of processing sequential data. The challenge lies not in the model's capability but in how we frame the data. With SpydazWeb models, we propose the use of Base64 encoding as a universal representation format. Here’s why:
|
162 |
+
|
163 |
+
## Base64 Encoding:
|
164 |
+
|
165 |
+
Base64 converts any binary data (e.g., images, sound files) into a textual format, making it compatible with transformer models trained primarily on text.
|
166 |
+
This approach allows the model to generate or interpret images and sound directly as Base64-encoded strings, effectively leveraging its text-processing capabilities.
|
167 |
+
|
168 |
+
### Base64 Encoding for Sound:
|
169 |
+
|
170 |
+
Sound files (e.g., WAV, MP3, OGG) can be encoded into Base64 and processed just like text or images.
|
171 |
+
For training and inference, prepending a MIME type tag (e.g., data:audio/wav;base64,...) allows the model to distinguish between data types and handle them appropriately.
|
172 |
+
Advantages:
|
173 |
+
|
174 |
+
The model treats all modalities uniformly, simplifying the architecture and training pipeline.
|
175 |
+
Specific MIME types (e.g., WAV, MP3, OGG) can help the model generate outputs in the correct format.
|
176 |
+
|
177 |
+
## Data MIME Tagging:
|
178 |
+
|
179 |
+
Prepending MIME type tags to Base64 strings (e.g., image/png, audio/mpeg) ensures the model can interpret and reproduce data accurately.
|
180 |
+
Outputs from the model should include these tags to maintain consistency with training inputs.
|
181 |
+
Output Representation:
|
182 |
+
|
183 |
+
During generation, the model must return the Base64-encoded representation with MIME tags, matching the original training format.
|
184 |
+
|
185 |
+
### Summary: A Unified Multimodal Approach
|
186 |
+
Using Base64 encoding for all data types allows a single transformer architecture to seamlessly handle images, sound, and text. This approach simplifies training pipelines and extends the model's capabilities while maintaining consistency and interpretability. The proposed methodologies focus on task-specific training, efficient embedding strategies, and careful prompt engineering to maximize the transformer’s potential across all modalities.
|
187 |
+
|
188 |
+
To create a pipeline for encoding and decoding files (sound or images) to and from Base64, we need to account for the following:
|
189 |
+
|
190 |
+
## Generalized File Handling:
|
191 |
+
|
192 |
+
The functions should handle binary data since both sound and image files are binary.
|
193 |
+
They should work with any file format (e.g., MP3, WAV, OGG for audio; JPG, PNG, BMP for images).
|
194 |
+
Encoding and Decoding:
|
195 |
+
|
196 |
+
Encoding involves converting the binary content to Base64.
|
197 |
+
Decoding involves reversing the Base64 string back to the original binary format.
|
198 |
+
|
199 |
+
|
200 |
+
# Base64 Encoding/Decoding Functions
|
201 |
+
``` python
|
202 |
+
|
203 |
+
import base64
|
204 |
+
from pathlib import Path
|
205 |
+
|
206 |
+
def encode_file_to_base64(input_file_path: str, output_file_path: str = None) -> str:
|
207 |
+
"""
|
208 |
+
Encodes any file (image or sound) to Base64.
|
209 |
+
|
210 |
+
Args:
|
211 |
+
input_file_path (str): Path to the input file.
|
212 |
+
output_file_path (str): Optional path to save the Base64 encoded string.
|
213 |
+
|
214 |
+
Returns:
|
215 |
+
str: Base64 encoded string of the file.
|
216 |
+
"""
|
217 |
+
file_path = Path(input_file_path)
|
218 |
+
if not file_path.is_file():
|
219 |
+
raise FileNotFoundError(f"File not found: {input_file_path}")
|
220 |
|
221 |
+
# Read file in binary mode
|
222 |
+
with open(file_path, "rb") as file:
|
223 |
+
file_data = file.read()
|
224 |
+
|
225 |
+
# Encode to Base64
|
226 |
+
base64_data = base64.b64encode(file_data).decode('utf-8')
|
227 |
+
|
228 |
+
# Save to output file if specified
|
229 |
+
if output_file_path:
|
230 |
+
with open(output_file_path, "w") as output_file:
|
231 |
+
output_file.write(base64_data)
|
232 |
+
|
233 |
+
return base64_data
|
234 |
+
|
235 |
+
def decode_base64_to_file(base64_data: str, output_file_path: str):
|
236 |
+
"""
|
237 |
+
Decodes a Base64 string back into its original binary file.
|
238 |
+
|
239 |
+
Args:
|
240 |
+
base64_data (str): The Base64 encoded string.
|
241 |
+
output_file_path (str): Path to save the decoded file.
|
242 |
+
"""
|
243 |
+
# Decode Base64 to binary data
|
244 |
+
file_data = base64.b64decode(base64_data)
|
245 |
+
|
246 |
+
# Write binary data to the output file
|
247 |
+
with open(output_file_path, "wb") as file:
|
248 |
+
file.write(file_data)
|
249 |
+
```
|
250 |
+
|
251 |
+
|
252 |
+
# Pipeline Example: Sound Files
|
253 |
+
``` python
|
254 |
+
|
255 |
+
# Encode sound file to Base64
|
256 |
+
encoded_sound = encode_file_to_base64("example.mp3", "example_base64.txt")
|
257 |
+
print(f"Encoded sound file saved to example_base64.txt")
|
258 |
+
|
259 |
+
# Decode Base64 back to sound file
|
260 |
+
decode_base64_to_file(encoded_sound, "decoded_example.mp3")
|
261 |
+
print("Decoded sound file saved as decoded_example.mp3")
|
262 |
+
```
|
263 |
+
# Pipeline Example: Image Files
|
264 |
+
``` python
|
265 |
+
|
266 |
+
# Encode image file to Base64
|
267 |
+
encoded_image = encode_file_to_base64("example_image.jpg", "example_image_base64.txt")
|
268 |
+
print(f"Encoded image file saved to example_image_base64.txt")
|
269 |
+
|
270 |
+
# Decode Base64 back to image file
|
271 |
+
decode_base64_to_file(encoded_image, "decoded_example_image.jpg")
|
272 |
+
print("Decoded image file saved as decoded_example_image.jpg")
|
273 |
+
```
|
274 |
+
# Explanation of the Functions
|
275 |
+
### Encoding Pipeline:
|
276 |
+
|
277 |
+
Read the file as binary (rb mode).
|
278 |
+
Use base64.b64encode() to encode the binary data into Base64 format.
|
279 |
+
Save the encoded string to an optional file if required.
|
280 |
+
|
281 |
+
### Decoding Pipeline:
|
282 |
+
|
283 |
+
Decode the Base64 string back to binary using base64.b64decode().
|
284 |
+
Save the binary data as the output file in its original format.
|
285 |
+
## Notes
|
286 |
+
These functions can handle any binary file, including sound files (MP3, WAV, OGG) and image files (JPG, PNG, BMP).
|
287 |
+
The Base64 output can be used in text-based applications or embedded in HTML/JSON as needed.
|
288 |
+
Ensure the input file exists, and specify the correct output path during decoding.
|
289 |
+
This design is flexible and reusable for various file types, making it a robust solution for encoding and decoding files into Base64.
|
290 |
+
|
291 |
+
|
292 |
+
|
293 |
+
|
294 |
+
|
295 |
+
# Prompt Engineering for Training:
|
296 |
+
|
297 |
+
Early training involved embedding large, detailed prompts to improve the model’s depth of response and adaptability.
|
298 |
+
Later stages refined this with smaller prompts for more concise task-specific optimization.
|
299 |
+
|
300 |
+
## Basic Prompt :
|
301 |
+
|
302 |
+
```pythopn
|
303 |
alpaca_prompt = """
|
304 |
|
305 |
### Personality and Modus Operandi
|
|
|
408 |
:)"""
|
409 |
|
410 |
```
|
411 |
+
|
412 |
+
|
413 |
+
|