::: ::: ::::::::::: :::::::: ::::::::: :::::::: :::::::::
:+:+: :+:+: :+: :+: :+: :+: :+: :+: :+: :+: :+:
+:+ +:+:+ +:+ +:+ +:+ +:+ +:+ +:+ +:+ |:| +:+
+#+ +:+ +#+ +#+ +#+ +#++:++#: +#+ +:+ |#| +:+
+#+ +#+ +#+ +#+ +#+ +#+ +#+ +#+ |#| +#+
### ### ### ### ### ### ### ### ### ### ###
### ### ########### ######## ### ### ######## #########
MICROD v1.0 (micro-distill-grpo-vae)
This model was made with the Micro Distillery app available at:
webxos.netlify.app/MICROD
-Model Distillation Training: Simulate GRPO optimization with VAE filtering for small LLMs (42M-345M params).
-Policy Experimentation: Test group sizes, KL penalties, cache reuse for RLHF-like training.
-VAE Filtering: Apply latent space compression to improve distillation quality.
-Sandbox Testing: Execute safe Python code with feedback masking.
-Export & Deployment: Generate deployable models for inference in various frameworks.
-Offline Usage: PWA supports offline training simulation and exports.
- Downloads last month
- 155
Model tree for webxos/microd_v1
Base model
openai-community/gpt2