π₯ Adversarial fine-tuning lab for small LLMs (1Bβ3B). Break models βοΈ, harden them π‘οΈ, and measure what actually improves π.
TΕki (ιΆε¨) β ceramic, shaped under pressure.
Models, like clay, only reveal their strength when stress-tested. TΕki is about forcing models through pressure β adversarial inputs β and reshaping them into something more robust.
TΕki is an end-to-end adversarial ML lab:
-
Generate adversarial prompts (jailbreaks, edge cases, failure modes)
-
Fine-tune models using LoRA / QLoRA (MLX or HuggingFace)
-
Evaluate robustness before and after training
-
Publish:
- adversarial datasets π¦
- hardened model weights π§
- evaluation reports π
LLMs are brittle.
- They fail under adversarial prompts
- They overfit to narrow behaviors
- Thereβs little systematic research on small model robustness
Most teams:
test a few prompts and call it βsafeβ
TΕki answers:
Do models actually get safer β or just better at passing tests?
- Adversarial ML & red-teaming
- LoRA / QLoRA fine-tuning
- Dataset construction & curation
- Robustness evaluation & benchmarking
- π¦ Rust CLI β orchestration, experiments, pipelines
- π Python core β training, generation, evaluation
git clone https://github.com/yourusername/toki.git
cd toki
cargo buildBreak the model. Fix the model. Prove it.
If you want next step, I can: β unify all 4 under a Konjo umbrella README + architecture diagram (thatβs what really makes this pop in interviews)