Hi everyone,
I’ve uploaded the weights for a side-project I’ve been working on: Reson.
It’s a fine-tune of LLaMA-7B (LoRA, PEFT + bitsandbytes 4-bit). The dataset is ~11k instruction–response pairs that I wrote/curated, focused less on benchmarks and more on “how the model thinks”.
The aim wasn’t to squeeze out another leaderboard model, but to see what happens if you push a model toward:
- reflecting on its own process (meta-cognition),
- recursive / loop reasoning,
- cross-domain adaptability,
- edge cases like deception/strategy (to simulate human-like flexibility).
Training ran locally, final loss around 0.33 (It was in different batch)
Weights are here: Nexus-Walker/Reson · Hugging Face
What I’d like feedback on:
- How do you evaluate this kind of behavior? Standard metrics don’t capture it.
- How to keep the behavior stable without catastrophic forgetting as dataset grows?
- Any prior work I should read that tried something similar?
This is still experimental and definitely rough around the edges, but I think it shows interesting “proto-agent” behaviors worth exploring.
Curious to hear your thoughts.