Are there any LLMs with less than 1m parameters?
I know that's a weird request and the model would be useless, but I'm doing a proof-of-concept port of llama2.c to DOS and I want a model that can fit inside 640 KB of RAM.
Anything like a 256K or 128K model?
I want to get LLM inferencing working on the original PC. 😆