Renato Candido

Renato Candido at

https://github.com/Mega4alik/ollm a lightweight Python library for large-context LLM inference that enables running models like gpt-oss-20B, qwen3-next-80B or Llama-3.1-8B-Instruct on 100k context using ~$200 consumer GPU with 8GB VRAM.