the website is just using go standard library and templ.
the voice is grok tts. the images are generate by nano banana from prompts written by claude, according to styles. it is all generated by a program called lg: https://yagnipedia.com/wiki/lg
the blog pages are markdown on disk, generated dynamically for each request, cached by cloudflare.
GPU/CUDA: Yes, but disabled by default for faster builds. To enable: remove LLAMA_CUDA = "OFF" from config.toml and rebuild with CUDA toolkit installed.
Rust library: Absolutely! Add shimmy = { version = "0.1.0", features = ["llama"] } to Cargo.toml. Use the inference engine directly:
let engine = shimmy::engine::llama::LlamaEngine::new();
let model = engine.load(&spec).await?;
let response = model.generate("prompt", opts, None).await?;
No need to spawn processes - just import and use the components directly in your Rust code.
I like the name, it’s a fun portmanteau of Tkinter and intergalactic. It’s also more search engine friendly and less likely to hit a name collision with another project.
can you share your prompt go make the website , what is the stack