- release WebGPU support everywhere, also embed llama.cpp or something similar for non GPU users
- add UI for easy model downloading and sharing among sites
- write the LLM browser API that enables easy access and sets the standard
- add security: "this website wants to use local LLM. Allow?"