If this is useful, a ❤️ helps others find it. Everything I keep looking up when building Tauri v2 apps — in one place. // Define #[tauri::command] fn greet(name: String) -> String { format!("Hello, {}!", name) } // With error handling #[tauri::command] fn read_file(path: String) -> Result { std::fs::read_to_string(path).map_err(|e| e.to_string()) } // Async #[tauri::command] async fn fet
Hermes Agent from Nous Research is a model-agnostic, tool-using assistant you run locally or on a VPS. Hermes does not lock you into one surface. You can use the classic hermes / hermes chat CLI, the full-screen hermes --tui session, a long-running hermes gateway for Telegram, Discord, Slack, and other messaging platforms, hermes dashboard for a local browser UI when the web extra is installed.
State that survives a docker compose down is one of those things you don't think about, until your test suite needs it, your local dev needs it, and your CI pipeline absolutely doesn't. LocalStack handles persistence with one switch (PERSISTENCE=1) and it's a Pro-only feature. Floci ships four storage modes, all free, all in core, with per-service overrides. Pick the right tradeoff for the job.
If this is useful, a ❤️ helps others find it. Everything I keep looking up when building with Gemini — in one place. Model Context Best for gemini-2.5-flash-preview 1M tokens General use, thinking, fast gemini-2.5-pro-preview 1M tokens Complex reasoning, best quality gemini-1.5-flash 1M tokens Stable, production-ready gemini-1.5-pro 2M tokens Longest context gemini-2.0-flash-lite 1M