An opinionated list of Python frameworks, libraries, tools, and resources
If this is useful, a ❤️ helps others find it. Everything I keep looking up when building Tauri v2 apps — in one place. // Define #[tauri::command] fn greet(name: String) -> String { format!("Hello, {}!", name) } // With error handling #[tauri::command] fn read_file(path: String) -> Result { std::fs::read_to_string(path).map_err(|e| e.to_string()) } // Async #[tauri::command] async fn fet
Hermes Agent from Nous Research is a model-agnostic, tool-using assistant you run locally or on a VPS. Hermes does not lock you into one surface. You can use the classic hermes / hermes chat CLI, the full-screen hermes --tui session, a long-running hermes gateway for Telegram, Discord, Slack, and other messaging platforms, hermes dashboard for a local browser UI when the web extra is installed.
I needed to coordinate background scripts running across different machines. The obvious answer was Redis. Everyone uses Redis for this. The tutorials all use Redis. The Stack Overflow answers all say "just use Redis." So I looked at what deploying Redis would actually cost me: A running Redis server I had to maintain A broker to connect workers to it Celery or RQ on top of that Memory-based stora
If this is useful, a ❤️ helps others find it. Everything I keep looking up when building with Gemini — in one place. Model Context Best for gemini-2.5-flash-preview 1M tokens General use, thinking, fast gemini-2.5-pro-preview 1M tokens Complex reasoning, best quality gemini-1.5-flash 1M tokens Stable, production-ready gemini-1.5-pro 2M tokens Longest context gemini-2.0-flash-lite 1M