In Q3 2024, our 12-person platform team slashed log ingestion spend by 35% in 90 days, moving from a brittle Elasticsearch-based pipeline to a tuned Vector 0.30 and Loki 3.0 stack—without losing a single log or breaking our 99.95% SLA. GameStop makes $55.5B takeover offer for eBay (279 points) Talking to 35 Strangers at the Gym (144 points) Newton's law of gravity passes its biggest test (15
If this is useful, a ❤️ helps others find it. Everything I keep looking up when building Tauri v2 apps — in one place. // Define #[tauri::command] fn greet(name: String) -> String { format!("Hello, {}!", name) } // With error handling #[tauri::command] fn read_file(path: String) -> Result { std::fs::read_to_string(path).map_err(|e| e.to_string()) } // Async #[tauri::command] async fn fet
Hermes Agent from Nous Research is a model-agnostic, tool-using assistant you run locally or on a VPS. Hermes does not lock you into one surface. You can use the classic hermes / hermes chat CLI, the full-screen hermes --tui session, a long-running hermes gateway for Telegram, Discord, Slack, and other messaging platforms, hermes dashboard for a local browser UI when the web extra is installed.
We Cut Compliance Costs by 40% Using Pulumi 3.140 and Chef 18 for Multi-Cloud AWS and GCP Modern multi-cloud environments offer unmatched flexibility, but they also introduce complex compliance challenges. For our team managing hybrid infrastructure across AWS and GCP, manual policy enforcement and fragmented tooling were driving up compliance costs by 22% year-over-year. By integrating Pulumi 3
In Q3 2024, our 12-person platform engineering team reduced confirmed security incidents by 41.7% (from 72 to 42 per quarter) after rolling out Trivy 0.50 for pre-deployment scanning and Falco 0.40 for runtime detection across 142 production microservices. We didn’t rewrite our CI/CD pipeline, we didn’t hire a dedicated security team, and we didn’t spend a dime on enterprise security tools. Here’s
If this is useful, a ❤️ helps others find it. Everything I keep looking up when building with Gemini — in one place. Model Context Best for gemini-2.5-flash-preview 1M tokens General use, thinking, fast gemini-2.5-pro-preview 1M tokens Complex reasoning, best quality gemini-1.5-flash 1M tokens Stable, production-ready gemini-1.5-pro 2M tokens Longest context gemini-2.0-flash-lite 1M