Archive
Discover and discuss technology tools
Explore the Tiscuss archive by category or keyword, then jump into conversations around what matters most.
Tilde.run: AI Agent Sandbox with Versioned Filesystem
Tilde.run: Revolutionizing AI Development with a Versioned Filesystem Tilde.run is a cutting edge platform designed to streamline AI development by providing a …
Running Parallel Pi Agents on a Local Sandbox: AI Infrastructure
Running Parallel Pi Agents on a Local Sandbox: AI Infrastructure In today's rapidly evolving tech landscape, running parallel Pi agents on a local sandbox has e…
AI Safety Measures: Controlling AI Agents' Destructive Actions
Saw a case recently where an AI coding agent ended up wiping a database in seconds. It made me think about how most agent setups are wired: agent decides → executes query → done There’s usually logging-tracing but those all happen after the action. If your agent has access to systems like a DB, are you: restricting it to read-only? running everything in staging/sandbox? relying on prompt-level safeguards? or putting some kind of control layer in between?
Explore Agentic AI with Free Interactive Curriculum on AgentSwarms
Hey Everyone, Over the last few months, I noticed a massive gap in how we learn about Agentic AI. There are a million theoretical blog posts and dense whitepapers on RAG, tool calling, and swarms, but almost nowhere to just sit down, run an agent, break it, and see how the prompt and tools interact under the hood. So, I built **AgentSwarms**.fyi It’s a free, interactive curriculum for Agentic AI. Instead of just reading, you run live agents alongside the lessons. **What it covers:** * Prompt engineering & system messages (seeing how temperature and persona change behavior). * RAG (Retrieval-Augmented Generation) vs. Fine-tuning. * Tool / Function Calling (OpenAI schemas, MCP servers). * Guardrails & HITL (Human-in-the-Loop) for safe deployments. * Multi-Agent Swarms (orchestrators vs. peer-to-peer handoffs). **The Tech/Setup:** You don't need to install anything or provide API keys to start. The "Learn Mode" is completely free and sandboxed. If you want to mess around with your own models, there's a "Build Mode" where you can plug in your own keys (OpenAI, Anthropic, Gemini, local models, etc.). I’d love for this community to tear it apart. What agent patterns am I missing? Is the observability dashboard actually useful for debugging your traces? Let me know what you think.
Open-Source AI Infrastructure for Desktop Control
Open-source infrastructure for Computer-Use Agents. Sandboxes, SDKs, and benchmarks to train and evaluate AI agents that can control full desktops (macOS, Linux, Windows).