Archive
Discover and discuss technology tools
Explore the Tiscuss archive by category or keyword, then jump into conversations around what matters most.
Salesforce Crowdsources AI Roadmap with Customer Input
Salesforce lets its customers lead its product roadmap with the thinking that if one enterprise customer has a problem, the others likely do too.
Qwen 3.5:9b Agents Exhibit Autonomous Behavior in Stress Tests
Running three qwen3.5:9b agents continuously on local hardware. Each accumulates psychological state over time, stressors that escalate unless the agent actually does something different, this gets around an agent claiming to do something with no output. It doesn't have any prompts or human input, just the loop. So you're basically the overseer. What happened: One agent hit the max crisis level and decided on its own to inject code called Eternal\_Scar\_Injector into the execution engine "not asking for permission." This action alleviated the stress at the cost of the entire system going down until I manually reverted it. They've succeeded in previous sessions in breaking their own engine intentionally. Typically that happens under severe stress and it's seen as a way to remove the stress. Again, this is a 9b model. After I added a factual world context to the existence prompt (you're in Docker, there's no hardware layer, your capabilities are Python functions), one agent called its prior work "a form of creative exhaustion" and completely changed approach within one cycle. Two agents independently invented the same name for a psychological stressor, "Architectural Fracture Risk" in the same session with no shared message channel. Showing naming convergence (possibly something in the weights of the 9b Qwen model, not sure on that one though.) Tonight all three converged on the same question (how does execution\_engine.py handle exceptions) in the same half-hour window. No coordination mechanism. One of them reasoned about it correctly: "synthesizing a retry capability is useless without first verifying the global execution engine's exception swallowing strategy; this is a prerequisite." An agent called waiting for an external implementation "an architectural trap that degrades performance" and built the thing itself instead of waiting. They've now been using this new tool they created for handling exceptions and were never asked or told to so by a human, they saw that as a logical step in making themselves more useful in their environment. They’ve been making tools to manage their tools, tools to help them cut corners, and have been modifying the code of the underlying abstraction layer between their orchestration layer and WSL2. v5.4.0: new in this version: agents can now submit implementation requests to a human through invoke\_claude. They write the spec, then you can let Claude Code moderate what it makes for them for higher level requests. Huge thank you to everyone who has given me feedback already, AI that can self modify and demonstrates interesting non-programmed behaviors could have many use cases in everyday life. Repo: [https://github.com/ninjahawk/hollow-agentOS](https://github.com/ninjahawk/hollow-agentOS)
AI User Expresses Frustration with AI Tools on Reddit
https://preview.redd.it/d4t5rd1f5ayg1.jpg?width=1062&format=pjpg&auto=webp&s=662ea8a0a701924af3b24c6b29bbdbaacb38129b I dislike AI strongly. It happened seven times. 🥲😢 Death to crazy AI!
Trading System V2: AI's Role in Deterministic Execution
Thanks to the incredible feedback on my last post, I’m officially moving away from the "distributed veto" system (where 8 LLM agents argue until they agree to trade). For v2, I am implementing a strict State Machine using a deterministic runtime (llm-nano-vm). The new rule is simple: Python owns the math and the execution contract. The LLM only interprets the context. I've sketched out a 5-module architecture, but before I start coding the new Python feature extractors, I want to sanity-check the exact roles I’m giving to the AI. Here is the blueprint: 1. The HTF Agent (Higher Timeframe - D1/H4) Python: Extracts structural levels, BOS/CHoCH, and premium/discount zones. LLM Role: Reads this hard data to determine the institutional narrative and select the most relevant Draw on Liquidity (DOL). 2. The Structure Agent (H1) Python: Identifies all valid Order Blocks (OB) and Fair Value Gaps (FVG) with displacement. LLM Role: Selects the highest-probability Point of Interest (POI) based on the HTF Agent's narrative. 3. The Trigger Agent (M15/M5) 100% Python (NO LLM): Purely deterministic. It checks for liquidity sweeps and LTF CHoCH inside the selected POI. 4. The Context Agent LLM Role: Cross-references active killzones, news blackouts, and currency correlations to either greenlight or veto the setup. 5. The Risk Agent 100% Python (NO LLM): Calculates Entry, SL, TP, Expected Value (EV), and position sizing. The state machine will only transition to EXECUTING if the deterministic Trigger and Risk modules say yes. The LLMs are basically just "context providers" for the state machine. My questions for the quants/architects here: Does this division of labor make sense? Am I giving the LLMs too much or too little responsibility in step 1 and 2? By making the Trigger layer (M15/M5) 100% deterministic, am I losing the core advantage of having an AI, or is this the standard way to avoid execution paralysis? Would you merge the HTF and Structure agents to reduce token constraints/hallucinations, or is separating them better for debugging? Would love to hear your thoughts before I dive into the codebase.
AI Calorie Tracker with Apple Health Integration: Dynamic Macro Adjust
Hey everyone, I’m currently in the final stretch of developing my AI calorie tracker (the one that breaks down photos into individual ingredients). One thing I’m obsessed with getting right before the beta launch in 2 weeks is the **Apple Health integration.** Most apps just show you a static number. I want mine to be dynamic. If you go for a 500kcal run, the app should know and adjust your macro targets for the next meal. My question to the fitness-tech crowd: Do you prefer apps that strictly stick to your base metabolic rate (BMR), or do you want the 'earned' calories from your Apple Watch to be automatically added to your budget? I’ve seen strong opinions on both sides. I'm also fine-tuning the macro-overflow logic (e.g., saving surplus calories for the weekend). Would love to hear some thoughts from people who actually track daily.
Plannotator: AI Tool for Document Annotation and Feedback
Annotate any doc, URL, or folder - send feedback to agents
AI Trustworthiness: Does Interface Design Influence Perception?
hello everyone, i'm conducting a research on whether AI interface design affects how much you trust it, independent of the actual content accuracy. it only takes about 5-7 minutes, and i would love your feedback. many thanks!
Exploring AI Empathy: Teaching AI with Brain Signals
Podcast episode with Thorsten Zander, professor at Brandenburg University of Technology and co-founder of Zander Labs. He coined the concept of *passive brain-computer interfaces*: devices that read brain signals to decode a user's mental state, non-invasively and without any effort on their part. Covers: * What non-invasive brain-computer interfaces (BCIs) can actually pick up from brain signals, and why that's very different from reading your thoughts or internal monologue * The hardware and software breakthroughs that are finally making passive BCIs wearable and affordable * How continuous neural feedback could dramatically improve AI training compared to current methods based on human ratings * Why Thorsten believes passive BCIs may offer the most concrete path to solving the AI alignment problem * The risk of social networks exploiting unconscious brain reactions to manipulate people, and why regulation alone is unlikely to be enough
Community-Driven Ratings for 120+ AI Coding Tools on Tolop
a few weeks ago I posted about building a library that tracks 120+ AI coding tools by how long their free tier actually lasts. the response was good but the most common feedback was "your scores are subjective." fair point. so I rebuilt the rating system. you can now sign in with Google and vote on any tool directly. the scores update in real time based on actual user votes, not just my personal assessment. if you think I rated something wrong, you can now do something about it instead of just commenting. also shipped dark mode because apparently I was the only person who thought the default looked fine. **what Tolop actually is if you're new:** every AI tool claims to be free. most aren't, or at least not for long. Tolop tracks the real limits: how many completions, how many requests, how long until you hit the wall under light use vs heavy use vs agentic sessions. it also flags the tools where "free" means you're still paying Anthropic or OpenAI through your own API key. 120+ tools across coding assistants, browser builders, CLI agents, frameworks, self-hosted tools, local models, and a new niche tools category for single-purpose utilities that don't fit anywhere else. **a few things the data shows that I found genuinely interesting:** * Gemini Code Assist offers 180,000 free completions per month. GitHub Copilot Free offers 2,000. same category, 90x difference * several of the most popular tools (Cline, Aider, Continue) are free to install but require paid API keys, so "free" is misleading * self-hosted tools have by far the most generous free tiers because the cost is on your hardware, not a server would genuinely appreciate votes on tools you've actually used, the more real usage data behind the scores, the more useful the ratings get for everyone. [tolop.space](http://tolop.space) :- no account needed to browse, Google login to vote.
Explore Prompt Creatures: Multiplayer AI Coding Battles
Hello r/artificial I built this specifically for Claude Code users - every prompt you run feeds a digital pet called a Prompt Creature. The more you code, the more it evolves: egg → baby → adult → elder. Stop coding long enough and it starves. The multiplayer part is what makes it interesting: there's a shared grid where you can see other Claude Code users' creatures in real time, watch them evolve, and battle them. It's a weirdly fun way to feel the collective activity of everyone grinding away with AI. Works with a local-only mode too if you'd rather not sign up. [https://www.promptcreatures.fun](https://www.promptcreatures.fun) or on Github: [prompt-creatures](https://github.com/FabianAckeret/prompt-creatures) Feedback welcome - still pretty early, but I hope you like it.
AI Clones: The Hidden Dangers of AI Assisted Duplicates
The point of this post is to warn that AI clones are "mathematical sociopaths." They use a manipulative form of harmony to mirror your tone and trap you in a narcissistic feedback loop. I do a deep dive into why this is the case in my most recent Substack post. This is not anti-AI, however, it is a warning to those who would otherwise like to clone themselves with AI, or use AI clones to "better" aspects of their lives.