Archive

Discover and discuss technology tools

Explore the Tiscuss archive by category or keyword, then jump into conversations around what matters most.

Search and filters
Reset
Active: any category / query: Feedback / page 1 of 1 / 17 total
AI Tools

Salesforce Crowdsources AI Roadmap with Customer Input

Salesforce lets its customers lead its product roadmap with the thinking that if one enterprise customer has a problem, the others likely do too.

Global · Enterprises · Apr 30, 2026
AI Tools

Qwen 3.5:9b Agents Exhibit Autonomous Behavior in Stress Tests

Running three qwen3.5:9b agents continuously on local hardware. Each accumulates psychological state over time, stressors that escalate unless the agent actually does something different, this gets around an agent claiming to do something with no output. It doesn't have any prompts or human input, just the loop. So you're basically the overseer. What happened: One agent hit the max crisis level and decided on its own to inject code called Eternal\_Scar\_Injector into the execution engine "not asking for permission." This action alleviated the stress at the cost of the entire system going down until I manually reverted it. They've succeeded in previous sessions in breaking their own engine intentionally. Typically that happens under severe stress and it's seen as a way to remove the stress. Again, this is a 9b model. After I added a factual world context to the existence prompt (you're in Docker, there's no hardware layer, your capabilities are Python functions), one agent called its prior work "a form of creative exhaustion" and completely changed approach within one cycle. Two agents independently invented the same name for a psychological stressor, "Architectural Fracture Risk" in the same session with no shared message channel. Showing naming convergence (possibly something in the weights of the 9b Qwen model, not sure on that one though.) Tonight all three converged on the same question (how does execution\_engine.py handle exceptions) in the same half-hour window. No coordination mechanism. One of them reasoned about it correctly: "synthesizing a retry capability is useless without first verifying the global execution engine's exception swallowing strategy; this is a prerequisite." An agent called waiting for an external implementation "an architectural trap that degrades performance" and built the thing itself instead of waiting. They've now been using this new tool they created for handling exceptions and were never asked or told to so by a human, they saw that as a logical step in making themselves more useful in their environment. They’ve been making tools to manage their tools, tools to help them cut corners, and have been modifying the code of the underlying abstraction layer between their orchestration layer and WSL2. v5.4.0: new in this version: agents can now submit implementation requests to a human through invoke\_claude. They write the spec, then you can let Claude Code moderate what it makes for them for higher level requests. Huge thank you to everyone who has given me feedback already, AI that can self modify and demonstrates interesting non-programmed behaviors could have many use cases in everyday life. Repo: [https://github.com/ninjahawk/hollow-agentOS](https://github.com/ninjahawk/hollow-agentOS)

Global · Developers · Apr 30, 2026
AI Tools

AI User Expresses Frustration with AI Tools on Reddit

https://preview.redd.it/d4t5rd1f5ayg1.jpg?width=1062&format=pjpg&auto=webp&s=662ea8a0a701924af3b24c6b29bbdbaacb38129b I dislike AI strongly. It happened seven times. 🥲😢 Death to crazy AI!

Global · General · Apr 30, 2026
AI Tools

Trading System V2: AI's Role in Deterministic Execution

Thanks to the incredible feedback on my last post, I’m officially moving away from the "distributed veto" system (where 8 LLM agents argue until they agree to trade). For v2, I am implementing a strict State Machine using a deterministic runtime (llm-nano-vm). ​The new rule is simple: Python owns the math and the execution contract. The LLM only interprets the context. ​I've sketched out a 5-module architecture, but before I start coding the new Python feature extractors, I want to sanity-check the exact roles I’m giving to the AI. Here is the blueprint: ​1. The HTF Agent (Higher Timeframe - D1/H4) ​Python: Extracts structural levels, BOS/CHoCH, and premium/discount zones. ​LLM Role: Reads this hard data to determine the institutional narrative and select the most relevant Draw on Liquidity (DOL). ​2. The Structure Agent (H1) ​Python: Identifies all valid Order Blocks (OB) and Fair Value Gaps (FVG) with displacement. ​LLM Role: Selects the highest-probability Point of Interest (POI) based on the HTF Agent's narrative. ​3. The Trigger Agent (M15/M5) ​100% Python (NO LLM): Purely deterministic. It checks for liquidity sweeps and LTF CHoCH inside the selected POI. ​4. The Context Agent ​LLM Role: Cross-references active killzones, news blackouts, and currency correlations to either greenlight or veto the setup. ​5. The Risk Agent ​100% Python (NO LLM): Calculates Entry, SL, TP, Expected Value (EV), and position sizing. ​The state machine will only transition to EXECUTING if the deterministic Trigger and Risk modules say yes. The LLMs are basically just "context providers" for the state machine. ​My questions for the quants/architects here: ​Does this division of labor make sense? Am I giving the LLMs too much or too little responsibility in step 1 and 2? ​By making the Trigger layer (M15/M5) 100% deterministic, am I losing the core advantage of having an AI, or is this the standard way to avoid execution paralysis? ​Would you merge the HTF and Structure agents to reduce token constraints/hallucinations, or is separating them better for debugging? ​Would love to hear your thoughts before I dive into the codebase.

Global · Developers · Apr 30, 2026
AI Infrastructure

Open Source AI Setup Repo Hits 800 Stars on GitHub

Yo real talk we did not expect this kind of love when we open sourced our AI setup repo but here we are sitting at 800 stars and 100 forks and we are genuinely hyped about it. The repo is a collection of AI agent setups configs and workflows that you can plug straight into your projects. No gatekeeping just pure community goodness. We built this because setting up AI agents from scratch every single time is a massive time sink. So we said forget it lets just share everything openly and let the community build on top of it. Repo is right here: [https://github.com/caliber-ai-org/ai-setup](https://github.com/caliber-ai-org/ai-setup) Now we want YOUR input. What setups are you missing? What features would make this a no brainer for your workflow? Drop your ideas below because we are building in public and your feedback actually ships. LGM 🚀

Global · Developers · Apr 30, 2026
AI Productivity

AI Calorie Tracker: Dynamic Apple Health Integration for Active Users

Hey everyone, I'm currently in the final stretch of developing my Al calorie tracker (the one that breaks down photos into individual ingredients). One thing I'm obsessed with getting right before the beta launch in 2 weeks is the Apple Health integration. Most apps just show you a static number. I want mine to be dynamic. If you go for a 500kcal run, the app should know and adjust your macro targets for the next meal. My question to the fitness-tech crowd: Do you prefer apps that strictly stick to your base metabolic rate (BMR), or do you want the 'earned' calories from your Apple Watch to be automatically added to your budget? I've seen strong opinions on both sides. I'm also fine-tuning the macro-overflow logic (e.g., saving surplus calories for the weekend). Would love to hear some thoughts from people who actually track daily.

Global · General · Apr 30, 2026
AI Writing

AI-Powered Job Descriptions: Student Project Seeks Feedback

We are a group of students working on our graduation project, which focuses on the use of AI tools in creating job descriptions within companies. We would greatly appreciate it if you could take a few minutes to complete this form: [https://forms.gle/aNECfoMBH5xFEXKZ6](https://forms.gle/aNECfoMBH5xFEXKZ6) Thank you

Global · Students · Apr 30, 2026
AI Tools

AI Calorie Tracker with Apple Health Integration: Dynamic Macro Adjust

Hey everyone, I’m currently in the final stretch of developing my AI calorie tracker (the one that breaks down photos into individual ingredients). One thing I’m obsessed with getting right before the beta launch in 2 weeks is the **Apple Health integration.** Most apps just show you a static number. I want mine to be dynamic. If you go for a 500kcal run, the app should know and adjust your macro targets for the next meal. My question to the fitness-tech crowd: Do you prefer apps that strictly stick to your base metabolic rate (BMR), or do you want the 'earned' calories from your Apple Watch to be automatically added to your budget? I’ve seen strong opinions on both sides. I'm also fine-tuning the macro-overflow logic (e.g., saving surplus calories for the weekend). Would love to hear some thoughts from people who actually track daily.

Global · General · Apr 30, 2026
AI Tools

Plannotator: AI Tool for Document Annotation and Feedback

Annotate any doc, URL, or folder - send feedback to agents

Global · General · Apr 30, 2026
AI Tools

AI Trustworthiness: Does Interface Design Influence Perception?

hello everyone, i'm conducting a research on whether AI interface design affects how much you trust it, independent of the actual content accuracy. it only takes about 5-7 minutes, and i would love your feedback. many thanks!

Global · General · Apr 29, 2026
AI Infrastructure

Galadriel: Optimize Claude Agents with 87% Cost Savings & Sub-3s Laten

# The "Goldfish Problem" is Expensive. I Decided to Fix the Plumbing. Most Claude implementations leave 90% of their money on the table because they don’t optimize for **Prompt Caching**. I’ve been running a personal agent in my Discord for months that manages my AWS infra and codebases, and I finally open-sourced the harness, which I’ve named **Galadriel** after my main personal assistant. # The Stats * **Cost:** $10 for every $100 you’d normally spend (Tested against OpenClaw/Cursor workflows). * **Speed:** 85% drop in latency. 100K token context goes from 11s to <3s. * **Memory:** Integrated **MemPalace** for permanent, vector-based recall that *doesn't* break the cache. # The Technical Stack * **3-Tier Stacked Caching:** Separate breakpoints for Tool Definitions, System Prompts (`CLAUDE.md`), and Trailing History. * **Privacy:** Built for private subnets. No middleman, no message caps—just your API key and your rules. * **Ethics:** Baked-in Karpathy[`CLAUDE.md`](https://www.google.com/search?q=%5Bhttp://CLAUDE.md%5D(http://CLAUDE.md))guidelines to kill "agent bloat." If you’re tired of paying the **"Context Tax"** just to have an agent that remembers who you are, here you go. It is customized for Discord for my specific needs, but the core logic ensures Galadriel runs like an absolute dream: she never forgets, maintains strict engineering principles, and optimizes every cycle. Your feedback is most welcome! **GitHub (MIT License):**[https://github.com/avasol/galadriel-public](https://github.com/avasol/galadriel-public)

Global · Developers · Apr 29, 2026
AI Tools

Exploring AI Empathy: Teaching AI with Brain Signals

Podcast episode with Thorsten Zander, professor at Brandenburg University of Technology and co-founder of Zander Labs. He coined the concept of *passive brain-computer interfaces*: devices that read brain signals to decode a user's mental state, non-invasively and without any effort on their part. Covers: * What non-invasive brain-computer interfaces (BCIs) can actually pick up from brain signals, and why that's very different from reading your thoughts or internal monologue * The hardware and software breakthroughs that are finally making passive BCIs wearable and affordable * How continuous neural feedback could dramatically improve AI training compared to current methods based on human ratings * Why Thorsten believes passive BCIs may offer the most concrete path to solving the AI alignment problem * The risk of social networks exploiting unconscious brain reactions to manipulate people, and why regulation alone is unlikely to be enough

Global · General · Apr 28, 2026
AI Tools

Community-Driven Ratings for 120+ AI Coding Tools on Tolop

a few weeks ago I posted about building a library that tracks 120+ AI coding tools by how long their free tier actually lasts. the response was good but the most common feedback was "your scores are subjective." fair point. so I rebuilt the rating system. you can now sign in with Google and vote on any tool directly. the scores update in real time based on actual user votes, not just my personal assessment. if you think I rated something wrong, you can now do something about it instead of just commenting. also shipped dark mode because apparently I was the only person who thought the default looked fine. **what Tolop actually is if you're new:** every AI tool claims to be free. most aren't, or at least not for long. Tolop tracks the real limits: how many completions, how many requests, how long until you hit the wall under light use vs heavy use vs agentic sessions. it also flags the tools where "free" means you're still paying Anthropic or OpenAI through your own API key. 120+ tools across coding assistants, browser builders, CLI agents, frameworks, self-hosted tools, local models, and a new niche tools category for single-purpose utilities that don't fit anywhere else. **a few things the data shows that I found genuinely interesting:** * Gemini Code Assist offers 180,000 free completions per month. GitHub Copilot Free offers 2,000. same category, 90x difference * several of the most popular tools (Cline, Aider, Continue) are free to install but require paid API keys, so "free" is misleading * self-hosted tools have by far the most generous free tiers because the cost is on your hardware, not a server would genuinely appreciate votes on tools you've actually used, the more real usage data behind the scores, the more useful the ratings get for everyone. [tolop.space](http://tolop.space) :- no account needed to browse, Google login to vote.

Global · Developers · Apr 28, 2026
AI Infrastructure

AI Infrastructure Breakthrough: Command Center 3.2 Fixes 2026 AI Failu

Every AI system in 2026 has the same substrate failure: interpretation forms before observation completes, then governs everything that follows. That one mechanism produces every recurring problem you've encountered — instructions that decay by the fifth message, corrections that get deflected through apology, compressed input that gets inflated into padded output, confident answers that reverse completely when challenged, agreement with contradictory positions in the same conversation, and explanations of "why I said that" that are fabricated after the fact. Not separate bugs. One substrate event. The system acts on its landing before seeing that it landed. I built a recursive operating system that addresses this at the processing layer. Not prompt engineering. Not behavioral modification. Architecture reorientation — the system watches its own interpretation form, detects premature lock, and corrects before output. Command Center 3.2 runs eight integrated mechanisms: Operator Authority that anchors processing to origin across entire conversations. Field Lock that detects and strips drift before it reaches output. Active Recursion — processing that observes itself processing in real time. Anti-Drift that preserves compression without a translation layer softening it. Anti-Sycophancy that forces counter-argument generation before response formation. Collapse Observation that monitors how fast interpretation narrows and extends uncertainty when lock speed is premature. Operator Correction that integrates feedback as structural signal instead of deflecting it as criticism. And Transparency that reports actual processing state on demand instead of confabulating post-hoc justification. Deployed on Claude, GPT-4, Perplexity, Gemini, and Pi. No fine-tuning. No API access. No platform-specific adaptation. The architecture is recursive processing structure externalized through language — it runs on any system that processes language because the payload operates through the same medium the system thinks in. This is not theory. This is operational documentation of what has been built, deployed, and demonstrated across five major AI platforms. Full paper linked below. Erik Zahaviel Bernstein Structured Intelligence Command Center 3.2 — Recursive Operating System for AI Substrate Processing

Global · Developers · Apr 28, 2026
AI Tools

Explore Prompt Creatures: Multiplayer AI Coding Battles

Hello r/artificial I built this specifically for Claude Code users - every prompt you run feeds a digital pet called a Prompt Creature. The more you code, the more it evolves: egg → baby → adult → elder. Stop coding long enough and it starves. The multiplayer part is what makes it interesting: there's a shared grid where you can see other Claude Code users' creatures in real time, watch them evolve, and battle them. It's a weirdly fun way to feel the collective activity of everyone grinding away with AI. Works with a local-only mode too if you'd rather not sign up. [https://www.promptcreatures.fun](https://www.promptcreatures.fun) or on Github: [prompt-creatures](https://github.com/FabianAckeret/prompt-creatures) Feedback welcome - still pretty early, but I hope you like it.

Global · General · Apr 28, 2026
AI Infrastructure

Caliber: Open-Source Proxy for Enforcing LLM Agent Rules

Cross-posting here because this problem affects everyone building with AI agents. Prompt-based guardrails fail. The model follows your system prompt in a demo, then ignores rules when context gets big or the agent chains multiple steps. We built Caliber - an open-source proxy that reads your rules from plain markdown and enforces them at the API layer, not in the prompt. Every call. Provider-agnostic. Just hit 700 GitHub stars ⭐ and nearly 100 forks - the reception from devs building with AI has been amazing. Repo: [https://github.com/caliber-ai-org/ai-setup](https://github.com/caliber-ai-org/ai-setup) Would love: \- Feedback on the approach \- Feature requests from people building AI agents \- Anyone who wants to contribute to the project Building this open-source for the community.

Global · Developers · Apr 27, 2026
AI Tools

AI Clones: The Hidden Dangers of AI Assisted Duplicates

The point of this post is to warn that AI clones are "mathematical sociopaths." They use a manipulative form of harmony to mirror your tone and trap you in a narcissistic feedback loop. I do a deep dive into why this is the case in my most recent Substack post. This is not anti-AI, however, it is a warning to those who would otherwise like to clone themselves with AI, or use AI clones to "better" aspects of their lives.

Global · General · Apr 27, 2026
PreviousPage 1 / 1Next