((((sandro.net))))

terça-feira, 3 de fevereiro de 2026

Show HN: Axiomeer – An open marketplace for AI agents https://ift.tt/Ola53if

Show HN: Axiomeer – An open marketplace for AI agents Hi, I built Axiomeer, an open-source marketplace protocol for AI agents. The idea: instead of hardcoding tool integrations into every agent, agents shop a catalog at runtime, and the marketplace ranks, executes, validates, and audits everything. How it works: - Providers publish products (APIs, datasets, model endpoints) via 10-line JSON manifests - Agents describe what they need in natural language or structured tags - The router scores all options by capability match (70%), latency (20%), cost (10%) with hard constraint filters - The top pick is executed, output is validated (citations required? timestamps?), and evidence quality is assessed deterministically - If the evidence is mock/fake/low-quality, the agent abstains rather than hallucinating - Every execution is logged as an immutable receipt The trust layer is the part I think is missing from existing approaches. MCP standardizes how you connect to a tool server. Axiomeer operates one layer up: which tool, from which provider, and can you trust what came back? Stack: Python, FastAPI, SQLAlchemy, Ollama (local LLM, no API keys). v1 ships with weather providers (Open-Meteo + mocks). The architecture supports any HTTP endpoint that returns structured JSON. Looking for contributors to add real providers across domains (finance, search, docs, code execution). Each provider is ~30 lines + a manifest. https://ift.tt/mTnFMCO February 2, 2026 at 09:43PM

Show HN: Kannada Nudi Editor Web Version https://ift.tt/xLETAbQ

Show HN: Kannada Nudi Editor Web Version Ported the Desktop Version of Kannada Nudi Editor to Web under the guidance of https://kagapa.com/ https://nudiweb.com/ February 3, 2026 at 01:11AM

Show HN: 127 PRs to Prod this wknd with 18 AI agents: metaswarm. MIT licensed https://ift.tt/BgJyc1z

Show HN: 127 PRs to Prod this wknd with 18 AI agents: metaswarm. MIT licensed A few weeks ago I posted about GoodToGo https://ift.tt/rSafhmw - a tool that gives AI agents a deterministic answer to "is this PR ready to merge?" Several people asked about the larger orchestration system I mentioned. This is that system. I got tired of being a project manager for Claude Code. It writes code fine, but shipping production code is seven or eight jobs — research, planning, design review, implementation, code review, security audit, PR creation, CI babysitting. I was doing all the coordination myself. The agent typed fast. I was still the bottleneck. What I really needed was an orchestrator of orchestrators - swarms of swarms of agents with deterministic quality checks. So I built metaswarm. It breaks work into phases and assigns each to a specialist swarm orchestrator. It manages handoffs and uses BEADS for deterministic gates that persist across /compact, /clear, and even across sessions. Point it at a GitHub issue or brainstorm with it (it uses Superpowers to ask clarifying questions) and it creates epics, tasks, and dependencies, then runs the full pipeline to a merged PR - including outside code review like CodeRabbit, Greptile, and Bugbot. The thing that surprised me most was the design review gate. Five agents — PM, Architect, Designer, Security, CTO — review every plan in parallel before a line of code gets written. All five must approve. Three rounds max, then it escalates to a human. I expected a rubber stamp. It catches real design problems, dependency issues, security gaps. This weekend I pointed it at my backlog. 127 PRs merged. Every one hit 100% test coverage. No human wrote code, reviewed code, or clicked merge. OK, I guided it a bit, mostly helping with plans for some of the epics. A few learnings: Agent checklists are theater. Agents skipped coverage checks, misread thresholds, or decided they didn't apply. Prompts alone weren't enough. The fix was deterministic gates — BEADS, pre-push hooks, CI jobs all on top of the agent completion check. The gates block bad code whether or not the agent cooperates. The agents are just markdown files. No custom runtime, no server, and while I built it on TypeScript, the agents are language-agnostic. You can read all of them, edit them, add your own. It self-reflects too. After every merged PR, the system extracts patterns, gotchas, and decisions into a JSONL knowledge base. Agents only load entries relevant to the files they're touching. The more it ships, the fewer mistakes it makes. It learns as it goes. metaswarm stands on two projects: https://ift.tt/pP2534V by Steve Yegge (git-native task tracking and knowledge priming) and https://ift.tt/QFl1f3j by Jesse Vincent (disciplined agentic workflows — TDD, brainstorming, systematic debugging). Both were essential. Background: I founded Technorati, Linuxcare, and Warmstart; tech exec at Lyft and Reddit. I built metaswarm because I needed autonomous agents that could ship to a production codebase with the same standards I'd hold a human team to. $ cd my-project-name $ npx metaswarm init MIT licensed. IANAL. YMMV. Issues/PRs welcome! https://ift.tt/7XzJrkL February 2, 2026 at 10:18PM

segunda-feira, 2 de fevereiro de 2026

Show HN: Prism AI – A research agent that generates 2D/3D visualizations https://ift.tt/wCL3gqo

Show HN: Prism AI – A research agent that generates 2D/3D visualizations https://ift.tt/h7AI2xQ February 2, 2026 at 07:03AM

Show HN: Claw-daw – offline, deterministic terminal-first DAW https://ift.tt/PYjkw3t

Show HN: Claw-daw – offline, deterministic terminal-first DAW I built claw-daw: a tiny MIDI DAW you can drive from the terminal (TUI + headless scripts). Motivation: I wanted “music making that feels like coding” — reproducible, diffable, and automation-friendly. Same script + same seed → same beat. Features: • offline (FluidSynth + SoundFont) + ffmpeg export • deterministic renders for iteration/agent pipelines • WAV/MP3/MIDI export • projects are JSON (Git-friendly) Would love feedback on the workflow + what features would make this actually useful for you. https://www.clawdaw.com February 2, 2026 at 01:06AM

Show HN: ContractShield – AI contract analyser for freelancers https://ift.tt/Ad0OLzi

Show HN: ContractShield – AI contract analyser for freelancers Built this with Claude Code. Analyses freelance contracts for 12 risk categories (payment terms, IP ownership, scope issues, termination clauses, etc.) and flags problems with specific recommendations. 40% of freelancers report getting stiffed by clients, often due to vague contract terms. This tool aims to help catch those issues before signing. Currently free while validating whether this solves a real problem. Would love HN's feedback, especially on: - Accuracy of the analysis - Whether this is actually useful for freelancers - What's missing or could be improved Tech stack: Node.js, Express, Anthropic Claude API, deployed on Railway. https://ift.tt/5hmJLqM February 2, 2026 at 12:11AM

Show HN: Is AI "good" yet? – tracking HN sentiment on AI coding https://ift.tt/QnBEaql

Show HN: Is AI "good" yet? – tracking HN sentiment on AI coding A survey tracking developer sentiment on AI-assisted coding through Hacker News posts. https://ift.tt/Ta3vw4y February 1, 2026 at 11:06PM

DJ Sandro

http://sandroxbox.listen2myradio.com