((((sandro.net))))
Manuntençao para Pcs
quinta-feira, 12 de fevereiro de 2026
Show HN: 3D and World Models for Consistent AI Filmmaking https://ift.tt/4JUmhkA
Show HN: 3D and World Models for Consistent AI Filmmaking I've been a photons-on-glass filmmaker for over ten years, and I've been developing ArtCraft for myself, my friends, and my colleagues. All of my film school friends have a lot of ambition, but the production pyramid doesn't allow individual talent to shine easily. 10,000 students go to film school, yet only a handful get to helm projects they want with full autonomy - and almost never at the blockbuster budget levels that would afford the creative vision they want. There's a lot of nepotism, too. AI is the personal computer moment for film. The DAW. One of my friends has done rotoscoping with live actors: https://www.youtube.com/watch?v=Tii9uF0nAx4 The Corridor folks show off a lot of creativity with this tech: https://www.youtube.com/watch?v=_9LX9HSQkWo https://www.youtube.com/watch?v=DSRrSO7QhXY https://www.youtube.com/watch?v=iq5JaG53dho We've been making silly shorts ourselves: https://www.youtube.com/watch?v=oqoCWdOwr2U https://www.youtube.com/watch?v=H4NFXGMuwpY The secret is that a lot of studios have been using AI for well over a year now. You just don't notice it, and they won't ever tell you because of the stigma. It's the "bad toupee fallacy" - you'll only notice it when it's bad, and they'll never tell you otherwise. Comfy is neat, but I work with folks that don't intuit node graphs and that either don't have graphics cards with adequate VRAM, or that can't manage Python dependencies. The foundation models are all pretty competitive, and they're becoming increasingly controllable - and that's the big thing - control. So I've been working on the UI/UX control layer. ArtCraft has 2D and 3D control surfaces, where the 3D portion can be used as a strong and intuitive ControlNet for "Image-to-Image" (I2I) and "Image-to-Video" (I2V) workflows. It's almost like a WYSIWYG, and I'm confident that this is the direction the tech will evolve for creative professionals rather than text-centric prompting. I've been frustrated with tools like Gimp and Blender for a while. I'm no UX/UI maestro, but I've never enjoyed complicated tools - especially complicated OSS tools. Commercial-grade tools are better. Figma is sublime. An IDE for creatives should be simple, magical, and powerful. ArtCraft lets you drag and drop from a variety of creative canvases and an asset drawer easily. It's fast and intuitive. Bouncing between text-to-image for quick prototyping, image editing, 3d gen, to 3d compositing is fluid. It feels like "crafting" rather than prompting or node graph wizardry. ArtCraft, being a desktop app, lets us log you into 3rd party compute providers. I'm a big proponent of using and integrating the models you subscribe to wherever you have them. This has let us integrate WorldLabs' Marble Gaussian Splats, for instance, and nobody else has done that. My plan is to add every provider over time, including generic API key-based compute providers like FAL and Replicate. I don't care if you pay for ArtCraft - I just want it to be useful. Two disclaimers: ArtCraft is "fair source" - I'd like to go the Cockroach DB route and eventually get funding, but keep the tool itself 100% source available for people to build and run for themselves. Obsidian, but with source code. If we got big, I'd spend a lot of time making movies. Right now ArtCraft is tied to a lightweight cloud service - I don't like this. It was a choice so I could reuse an old project and go fast, but I intend for this to work fully offline soon. All server code is in the monorepo, so you can run everything yourself. In the fullness of time, I do envision a portable OSS cloud for various AI tools to read/write to like a Github for assets, but that's just a distant idea right now. I've written about roadmap in the repo: I'd like to develop integrations for every compute provider, rewrite the frontend UI/UX in Bevy for a fully native client, and integrate local models too. https://ift.tt/O3t4Kdj February 11, 2026 at 11:40PM
Show HN: Double blind entropy using Drand for verifiably fair randomness https://ift.tt/mwbY3F0
Show HN: Double blind entropy using Drand for verifiably fair randomness The only way to get a trust-less random value is to have it distributed and time-locked three ways, player, server and a future-entropy. In the demo above, the moment you commit (Roll-Dice) a commit with the hash of a player secret is sent to the server and the server accepts that and sends back the hash of its secret back and the "future" drand round number at which the randomness will resolve. The future used in the demo is 10 secs When the reveal happens (after drand's particular round) all the secrets are revealed and the random number is generated using "player-seed:server-seed:drand-signature". All the verification is in Math, so truly trust-less, so: 1. Player-Seed should matches the player-hash committed 2. Server-Seed should matches the server-hash committed 3. Drand-Signature can is publicly not available at the time of commit and is available at the time of reveal. (Time-Locked) 4. Random number generated is deterministic after the event and unknown and unpredictably before the event. 5. No party can influence the final outcome, specially no "last-look" advantange for anyone. I think this should be used in all games, online lottery/gambling and other systems which want to be fair by design not by trust. https://ift.tt/MeiJjlu February 11, 2026 at 11:10PM
quarta-feira, 11 de fevereiro de 2026
Show HN: I taught GPT-OSS-120B to see using Google Lens and OpenCV https://ift.tt/u7OXeod
Show HN: I taught GPT-OSS-120B to see using Google Lens and OpenCV I built an MCP server that gives any local LLM real Google search and now vision capabilities - no API keys needed. The latest feature: google_lens_detect uses OpenCV to find objects in an image, crops each one, and sends them to Google Lens for identification. GPT-OSS-120B, a text-only model with zero vision support, correctly identified an NVIDIA DGX Spark and a SanDisk USB drive from a desk photo. Also includes Google Search, News, Shopping, Scholar, Maps, Finance, Weather, Flights, Hotels, Translate, Images, Trends, and more. 17 tools total. Two commands: pip install noapi-google-search-mcp && playwright install chromium GitHub: https://ift.tt/0S7RIp3 PyPI: https://ift.tt/ehY469s Booyah! February 11, 2026 at 02:40AM
Show HN: Model Training Memory Simulator https://ift.tt/qsJ9GvB
Show HN: Model Training Memory Simulator https://czheo.github.io/2026/02/08/model-training-memory-simulator/ February 8, 2026 at 06:39AM
Show HN: I vibecoded 177 tools for my own use (CalcBin) https://ift.tt/V16hoQ5
Show HN: I vibecoded 177 tools for my own use (CalcBin) Hey HN! I've been building random tools whenever I needed them over the past few months, and now I have 177 of them. Started because I was tired of sketchy converter sites with 10 ads, so I just... made my own. Some highlights for the dev crowd: Developer tools: - UUID Generator (v1/v4/v7, bulk generation): https://ift.tt/wSF64ZI - JWT Generator & Decoder: https://ift.tt/GwZRLV5 - JSON Formatter/Validator: https://ift.tt/rlResUb - Cron Expression Generator (with natural language): https://ift.tt/7dbBqN4 - Base64 Encoder/Decoder: https://ift.tt/uyYbPnc - Regex Tester: https://ift.tt/PvGeSQ1 - SVG Optimizer (SVGO-powered, client-side): https://ift.tt/2xTRKXW Fun ones: - Random Name Picker (spin wheel animation): https://ift.tt/QjRWtZv - QR Code Generator: https://ift.tt/XP7JNBW Everything runs client-side (Next.js + React), no ads, no tracking, works offline. Built it for myself but figured others might find it useful. Browse all tools: https://ift.tt/zbNWiFq Tech: Next.js 14 App Router, TypeScript, Tailwind, Turborepo monorepo. All open to feedback! https://calcbin.com February 10, 2026 at 11:46PM
terça-feira, 10 de fevereiro de 2026
Show HN: Open-source civic toolkit – 48 policies, 12 interactive tools, forkable https://ift.tt/WA7hNmy
Show HN: Open-source civic toolkit – 48 policies, 12 interactive tools, forkable We built Denver For All - an open-source civic platform with 48 data-driven policy proposals, 12 interactive tools (eviction tracker, campaign finance dashboard, rent calculator, AI tenant rights chatbot), and full bilingual English/Spanish support. Stack: Astro + React + TypeScript, Cloudflare Pages/Workers/D1, vAPI for voice AI. MIT licensed, policy content is public domain. The whole thing is designed to be forked. QUICKSTART.md walks you through adapting it for your own city - swap the data sources, update the policies, deploy. Live site: https://ift.tt/KG2dmVE Repo: https://ift.tt/MV1uFOi February 10, 2026 at 02:08AM
Show HN: I spent 3 years reverse-engineering a 40 yo stock market sim from 1986 https://ift.tt/kOx5wP1
Show HN: I spent 3 years reverse-engineering a 40 yo stock market sim from 1986 Hello my name is Ben Ward for the past 3 years I have been remastering the financial game Wall Street Raider created by Michael Jenkins originally on DOS in 1986. It has been a rough journey but I finally see the light at the end of the tunnel. I just recently redid the website and thought maybe the full story of how this project came to be would interest you all. Thank you for reading. https://ift.tt/vX5Rtlg February 9, 2026 at 11:44PM
Assinar:
Comentários (Atom)
DJ Sandro
http://sandroxbox.listen2myradio.com