((((sandro.net))))

domingo, 18 de janeiro de 2026

Show HN: A self-custody medical records prototype (lessons learned) https://ift.tt/W4Ip9gj

Show HN: A self-custody medical records prototype (lessons learned) Hi HN, I built an early prototype exploring whether self-custody medical records can work in practice, using cryptographic proofs without putting sensitive health data on-chain. The problem I’m testing against is healthcare data fragmentation in Indonesia, where patient records are siloed across hospitals and often unavailable in emergencies. Blockchain is used only as an immutable audit layer; the system is designed to work even if the chain changes. Key design choices: - No medical data on-chain (hashes only, for verification and audit) - All records encrypted off-chain - Patients control access via QR-based sharing (doctors don’t touch crypto) - Blockchain treated as a verification layer, not storage Lessons learned so far: - Hospitals won’t run blockchain infrastructure - Doctors won’t manage private keys - UX matters more than cryptography - Key recovery is harder than expected - Regulation shapes architecture early This is not production-ready and doesn’t solve regulation, key recovery, or hospital interoperability yet. I’m mainly looking for critical feedback: - Where this approach is fundamentally flawed - What simpler designs I should consider instead - Healthcare practitioners’ reality checks Repo and technical details are in the README. Happy to answer questions. https://ift.tt/ClPwrI7 January 18, 2026 at 12:53AM

Show HN: Hekate – A Zero-Copy ZK Engine Overcoming the Memory Wall https://ift.tt/7a0Wlnf

Show HN: Hekate – A Zero-Copy ZK Engine Overcoming the Memory Wall Most ZK proving systems are optimized for server-grade hardware with massive RAM. When scaling to industrial-sized traces (2^20+ rows), they often hit a "Memory Wall" where allocation and data movement become a larger bottleneck than the actual computation. I have been developing Hekate, a ZK engine written in Rust that utilizes a Zero-Copy streaming model and a hybrid tiled evaluator. To test its limits, I ran a head-to-head benchmark against Binius64 on an Apple M3 Max laptop using Keccak-256. The results highlight a significant architectural divergence: At 2^15 rows: Binius64 is faster (147ms vs 202ms), but Hekate is already 10x more memory efficient (44MB vs ~400MB). At 2^20 rows: Binius64 hits 72GB of RAM usage, entering swap hell on a laptop. Hekate processes the same workload in 4.74s using just 1.4GB of RAM. At 2^24 rows (16.7M steps): Hekate finishes in 88s with a peak RAM of 21.5GB. Binius64 is unable to complete the task due to OOM/Swap on this hardware. The core difference is "Materialization vs. Streaming". While many engines materialize and copy massive polynomials in RAM during Sumcheck and PCS operations, Hekate streams them through the CPU cache in tiles. This shifts the unit economics of ZK proving from $2.00/hour high-memory cloud instances to $0.10/hour commodity hardware or local edge devices. I am looking for feedback from the community, especially those working on binary fields, GKR, and memory-constrained SNARK/STARK implementations. January 18, 2026 at 12:52AM

sábado, 17 de janeiro de 2026

Show HN: A smart camera that detects eye movements during REM sleep https://ift.tt/Uo4ueHM

Show HN: A smart camera that detects eye movements during REM sleep https://ift.tt/bg93ELq January 17, 2026 at 04:45AM

Show HN: Building the ClassPass for coworking spaces, would love your thoughts https://ift.tt/3opYHlq

Show HN: Building the ClassPass for coworking spaces, would love your thoughts Growing up in a family business focused on coworking and shared spaces, I saw that many people were looking for a coworking space to use for a day. They weren't ready to jump into a long-term agreement. So I created LANS to simplify coworking. Our platform allows users to buy a day pass to a coworking space in seconds. The process is simple: book your pass, arrive at the space, give your name at the front desk, and you're in. Where we are Live in San Francisco with several coworking partners. Recently started expanding beyond the Bay. 10K paid users in San Francisco. Day passes priced between $18 and $25. What we’re seeing Users often use this service. They rotate locations during the week to fit their needs and schedules. For spaces, it’s incremental usage and new foot traffic during the workday. Outside dense city centers, onboarding new spaces tends to be faster. Many suburban areas host nice boutique coworking spaces. But, they often miss a strong online presence. Day passes quickly appeal to both operators and users. What we’re working on Expanding to more cities. Adding supply while keeping quality consistent. Learning which product decisions actually improve repeat usage. Would love feedback from HN: Does this resonate with how you work today? Have you used coworking day passes before? Would you dump your coworking membership for this? https://lans.app/ January 17, 2026 at 01:54AM

Show HN: Making Claude Code sessions link-shareable https://ift.tt/UgtedH7

Show HN: Making Claude Code sessions link-shareable Hey HN! My name is Omkar Kovvali and I've been wanting to share my CC sessions with friends / save + access them easily,so I decided to make an MCP server to do so! /share -> Get a link /import -> resume a conversation in your Claude Code All shared sessions are automatically sanitized, removing api keys, tokens, and secrets. Give it a try following the Github/npm instructions linked below - would love feedback! https://ift.tt/oiNtrsR https://ift.tt/d4xhGDT January 16, 2026 at 11:50PM

sexta-feira, 16 de janeiro de 2026

Show HN: Hc: an agentless, multi-tenant shell history sink https://ift.tt/cgEwnDb

Show HN: Hc: an agentless, multi-tenant shell history sink This project is a tool for engineers who live in the terminal and are tired of losing their command history to ephemeral servers or fragmented `.bash_history` files. If you’re jumping between dozens of boxes, many of which might be destroyed an hour later, your "local memory" (the history file) is essentially useless. This tool builds a centralized, permanent brain for your shell activity, ensuring that a complex one-liner you crafted months ago remains accessible even if the server it ran on is long gone. The core mechanism wants to be a "zero-touch" capture that happens at the connection gateway level. Instead of installing logging agents or scripts on every target machine, the tool reconstructs your terminal sessions from raw recording files generated by the proxy you use to connect. This "in-flight" capture means you get a high-fidelity log of every keystroke and output without ever having to touch the configuration of the remote host. It’s a passive way to build a personal knowledge base while you work. To handle the reality of context-switching, the tool is designed with a "multi-tenant" architecture. For an individual engineer, this isn't about managing different users, but about isolating project contexts. It automatically categorizes history based on the specific organization or project tags defined at the gateway. This keeps your work for different clients or personal side-projects in separate buckets, so you don't have to wade through unrelated noise when you're looking for a specific solution. In true nerd fashion, the search interface stays exactly where you want it: in the command line. There is no bloated web UI to slow you down. The tool turns your entire professional history into a searchable, greppable database accessible directly from your terminal. Please read the full story [here]( https://ift.tt/qnZSDCt... ) https://ift.tt/PZF690W January 16, 2026 at 05:13AM

Show HN: BGP Scout – BGP Network Browser https://ift.tt/13APl6o

Show HN: BGP Scout – BGP Network Browser Hi HN, When working with BGP data, I kept running into the same friction: it’s easy to get raw data, but surprisingly hard to browse networks over time — especially by when they appeared, where they operate, and what they actually look like at a glance. I built a small tool, bgpscout.io, to scratch that itch. It lets you: Browse ASNs by registration date and geography See where a given network appears to have presence View commonly scattered public data about an ASN in one place Save searches to track when new networks matching certain criteria appear All of this data is public already; the goal was to make exploration faster and less painful. I haven’t invested heavily in expanding it yet. Before doing so, I’m curious: Is this solving a real problem for you? What would make something like this actually useful in day-to-day work? Feedback is welcome. https://bgpscout.io/ January 15, 2026 at 09:52PM

DJ Sandro

http://sandroxbox.listen2myradio.com