Security Ops News

[P] [<] [N] [>] | Pages: 1 2 3 4 5

> Post #47006594 by davidbarker | 200 points | 134 comments | 1h ago
GPT-5.2 derives a new result in theoretical physics
(No body text)
> Post #47003064 by ozzyphantom | 1026 points | 503 comments | 6h ago
Apple, fix my keyboard before the timer ends or I'm leaving iPhone
(No body text)
> Post #47007656 by danso | 18 points | 15 comments | 16m ago
The EU moves to kill infinite scrolling
(No body text)
> Post #47001871 by penguin_booze | 604 points | 114 comments | 8h ago
Monosketch
(No body text)
> Post #46939613 by zdw | 140 points | 15 comments | 4d ago
Sandwich Bill of Materials
(No body text)
> Post #47006393 by austinwang115 | 37 points | 8 comments | 2h ago
Show HN: Skill that lets Claude Code/Codex spin up VMs and GPUs
I've been working on CloudRouter, a skill + CLI that gives coding agents like Claude Code and Codex the ability to start cloud VMs and GPUs.

When an agent writes code, it usually needs to start a dev server, run tests, open a browser to verify its work. Today that all happens on your local machine. This works fine for a single task, but the agent is sharing your computer: your ports, RAM, screen. If you run multiple agents in parallel, it gets a bit chaotic. Docker helps with isolation, but it still uses your machine's resources, and doesn't give the agent a browser, a desktop, or a GPU to close the loop properly. The agent could handle all of this on its own if it had a primitive for starting VMs.

CloudRouter is that primitive — a skill that gives the agent its own machines. The agent can start a VM from your local project directory, upload the project files, run commands on the VM, and tear it down when it's done. If it needs a GPU, it can request one.

cloudrouter start ./my-project cloudrouter start --gpu B200 ./my-project cloudrouter ssh cr_abc123 "npm install && npm run dev" Every VM comes with a VNC desktop, VS Code, and Jupyter Lab, all behind auth-protected URLs. When the agent is doing browser automation on the VM, you can open the VNC URL and watch it in real time. CloudRouter wraps agent-browser [1] for browser automation.

cloudrouter browser open cr_abc123 "http://localhost:3000" cloudrouter browser snapshot -i cr_abc123 # → @e1 [link] Home @e2 [link] Settings @e3 [button] Sign Out cloudrouter browser click cr_abc123 @e2 cloudrouter browser screenshot cr_abc123 result.png Here's a short demo: https://youtu.be/SCkkzxKBcPE

What surprised me is how this inverted my workflow. Most cloud dev tooling starts from cloud (background agents, remote SSH, etc) to local for testing. But CloudRouter keeps your agents local and pushes the agent's work to the cloud. The agent does the same things it would do locally — running dev servers, operating browsers — but now on a VM. As I stopped watching agents work and worrying about local constraints, I started to run more tasks in parallel.

The GPU side is the part I'm most curious to see develop. Today if you want a coding agent to help with anything involving training or inference, there's a manual step where you go provision a machine. With CloudRouter the agent can just spin up a GPU sandbox, run the workload, and clean it up when it's done. Some of my friends have been using it to have agents run small experiments in parallel, but my ears are open to other use cases.

Would love your feedback and ideas. CloudRouter lives under packages/cloudrouter of our monorepo https://github.com/manaflow-ai/manaflow.

[1] https://github.com/vercel-labs/agent-browser

> Post #47005509 by abelanger | 13 points | 3 comments | 3h ago
Building a TUI is easy now
(No body text)
> Post #46993587 by fabienpenso | 22 points | 10 comments | 1d ago
Show HN: Moltis – AI assistant with memory, tools, and self-extending skills
Hey HN. I'm Fabien, principal engineer, 25 years shipping production systems (Ruby, Swift, now Rust). I built Moltis because I wanted an AI assistant I could run myself, trust end to end, and make extensible in the Rust way using traits and the type system. It shares some ideas with OpenClaw (same memory approach, Pi-inspired self-extension) but is Rust-native from the ground up. The agent can create its own skills at runtime.

Moltis is one Rust binary, 150k lines, ~60MB, web UI included. No Node, no Python, no runtime deps. Multi-provider LLM routing (OpenAI, local GGUF/MLX, Hugging Face), sandboxed execution (Docker/Podman/Apple Containers), hybrid vector + full-text memory, MCP tool servers with auto-restart, and multi-channel (web, Telegram, API) with shared context. MIT licensed. No telemetry phoning home, but full observability built in (OpenTelemetry, Prometheus).

I've included 1-click deploys on DigitalOcean and Fly.io, but since a Docker image is provided you can easily run it on your own servers as well. I've written before about owning your content (https://pen.so/2020/11/07/own-your-content/) and owning your email (https://pen.so/2020/12/10/own-your-email/). Same logic here: if something touches your files, credentials, and daily workflow, you should be able to inspect it, audit it, and fork it if the project changes direction.

It's alpha. I use it daily and I'm shipping because it's useful, not because it's done.

Longer architecture deep-dive: https://pen.so/2026/02/12/moltis-a-personal-ai-assistant-bui...

Happy to discuss the Rust architecture, security model, or local LLM setup. Would love feedback.

> Post #47004312 by dawg91 | 102 points | 49 comments | 4h ago
IronClaw: a Rust-based clawd that runs tools in isolated WASM sandboxes
(No body text)
> Post #46945696 by surprisetalk | 24 points | 13 comments | 4d ago
Do Metaprojects
(No body text)
[P] [<] [N] [>] | Pages: 1 2 3 4 5