Rendered at 20:44:19 GMT+0000 (Coordinated Universal Time) with Cloudflare Workers.
pycassa 4 days ago [-]
Other things in the box:
- Ask-user tool. When the agent hits a judgment call (cost
confirmation, ambiguous field, a captcha), it pauses and asks you
in the chat, not in a page overlay. You answer, it resumes.
I'm really obsessed with the ask user tool on Claude Code,
and obviously I implemented it here also.
- Use oracle to plan complex tasks, take its help when stuck, and also to
create skills
- Sessions. Each task is its own session with its own tab(s) and
history. Switch between them and let the tasks run in the background.
- Bring your own key. Gemini, OpenAI, Anthropic, or any
OpenAI-compatible endpoint. No server of mine in the loop.
- Skills. Teach it or let it figure out a reusable flow and save it as
skill to reuse it.
- Auth handoff. When a login popup opens, the agent blocks, you
complete the auth, the agent picks back up. I purposefully didn't automate
things like auth/captcha, as the expectation of the current websites'
implementations isn't automation.
- Everything local. Traces of every run go to ~/.nimbus/traces/.
No telemetry, YET. Nothing reaches my servers, you just contact the LLM
providers directly.
theflyestpilot 3 days ago [-]
good uiux. Youtube video sound track also good. Very inviting.
pycassa 3 days ago [-]
thank you so much for checking it out. please try it out on macos. supports multiple LLM providers.
hyprcat 2 days ago [-]
Looks interesting. But how does it differ from your typical AI browsers that exist? Did you find any usage differences as UX is changed?
pycassa 2 days ago [-]
hi thanks for checking NIMBUS out.
i mainly wanted a single central place for
giving my intent no matter what it is. this is the central ux principle I used while building NIMBUS. and also why it isnt a chrome extension or an afterthought addition to current browser forks. wanted to see what agent native ui for a browser would look like. i dont think many agent browsers are doing this right now.
i just did alpha launch, so its too early to say, how users are using it. but it is in a state where i myself can actively dog food it.
i also dont consider this browser automation at all, its just the way browsers should be. giving the user the freedom to think about the task at hand and not worry about tab switching, different implementations of different websites. and at the same time give enough visibility into what the agent is doing.
i mainly wanted a single central place for giving my intent no matter what it is. this is the central ux principle I used while building NIMBUS. and also why it isnt a chrome extension or an afterthought addition to current browser forks. wanted to see what agent native ui for a browser would look like. i dont think many agent browsers are doing this right now.
i just did alpha launch, so its too early to say, how users are using it. but it is in a state where i myself can actively dog food it.
i also dont consider this browser automation at all, its just the way browsers should be. giving the user the freedom to think about the task at hand and not worry about tab switching, different implementations of different websites. and at the same time give enough visibility into what the agent is doing.