Ben Lang

SuperPowers AI - Real time ambient visual agents for phones and wearables

by
Claude-grade AI agents that see what you see—on your phone or glasses. Solve visual problems instantly, no coding needed.

Add a comment

Replies

Best
Rohan Arun

Hey Product Hunt 👋.

We noticed there were a lot of powerful tools like Claude code and Github that non-technical people didn't have access to, so for the past few months we decided to make it as easy as possible to level the playing field using real-time visual agents. The problem with existing tools:


❌ Not safe and scary to set up
❌ Requires hardware or tech knowledge of clouds
❌ Just bringing code to non-technical people doesn't solve the problem in terms of UI/UX


SuperPowers AI enables non-technical people to solve impossible problems by vibe-coding agents using voice and real time video.


Unlimited Cheap Computer Use


Instead of paying $200/mo for a Claude Max subscription, we figured out how to get it to work with cheaper, nearly free, models.

How?

Users can edit the voice commands to teach Super how to accomplish any complex multi-step actions on a Mac or Android using entirely english targets.

You don't actually need an expensive Mac mini or a Max subscription to automate everything you already do following this pattern.

Example Power:
Voice command: Get the news
Prompt:
1) Open Google News in chrome
2) Summarize the articles on the page

3) Email me the summary at email@domain.com

So you can get started NOW, for free, and start automating your Mac or Android within minutes!

At launch we support the Meta Display Glasses, Apple Vision Pro, Android XR like Luma Ultra, and SMS/Facetime/WhatsApp video calls to lower the barriers of access. Apple is currently reviewing the iPhone and Apple Vision Pro apps, so please start at getsupers.com on all devices.

Odeth N
Great idea for a wearables! Best of luck!
Rohan Arun
Reed Floren

Looks really cool @rohan_arun1 does it work with the regular Meta's or just the ones with the display? I have the RayBan's and the Oakley HSTN if you need me to test them out.

Rohan Arun

@reed_floren Yes it also works with meta 2 glasses through voice commands and shows the output on the phone instead of the glasses so if you can help test them that would be great!

Ron Palmeri

Very excited about today’s launch.

Real-time visual agents are going to allow non developers to do amazing things in the real world.

Imagine an angel on your shoulder that understands where are you, what you’re looking at, and can intuit your objective, all that with long running context and memory across devices and models.

@rohan_arun1 is the genius behind the Tech, and we’re both really looking forward to where the community takes “vision” into the world.

🚀

Rohan Arun

@ronp Yes super excited for today and it's been great working on this with you!

Abhinav Ramesh

so so cool! Haven't seen anyone building for those with wearables, all the best!

Rohan Arun

@abhinavramesh thanks let me know how you like it!

Mark Rezansoff

This is super interesting -> I can see using this for tracking my photo subjects (face to a name) on my volume sports jobs.

Rohan Arun

@mark_rezansoff If you're interested I can generated that power for you! It's very easy let me know

Rohan Arun

We have live support available to help set up devices in our Discord:

http://discord.gg/phoneclaw

Get started in minutes for free! :)

:-)

Congratulations on the launch! First cheatlayer, now this !! Looking forward to seeing how this product evolves.

Rohan Arun

@forthecool thanks for the support!

Brianna Lin

Very cool idea! Curious what use cases you’re seeing most so far from early users?

Rohan Arun

@brianna_lin So far the most common use cases are social agents, social mods, social posting, etc. There's a lot of requests for home improvement, plumbing, electrical, and fixing cars so we'll be working on those next. What would you like to see?

Jonathan Scanzi

The ambient, always-on framing is what makes this interesting to me most visual AI tools require you to intentionally invoke them, which creates enough friction that people just don't bother. Removing that trigger step could genuinely change the usage pattern. My question is around onboarding: how do you help users develop the mental model of when to trust the agent versus when it's going to hallucinate on something visual? That trust calibration is usually where these kinds of tools lose people in the first two weeks. Curious how you're handling the early activation loop.

Rohan Arun

@jscanzi This is a good question! Each "Power" is basically an "RL" environment, but vibe-coded by consumers. So there can be many mistakes initially, but the real time API manages the experience and records the feedback. So if there is hallucination, it acts as the reward/RL loop that iterates over time. The real time API is saving all of this feedback and has the ability to regenerate the powers based on the errors. If we can solve getting consumers to build RL environments in this feedback loop, we solve the data problem for robotics and dwarf current AI labs.

Jonathan Scanzi

@rohan_arun1 thank you for your feedback!

12
Next
Last