Launch HN: Poly (YC S22) – Cursor for Files
Is there something in particular that we are vulnerable to that doesn't also affect Google Drive, Dropbox, iCloud Drive, OneDrive, etc.?
With Google Drive, I choose which files to upload. It doesn't have broad access to everything on my computer.
Dropbox, iCloud, and OneDrive are just backup services, so in theory they could just back up your files as an encrypted blob and have no way to read them. Unfortunately, they don't encrypt them (which is partly why I don't use those services). But at least I have their "promise" that they won't read or analyze my files, which would make me feel better even if its a weak promise.
On the other hand, your service, by nature, is reading an analyzing all of my files using a remote server.
I don't know about the other services, but Dropbox _does_ read your files. https://help.dropbox.com/security/privacy-policy-faq
> We may build models that identify keywords and topics from a given document. These models may be trained on your documents and metadata, and power features within Dropbox such as improved search relevance, auto-sorting and organization features, and document summaries.
This app gives me the same heebie-jeebies as the "Warp" terminal that was heavily pushed (and then rebuked) on HN. I don't want to replace my file browser or terminal with a subscription service, full-stop. The most magical featureset on the market won't move my needle, but then again maybe I'm not the ideal customer for this kind of product.
However, at least for my use-case, this is a very infrequent problem. So, a monthly subscription and the security risk wouldn't be worth it. Though I'm certain there are people who work with files all day and for them, this might be god-send!
Feedback:
Supporting an enterprise air gapped solution of this clearly has huge value. It really doesn't matter where the data is stored if the indexing / embedding is happening on your infrastructure.
Enterprises with compliance requirements are quite likely the types of clients looking for ways to save time searching through petabytes of data.
1. We have an embedded agent that can read, edit, organize, and take actions on your files. This means it can read almost any media type, which is why it's "cursor for files" even though "isn't _cursor_, cursor for files?". In other words, Cursor is "Poly for text" :)
2. We provide you with an "IDE", I.e. a file browser. However, unlike Cursor we actually built our engine rather than relying on an existing open source one like VS Code
Lastly, agree about the enterprise solution. All in due time for sure!
Here's what we always tell founders about demo videos: "What works well for HN is raw and direct, with zero production values. Skip any introductions and jump straight into showing your product doing what it does best. Voiceover is good, but no logos or music!"
Only way I would ever use something like this is with a local/self-host model that I run myself on my own hardware, with meticulous control over what the thing can access on the internet.
We don't use transcription or any post processing. We simply embed the file. Our embedding has an additional inner dimension to support long duration content. So it's [N x D] where D is the embed dimension and N is an internal dimension that varies on the content.
Maybe I'm too old, but after I read the post I thought -- oh this is an "AI-first Quicksilver" -- who remembers that plugin for Mac? I don't think they stayed relevant enough
Do y'all have a solution for this here? Some kind of safety layer of some kind to easily review and optionally revert actions the agent has taken across my entire file system?
The search part of this is cool, but if the write side is solved, shut up and take my money.
Stupid, not good, unsatisfying reasons I agree!
Or if I work in finance, or healthcare, or law, or government, or a hardware design company, I don’t want my files leaving my network. Those are very important use cases, much more important than searching my personal laptop. I want this for WORK, not my little photo collection or notes or whatever.
This is a great use case for modern LLM/embedding models but gotta be local to be actually useful in the places where it’s most needed.
I think the valuable and difficult thing is high quality, fast, offline semantic search, including indexing (so my files don't go to your servers -- that's an immediate no-go for all the important industries).
I've tried my hand at this a couple times and it's hard! But really hope you succeed, because it would be a big game changer in how we organize and use information. Every company out there would buy this immediately.
I am curious to know if your agent can connect to a user’s custom MCP server in order to make tool/calls with files attached. For example I have a custom MCP server that catalogs the images which are submitted in the RPS 2.0 payload. I have a python wrapper script that convert’s Cursor’s stdio MCP interface to an http request, inserting htaccess credentials and scanning the tool call args for any files that it needs to attach to the payload. This lets me set up the wrapper script as an MCP server in the cursor settings. This lets the built in Auto-mode agents do some similar kinds of tasks like you are describing.
With that said, it sounds like your system has better agents and tools for working with things besides just text files, so I am wondering if perhaps your agentic tools might be a good fit for a situation where for example I have a messy filesystem full of images that may already be duplicates in the database. If I provide the MCP tools into my database, can your system and agents use tools that I define? Does your system have ability to attach access credentials if/when making calls to third party mcp/tools?
Also, while it looks visually stunning, it is really not helping much introducing the product. I find it rather confusing, really.
The first thing I noticed - when you searched for "upcycling", all results that you have shown hat the exact match keyword "upcyling" in their filename. If this is just a text grep, then the product is redundant: "everything" exists.
In case you really would get results for "upcycling" based on an internal LLM-based summary of disk content, that would be great - but your live demo does not show that.
Also, later down my "experience" you try to show how user input is handled. But as the "experience" is fixed size and can not be zoomed in, it is impossible to read any of the blurry text.
Again, I understand that this "experience" looks impressive. But it is a complete distraction and does not help at all to introduce your product.
I spent 5 minutes looking at this, one of that looking at a loading spinner like it's 2005 again, and I have no clue what your USP is.
And to be honest: Your HN post here also is not helping. Think "Dropbox + NotebookLM + Perplexity" - what? That makes no sense. It is a complete distraction. Explain what your product/service is, what it's USP is, and what value it provides to me.
Instead after a wall of text you are explaining me that one can select files by clicking on them.
I do not find the idea of bringing my data to your search engine attractive. The data should stay put and you should merely index it.
And since LLMs let you generate things, you also have a feature to do that. But who is going to want to create things -- armed with only a text prompt -- while they are searching? Rarely have I wanted to combine images like in your demo. And on the rare occasion I did -- like today! -- I used a GUI to do it in. I suppose creatives would find this feature more.
we experimented with something like this: https://local-note-taker.vercel.app/chat
repo in case you'd rather run it locally: https://github.com/tambo-ai/local-note-taker
Curious to compare how our two ideas are built differently, feel free to reach out (email in my profile)
Feedback: this site crashes my Safari (Sequoia) reliably, so I can't see the demo.
16 more comments available on Hacker News