Hacker News

Ask HN: AI code assistants and privacy/security

Hacker News - Thu, 04/24/2025 - 11:23pm

There's a lot of pressure to try every new flavour of LLM coding assistants.

Every day I see programming influencers reporting on their new favourite AI tool.

Everyone seems happy to let AIs soak up as much "context" as they can in order to get the best results.

My question is - how do we know what is being uploaded and accessed?

EG: Recently there's been chat about Cursor moving outside of a project directory and deleting folders.

Which gets me thinking... If you ask an AI to perform an action, or update some content, how do you limit the "context"?

What tooling can you use to provide fences around what AI accesses?

What LLM tools have the best track record on this?

Is everyone just asking LLMs about single files and expecting it to limit itself to related files? How do you trust it to stick to a "project" folder... which is a bit of a vague concept given it's usually just a folder on a wider filesystem that could also be accessed for "context".

Comments URL: https://news.ycombinator.com/item?id=43789822

Points: 1

# Comments: 0

Categories: Hacker News

The Cryptoint Library [pdf]

Hacker News - Thu, 04/24/2025 - 11:17pm
Categories: Hacker News

Ask HN: I am at a loss. What shall I do?

Hacker News - Thu, 04/24/2025 - 11:03pm

I started a startup about 1.5 years ago and raised about $100k. However, most of that is spent on ads, travel and meeting with accelerator mandatory workshops. I have $50k in bank now. I am not drawing salary and living on my savings at this point.

It is incredibly hard to survive without any income on the west coast. I have considered moving but all other obligations doesn't let that happen.

I don't know if I am burned out or something is wrong with my brain. I am just numb to everything since 2023 when my father passed away.

So far we have made $4k but that's it. We released a product in February after significant delay in development. That product flopped. We have made one sale.

I know talking to customer is a cliche at this point. But, I am just not finding anyone to talk to. I have little to no network. Most of our software is for Marketing folks. I am using LinkedIn as primary channel and I do get 3% response. But, it is either not interested or sometime in the future.

I am at a crosspoint now. I don't know if I should continue to work on my startup.

I am relying on drawing from my savings but it is not sustainable. I desperately want to make it work and earn at least living expenses through my work.

I have spent a decade or more in tech but as an introvert and partly autist, I have kept to myself.

How do I find users to talk to and how can I reach out to them? I am finding that building without verifying or talking to users is a costly affair.

I would love to get some guidance.

Comments URL: https://news.ycombinator.com/item?id=43789741

Points: 4

# Comments: 2

Categories: Hacker News

FedRAMP Marketplace

Hacker News - Thu, 04/24/2025 - 10:45pm
Categories: Hacker News

Do Not Train" Meta Tags: The Robots.txt of AI – Will Anyone Respect Them?

Hacker News - Thu, 04/24/2025 - 10:41pm

I've been noticing more creators and platforms quietly adding things like to their pages - kind of like a robots.txt, but for LLMs. For those unfamiliar, robots.txt is a standard file websites use to tell search engines which pages they shouldn't crawl. These new "noai" tags serve a similar purpose, but for AI training models instead of search crawlers.

Some examples of platforms implementing these opt-out mechanisms: - Sketchfab now offers creators an option to block AI training in their account settings - DeviantArt pioneered these tags as part of their content protection approach - ArtStation added both meta tags and updated their Terms of Service - Shutterstock created a compensation model for contributors whose images are used in AI training

But here's where things get concerning - there's growing evidence these tags are being treated as optional suggestions rather than firm boundaries:

- Various creators have reported issues with these tags being ignored. For instance, a discussion on DeviantArt (https://www.deviantart.com/lumaris/journal/NoAI-meta-tag-is-NOT-honored-by-DA-941468316) documents cases where the tags weren't honored, with references to GitHub conversations showing implementation issues

- In a GitHub pull request for an image dataset tool (https://github.com/rom1504/img2dataset/pull/218), developers made respecting these tags optional rather than default, which one commenter described as having "gutted it so that we can wash our hands of responsibility without actually respecting anyone's wishes"

- Raptive Support, a company implementing these tags, admits they "are not yet an industry standard, and we cannot guarantee that any or all bots will respect them" (https://help.raptive.com/hc/en-us/articles/13764527993755-NoAI-Meta-Tag-FAQs)

- A proposal to the HTML standards body (https://github.com/whatwg/html/issues/9334) acknowledges these tags don't enforce consent and compliance "might not happen short of robust regulation"

Some creators have become so cynical that one prominent artist David Revoy announced they're abandoning tags like #NoAI because "the damage has already been done" and they "can't remove [their] art one by one from their database." (https://www.davidrevoy.com/article977/artificial-inteligence-why-i-ll-not-hashtag-my-art-humanart-humanmade-or-noai)

This raises several practical questions:

- Will this actually work in practice without enforcement mechanisms?

- Could it be legally enforceable down the line?

- Has anyone successfully used these tags to prevent unauthorized training?

Beyond the technical implementation, I think this points to a broader conversation about creator consent in the AI era. Is this more symbolic - a signal that people want some version of "AI consent" for the open web? Or could it evolve into an actual standard with teeth?

I'm curious if folks here have added something like this to their own websites or content. Have you implemented any technical measures to detect if your content is being used for training anyway? And for those working in AI: what's your take on respecting these kinds of opt-out signals?

Would love to hear what others think.

Comments URL: https://news.ycombinator.com/item?id=43789634

Points: 1

# Comments: 1

Categories: Hacker News

Pages