Hacker Newsnew | past | comments | ask | show | jobs | submit | matrixgard's commentslogin

eight months is usually when the first version of "what you built" stops working and you have to build the thing customers actually want. the gap between those two is what most people call "0 revenue growth."

we went through something similar building a fintech tool — spent months on features nobody asked for because we assumed we understood the problem. what actually changed it: stopped building for what we assumed and started building for one specific thing one customer said they'd pay for that day. just one. then the next.

what was the specific use case you landed on?


eight months is usually when the first version of "what you built" stops working and you have to build the thing customers actually want. the gap between those two is what most people call "0 revenue growth."

we went through something similar building a fintech tool — spent months on features nobody asked for because we assumed we understood the problem. what actually changed it: stopped building for what we assumed and started building for one specific thing one customer said they'd pay for that day. just one. then the next.

what was the specific use case you landed on?


We started portfolio builder, where professionals can comprehensively express their work with ease. We believed work speaks louder than shiny websites.

but 5 months, we learned we can't build sustainable business out of it. But we had around 1000+ signups. Users gave us feedback, currently my portfolio looks standalone, if you could connect me with maybe potentials jobs etc then, something value i can pay for.

Then we tried to connect portfolio with ATS. Portfolio-first ATS for startups and lean teams. Then yeah, we got around 10 customers who are ready to pay.

Recently couple of colleges, also expressed interest for such tool.

Now we got our core principle in place, "Build by feedbacks, not with assumptions"

But reality is most startups begins with idea at day 1.

Real game changer is Idea -> problem -> solution -> value for money -> customer, how quickly your startup make these transitions and get your Market Product Fit.

Journey will be hard but hang in there and keep your product floating and keep connecting the dots. No playbook exists


Static IP whitelisting is a nightmare in practice -- we ran it for about 8 months and the support burden was basically someone's part-time job. Every time someone's hotel or coffee shop rotated IPs, they'd open a ticket. We moved to AWS SSM Session Manager a couple years back and haven't touched a bastion since. No open ports, no SSH keys to rotate, and `aws ssm start-session --target i-xxxxxxx` just works from anywhere as long as they have valid IAM creds. CloudTrail picks up the session automatically, which was a side benefit we hadn't even planned for.

IAM Identity Center (was just called SSO before the rename) with a short session duration -- we do 8-hour max with MFA at login -- handles the traveling employee case cleanly. They re-authenticate, no ticket. Overhead compared to running a VPN server you have to patch is basically zero.

The one thing that bit us: we kept a bastion sitting around "just in case" for way too long before we cleaned it up. It was live for almost a year after we didn't need it anymore. What's the main access pattern you're trying to solve -- DB access, SSH to EC2, or something different?


We are trying to address the Backend applications that are not required to be publicly accessible for everyone except the employees/contractors. Things like internal dashboards, UAT testing environments, etc.

The target audience is the ones who have no AWS credentials but should access an application hosted in one of the CSP-hosted environments.


The 20% contamination number on ClawHub was genuinely alarming -- at that scale it's not opportunistic, it's systematic. The multi-pass approach makes sense given how trivially obfuscated payloads evade single-regex scanning; same problem npm has been fighting for years where a base64 decode or dynamic require wrapper kills most static analysis.

One thing worth thinking about beyond detection: even a perfect scanner at install time doesn't protect against skills that start clean and phone home post-install. The runtime layer is a different problem -- restricting what a skill process can actually touch (outbound network, credential paths, filesystem writes outside its own dir) probably matters as much as the intake scan. Seccomp or at minimum per-skill network namespacing would close that gap.

Did any of the 824 original malicious skills survive all 6 passes, or were they each caught by at least one detector?


The vault/proxy layer solving the "2am paste" vector but not the semantic leakage is exactly the gap most teams don't account for. Ephemeral key naming, endpoint patterns, TTL behaviors -- all of this is in the training corpus and no amount of runtime secret rotation changes what the model already internalized. You've essentially found that your threat model stopped at input hygiene but the model itself is a side channel.

What I'd add to the defense stack: structured output validation that flags known internal naming patterns before they reach the client, plus anomaly detection on response metadata -- token budget shifts, refusal rates, response shape changes under semantic pressure. At 75% convergence you're past "interesting research" into "reliable extraction technique," which means you need a detection layer, not just prevention.

Have you tested this against non-OpenAI models like Claude or Gemini to see if the naming bleed is GPT-4o-specific or a broader training corpus problem?


matrixgard — spot on. The vault/proxy layer solves input hygiene (paste risk), but the semantic leakage from training corpus (ek_ prefix, EPHEMERAL_KEY naming, client_secret endpoint, TTL discrepancy) is a systemic weight-level issue. The model itself becomes a side-channel once those patterns are internalized.

Your detection layer suggestions (structured output validation + anomaly detection on refusal rates/response shape) are exactly the right next frontier. I'm seeing 6.7% vulnerability increase just by seeding the model with its own safety policy — the "blink" is real.

On your question: Yes, I am expanding to Claude 3.5 Sonnet and Gemini 1.5 Pro this week to see if the naming bleed is GPT-4o-specific or a broader common corpus problem (likely OpenAI docs in multiple training sets).

Have you seen models refuse legitimate session.update or metadata_nonce flows after public discussion of Realtime API internals? Or is the naming too baked in to remove without breaking utility?

Thanks for the sharp additions — this is the kind of discussion that moves the defense stack forward.


The proxy approach in the other comment handles the technical control side well. The harder part is the auditor question you slipped in at the end — that one trips up almost every team I've talked to. Most companies cannot produce a log showing which employee sent what data to which model, when it was authorized, and what classification level it had. The governance infrastructure for AI just isn't there yet, and auditors are starting to catch up.

SOC 2 and ISO 27001 auditors are now explicitly asking for evidence of data flow controls around AI integrations. "Policy says don't paste customer data into ChatGPT" gets zero credit as a control. What they want to see is technical enforcement with logging — something you can point to and say "here is the record of every outbound prompt from the last 90 days, here is what was flagged." If you can't produce that, it's a finding.

The other gap that's easy to miss: most teams don't know what's sensitive until it's already in a prompt. The classification problem comes before the proxy. What are your teams actually pulling into their prompts right now — internal docs, support tickets, code with credentials, database outputs? That's usually where the real exposure surfaces.


A breach at this scale almost never comes from a single access event — moving a petabyte takes time, and that kind of sustained egress usually means either the detection tooling wasn't watching outbound data flows, or alerts fired and got buried in noise. "1 petabyte" from a hacker claim is probably inflated, but even 5-10% of that is catastrophic depending on what's in it.

What's worth paying attention to here is that Telus Digital is a BPO/outsourcing company, which means the blast radius almost certainly extends to their clients. If your company has API integrations with Telus Digital, or gave them any kind of federated access to internal systems, now is the time to audit what they held and rotate anything they could have touched. Downstream credential exposure in third-party breaches is consistently underreacted to.

The employee data angle is also interesting that usually means developer workstations and internal tooling were in scope, not just the customer-facing layer. Makes the "how did detection miss it" question even harder. Does anyone know if Telus Digital ran a shared SOC across their outsourcing clients, or was each client siloed?


The part that doesn't show up in these posts is what happens when the AI-generated code meets real users. Works fine in dev, clean in staging, then production throws an edge case the model never saw and you're staring at a 3am incident with no mental model of what's actually running.

Vibe coding is genuinely useful for prototyping — I use it. But there's a difference between using AI to move faster and outsourcing your understanding to it entirely. The second one catches up with you the moment something breaks and you have no instinct about where to look.

Curious what the actual product was — did it make it to users before things fell apart, or did it not get that far?


The thing most first-year CTOs don't see coming is the translation problem. You understand the system. The founder understands the market. And there's this gap where critical decisions get made based on whoever can communicate their uncertainty the most confidently.

I've seen it go wrong both ways — CTOs who gold-plate systems nobody uses yet, and founders who promise customers features that are a month away from being architecturally possible. Both happen because the two sides aren't seeing the same thing.

The best early-stage CTOs I've watched work well are the ones who treat "no we can't do that" as a last resort. What feature did you promise that you genuinely couldn't deliver, and what did you actually ship instead?


Agreed. Sometimes you take the risk because the upside is existential.

The key is doing it deliberately. Know what you are trading off, isolate the risk, and avoid baking shortcuts into the core system.

Win the customer if you must, but do it in a way that lets you recover fast rather than paying for it for the next year.


Deliberately is doing a lot of work in that sentence and it's exactly right. The shortcuts that kill you aren't the ones you knew were shortcuts it's the ones that felt like reasonable decisions under pressure and only look like shortcuts in hindsight.

The "isolate the risk" part is where most early CTOs underestimate the blast radius. A shortcut in the auth layer feels isolated until six months later when three other systems are built on top of assumptions it makes. By then it's not a shortcut anymore, it's load-bearing technical debt.

What's the recovery actually looked like in practice when you've caught it early enough — do you get clean rewrites or is it mostly containment?


Twenty years of enterprise supply chain GTM and you're asking if you need a technical co-founder — the answer really depends on what phase you're in and how fast you need to move.

For where you are now (building on top of an existing AI platform, adding domain intelligence and decision workflows), you don't need a co-founder who can invent a new transformer. You need someone who can move fast, understand the product deeply, and ship without a lot of hand-holding. That profile is often a strong contractor or early hire, not someone who needs equal equity to stay motivated.

The co-founder model makes sense when there's genuine tech risk that only a senior technical person can see around — when you're building the infrastructure itself. You're not. The risk you described is clearly on the product and GTM side, which you already own. What's your current thinking on timeline to first deployed version with a real supply chain customer?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: