Hacker Newsnew | past | comments | ask | show | jobs | submit | film42's commentslogin

It's a cash grab. More conversational AI means more folks running out of free or lower paid tier tokens faster, leading to more upsell opportunities. API users will pay more in output tokens by default.

Example, I asked Claude a high level question about p2p systems and it started writing code in 3 languages. Ignoring the code, asking a follow up about the fundamentals, it answered and then rewrote the code 3 times. After a few minutes I hit a token limit for the first time.


It's pretty ridiculous that the response style doesn't persist for Claude. You need to click into a menu to set it to 'concise' for every single conversation. If I forget to it's immediately apparent when it spits out an absurd amount of text for a simple question.

Claude is a great example of a great product coupled with shitty UX, UI and customer service all in one.

Is it just me or does it slow down significantly after 5 chats or so? Or the fact that you have to set the style for each chat.

Oh, and their sales support is so shit for teams and enterprises that in order to use it effectively, you have to literally make your team register for Claude Max 200 on their personal accounts.


I've had good results saying Do not code, focus on architecture first.

As another comment said, use planning mode. I don't use Claude code (I use cursor) and before they introduced planning mode, I would always say "without writing any code, design blah blah blah"

But now that there's planning mode it's a lot easier.


In claude code you should use Planning mode

I think the cash grab is that by far the biggest use case for these models is personal relationship. Chai AI is doing more tokens per month than Anthropic all together and its just personal relationships.

AKA... 70% of existing google cloud users have filled out a support ticket, which starts with a chat bot.


Maybe look at R2 or Wasabi instead of S3. That would cut your storage bill by 3x and take your cloud network bill to zero. IMO self-managing DBs always sucks no matter what you do.


Or you have PTSD after 10 years of being on-call 24/7 for your company's stack. I've built my next chapter around offloading the pager. Worth every penny.


They created their business on open source. Free software was their top of funnel. Free customers become paid customers, and fund the business. They are more than welcome to change this, but there is no way they don't end up with egg on their face, and that's what we're seeing here.


I have a feeling this is IBM dipping their toes into the water of Postgres after seeing Databricks acquire Neon and Snowflake acquire CrunchyData. Won't be surprised if IBM acquires them next year. It makes a lot of sense and I wish everyone the best of luck.

From the outset, CockroachDB has a "just scale it up" product, an open-source presence, a serverless product with pricing similar to DSQL from AWS, and enterprise support. There is a lot to work with there if you're IBM.


I'm a fan because it's something you can explicitly turn on and off. For my Docker based app, I really want to verify the completeness of imports. Preferably, at build and test time. In fact, most of the time I will likely disable lazy loading outright. But, I would really appreciate a faster loading CLI tool.

However, there is a pattern in python to raise an error if, say, pandas doesn't have an excel library installed, which is fine. In the future, will maintainers opt to include a bunch of unused libraries since they won't negatively impact startup time? (Think pandas including 3-4 excel parsers by default, since it will only be loaded when called). It's a much better UX, but, now if you opt out of lazy loading, your code will take longer to load than without it.


My guess is both will look about the same with real world workloads. Worker is certainly more predictable which is safer in general. That said, I appreciate the callout about signal throughput on workers (fewer connections farm to more processes vs each connection getting its own io_uring setup with upper bound being the throughput for a single process). Again, I doubt it makes any difference for 99.9999% of apps out there.


Very cool post. If Jeff Geerling is reading this, I wouldn't mind watching a video on each of these ;)


Check out saveitforparts on youtube, he does lots of this stuff.


Gabe just quit his dayjob to go full-time on SaveItForParts, so hopefully we'll be seeing even more cool stuff in the near future. Me personally, I'm hoping for a collab between him and Jeff. They've had some interaction already (Jeff donated Gabe a spare computer he had lying around) so maybe... That would be epic if it did happen.


They already did a quick collab with Jeff's PIs in space video. But I'm definitely hoping for more.


Aaah. I haven't watched that one yet, so I didn't even know. I'll definitely check that out later today.


I think I update Vundle like once every 3 years.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: