Hacker Newsnew | past | comments | ask | show | jobs | submit | lazarus01's commentslogin

Really cool experiment Phillip, thanks for sharing!

This makes me recall a conversation from a podcast with Sam Harris where he discusses the “pornography of doubt”

Here is the YouTube clip, less than a minute long

https://youtube.com/shorts/ybUfy3DZK0U?si=o0t8AiLZE4XEeEYV


Thanks! And appreciate you sharing the Harris clip. I sometimes listen to his meditations: https://philippdubach.com/posts/gratitude

Glad you enjoy Sam too!

I use to subscribe to waking up app and really enjoyed the enlightening discussions with other intellectuals in different domains of biology, psychology and science.


Yes, the current technology cannot replace an engineer.

The easiest way to understand why is by understanding natural language. A natural language like english is very messy and and doesn't follow formal rules. It's also not specific enough to provide instructions to a computer, that's why code was created.

The AI is incredibly dumb when it comes to complex tasks with long range contexts. It needs an engineer that understands how to write and execute code to give it precise instructions or it is useless.

Natural Language Processing is so complex, it started around the end of world war two and we are just now seeing innovation in AI where we can mimmick humans, where the AI can do certain things faster than humans. But thinking is not one of them.


LOL. Figuring out how to solve IMO-level math problems without "thinking" would be even more impressive than thinking itself. Now there's a parrot I'd buy.

It isn't thinking it's RL with reward hacking.

It's like taking a student who wins a gold in IMO math, but can't solve easier math problems, because they did not study those type of problems. Where a human who is good at IMO math generalizes to all math problems.

It's just memorizing a trajectory as part of a specific goal. That's what RL is.


It's like taking a student who wins a gold in IMO math, but can't solve easier math problems

I've tried to think of specific follow-up questions that will help me understand your point of view, but other than "Cite some examples of easier problems than a successful IMO-level model will fail at," I've got nothing. Overfitting is always a risk, but if you can overfit to problems you haven't seen before, that's the fault of the test administrators for reusing old problem forms or otherwise not including enough variety.

GPT itself suggests[1] that problems involving heavy arithmetic would qualify, and I can see that being the case if the model isn't allowed to use tools. However, arithmetic doesn't require much in the way of reasoning, and in any case the best reasoning models are now quite decent at unaided arithmetic. Same for the tried-and-true 'strawberry' example GPT cites, involving introspection of its own tokens. Reasoning models are much better at that than base models. Unit conversions were another weakness in the past that no longer seems to crop up much.

So what would some present-day examples be, where models that can perform complex CoT tasks fail on simpler ones in ways that reveal that they aren't really "thinking?"

1: https://chatgpt.com/share/695be256-6024-800b-bbde-fd1a44f281...


In response to your direct question -> https://gail.wharton.upenn.edu/research-and-insights/tech-re...

“ This indicates that while CoT can improve performance on difficult questions, it can also introduce variability that causes errors on “easy” questions the model would otherwise answer correctly.”

Other response to strawberry example; There are 25,000 people employed globally that repair broken responses and create training data, a big whack-a-mole effort to remediate embarrassing errors.


(Shrug) Ancient models are ancient. Please provide specific examples that back up your point, not obsolete .PDFs to comb through.

excellent article and appreciate the author sharing his perspective which is very valuable.

For me the main lesson is, don't let your ego develop from success. Any human is vulnerable to narcissism. It is an interesting phenomenon, where you can originate as a humble person who becomes successful, only to lose your great qualities, when your identity changes. With success you attract different people in your life who may be attracted only to your success and who don't have the stones to confront you on your bs.

Developing healthy self awareness comes from surrounding yourself with people that love you, but are not afraid to keep you honest if you do something out of character.



Is it wise to start to with deep learning without knowing machine learning?

That's a great question. Machine Learning is the overarching space where deep learning is a subspace of machine learning. So if you grasp some basic concepts of machine learning, then you can apply them to deep learning.

All the exciting innovation over the past 13 years comes from deep learning mainly in working with images and natural language.

Machine learning is good for tabular data problems, particularly decision trees, that work well to reduce uncertainty for business outcomes, like sales and marketing as one example.

Machine Learning Basics:

Linear regression - Y = Mx + B (predicts a future value) Classification (logistic regression) - Y = 1 / 1 + e^-(b0 + b1x) (predicts probability of a class or future event)

There is a common learning process between the two called gradient descent. It starts with the loss function, that measures the error between predictions and ground truth, where you backpropogate the errors as a feedback signal to update the learned weights which are the parameters of your ml model which is a more meaningful representation of your dataset that you train on.

In deep learning it's more appropriate for perception problems, like vision ,language and time sequences. It gets more complex where you are dealing with significantly more parameters in the millions, that are organized in hierarchical layer representation.

There are different layers for different types of learning representation, Convolutions for Images and RNN for Sequence to Sequence learning and many more examples of layers, which are the basis of all deep learning models.

So there is a small conceptual overlap; but I would say deep learning has a wider variety of interesting applications, is much more challenging to learn, but not impossible by any stretch.

There is no harm in giving it a try and diving in. If you get lost and drown in complexity, start with machine learning. It took me 3 years to grasp, so it's a marathon, not a sprint.

Hope this helps


I like Karpathy, we come from the same lineage and I am very proud of him for what he's accomplished, he's a very impressive guy.

In regards to deep learning, building deep learning architecture is one of my greatest joys in finding insights from perceptual data. Right now, I'm working on spatiotemporal data modeling to build prediction systems for urban planning to improve public transportation systems. I build ML infrastructure too and plan to release an app that deploys the model in the wild within event streams of transit systems.

It took me a month to master the basics and I've spent a lot of time with online learning, with Deeplearning.ai and skills.google. Deeplearning.ai is ok, but I felt the concepts a bit dated. The ML path at skills.google is excellent and gives a practical understanding of ML infrastructure, optimization and how to work with gpus and tpus (15x faster than gpus).

But the best source of learning for me personally and makes me a confident practitioner is the book by Francois Chollet, the creator of Keras. His book, "Deep Learning with Python", really removed any ambiguity I've had about deep learning and AI in general. Francois is extremely generous in how he explains how deep learning works, over the backdrop of 70 years of deep learning research. Francois keeps it updated and the third revision was made in September 2025 - its available online for free if you don't want to pay for it. He gives you the recipe for building a GPT and Diffusion models, but starts from the ground floor basics of tensor operations and computation graphs. I would go through it again from start to finish, it is so well written and enjoyable to follow.

The most important lesson he discusses is that "Deep learning is more of an art than a science". To get something working takes a good amount of practice and the results on how things work can't always be explained.

He includes notebooks with detailed code examples with Tensorflow, Pytorch and Jax as back ends.

Deep learning is a great skill to have. After reading this book, I can recreate scientific abstracts and deploy the models into production systems. I am very grateful to have these skills and I encourage anyone with deep curiosity like me to go all in on deep learning.


The project you mentioned you are working sounds interesting. Do you have more to share ?

I’m curious how ML/AI is leveraged in the domain of public transport. And what can it offer when compared to agent based models.


The project I’m working on emulates a scientific abstract. I’m not a scientist by any means, but am adapting an abstract to the public transit system in NYC. I will publish the project on my website when it’s done. I think it’s a few weeks away. I built the dataset, now doing experimental model training. If I can get acceptable accuracy, I will deploy in a production system and build a UI.

Here is a scientific abstract that inspired my to start building this system. -> https://arxiv.org/html/2510.03121

I am unfamiliar with agent based models, sorry I can’t offer any personal insight there, but I ran your question through Gemini and here is the AI response:

Based on the scientific abstract of the paper *"Real Time Headway Predictions in Urban Rail Systems and Implications for Service Control: A Deep Learning Approach"* (arXiv:2510.03121), agent-based models (ABMs) and deep learning (DL) approaches compare as follows:

### 1. Computational Efficiency and Real-Time Application

* *Deep Learning (DL):* The paper proposes a *ConvLSTM* (Convolutional Long Short-Term Memory) framework designed for high computational efficiency. It is specifically intended to provide real-time predictions, enabling dispatchers to evaluate operational decisions instantly. * *Agent-Based Models (ABM):* While the paper does not use ABMs, it contrasts its DL approach with traditional *"computationally intensive simulations"*—a category that includes microscopic agent-based models. ABMs often require significant processing time to simulate individual train and passenger interactions, making them less suitable for immediate, real-time dispatching decisions during operations.

### 2. Modeling Methodology

* *Deep Learning (DL):* The approach is *data-driven*, learning spatiotemporal patterns and the propagation of train headways from historical datasets. It captures spatial dependencies (between stations) and temporal evolution (over time) through convolutional filters and memory states without needing explicit rules for train behavior. * *Agent-Based Models (ABM):* These are typically *rule-based and bottom-up*, modeling the movement of each train "agent" based on signaling rules, spacing, and train-following logic. While highly detailed, they require precise calibration of individual agent parameters.

### 3. Handling Operational Control

* *Deep Learning (DL):* A key innovation in this paper is the direct integration of *target terminal headways* (dispatcher decisions) as inputs. This allows the model to predict the downstream impacts of a specific control action (like holding a train) by processing it as a data feature. * *Agent-Based Models (ABM):* To evaluate a dispatcher's decision in an ABM, the entire simulation must typically be re-run with new parameters for the affected agents, which is time-consuming and difficult to scale across an entire metro line in real-time.

### 4. Use Case Scenarios

* *Deep Learning (DL):* Optimized for *proactive operational control* and real-time decision-making. It is most effective when large amounts of historical tracking data are available to train the spatiotemporal relationships. * *Agent-Based Models (ABM):* Often preferred for *off-line evaluation* of complex infrastructure changes, bottleneck mitigation strategies, or microscopic safety analyses where the "why" behind individual train behavior is more important than prediction speed.


I just started building my own website today with Django. I’m doing it because I just enjoy doing it. Most of my work is in data and ML infrastructure and it is just killing me. Working on the front end has opened my mind to possibility and given me new inspiration.

I love hn and was inspired by all the devs who have their own site. I was drowning in work, but put the Django architecture together on vacation, started putting things together today and it’s been a blast.

I don’t enjoy social media and was thinking to posse intrinsically.

I appreciate this post and the authors perspective.


What features did you want for your personal site that lead to choosing Django (or a backend framework at all) instead of a static site generator?

SSGs are good for static sites with no interactivity or feedback. If you want interactivity or feedback, someone (you or a 3rd party service provider) is going to have to run a server.

If you're running a server anyway, it seems trivial to serve content dynamically generated from markdown - all an SSG pipeline adds is more dependencies and stuff to break.

I know there's a fair few big nerd blogs powered by static sites, but when you really consider the full stack and frequency of work that's being done or the number of 3rd party external services they're having to depend on, they'd have been better by many metrics if the nerds had just written themselves a custom backend from the start.


I just wanted to learn how to create an enterprise grade web application. I read a book on Django last year and did a few tutorials and enjoyed it. I also deploy infra on gcp and it works well there. It cost about $60/month for baseline hosting with light traffic/storage. I will probably use it for an interface for some of my ml projects. I was also looking into dart/flutter a much steeper learning curve for me personally.

This is pretty much how I began developing websites too. Except it was 2001 instead of 2026. And it was ASP (the classic ASP that predates ASP.NET) instead of Python. And I had a Windows 98 machine in my dorm room with Personal Web Server (PWS) running on it instead of GCP.

It could easily have been a static website, but I happened to stumble across PWS, which came bundled with a default ASP website. That is how I got started. I replaced the default index.asp with my own and began building from there. A nice bonus of this approach was that the default website included a server-side guestbook application that stored comments in an MS Access database. Reading through its source code taught me server-side scripting. I used that newfound knowledge to write my own server-side applications.

Of course, this was a long time ago. That website still exists but today most of it is just a collection of static HTML files generated by a Common Lisp program I wrote for myself. The only parts that are not static are the guestbook and comment forms, which are implemented in CL using Hunchentoot.


I remember ASP (application service provider, before cloud became synonymous with hosting), you are making me nostalgic. Back then I was in sales, I was selling real time inventory control, CRM and point of sale systems distributed over Citrix Metaframe in a secure datacenter. Businesses were just starting to get broadband connections. I would have to take customers to the datacenter to motivate them to let us host their data. Eight years later, google bought the building for $1.8b and eventually bought adjacent buildings as well.

We are talking about different ASPs. I am referring to Active Server Pages (ASP), the server-side scripting language supported by Personal Web Server (PWS) and Internet Information Services (IIS) on Windows. It is similar to PHP Hypertext Processor (PHP) and Java Server Pages (JSP) but for the Windows world. I began developing websites with ASP. Over the years, I dabbled with CGI, PHP, JSP, Python, etc. before settling on Common Lisp as my preferred choice for server-side programming.

Got it! It's amazing how many languages are out there... very interesting

$60/mo for a personal website is insane.

I agree. To be more clear, that $60 is an estimate for a small configuration and includes serverless infrastructure to process 500,000 requests per month, plus storage, including a 20gb sql database and 100gb of object storage to serve video and images. More ideal for an application. You run the app in a container and only get charged for the requests, the sql database is persistent, so that cost $20/month and object storage with egress is about $10/month.

Let me describe my setup, so that you can compare. I use a Contabo VPS for around 5 USD month to host my Wagtail (django-based) site. The DB also runs on the same infra and since it's SQLite I can back it up externally.

I probably wouldn't be able to handle 0.5M requests, but I am nowhere near getting them. If I start approaching such numbers I'll consider an upgrade.

Check out Wagtail if you'd like to have even more batteries included for your site, it was a delight building my site with it:

https://blog.miloslavhomer.cz/hello-wagtail/


Thank you for sharing your setup, I will certainly examine it and compare a bit later. I know my setup is a bit over the top, but it is the easiest to learn, since I live in gcp everyday. I certainly don't expect the .5m traffic, but that is one of the lower tiers for cloud run, serverless execution service. This is just a poc to get my fingers dirty with the MVT pattern.

Gotcha. Yes, with just a VPS you have to do a lot of busywork to get online - DNS, reverse proxy, docker, dev environment, DB setup and others.

I'd still recommend starting with SQLite, seems that by skipping a DB service you can save quite a few bucks.


Many good observations here. I had time to read 50% through.

>> I think I’m particularly suspicious of community, because as a writer and pedantic arsehole on the internet, I value truth-seeking behaviour. I want people to think and say things that are true, not just things that they have to believe for the sake of keeping their community happy.

Unfortunately, this is what happens with every group of people.

Our individual realities are highly subjective. A group of people who are part of a community construct a shared reality that they can all accept. If you don’t contribute to the shared reality, you are treated as someone who is problematic.

As humans we are social creatures. In our evolution, we develop cognitive systems that help us thrive in social structures. One system is called the social protection system. This system gets activated when we sense tension in relationships and sends a signal of fear to the subject that they risk being separated from a social group. This fear motivates people to maintain connection. So some people are intrinsically motivated by fear to maintain their status, sometimes unconsciously.

Our self esteem comes from two things, relationships and mastery. Healthy self esteem comes from connection to people who accept you for who you are, where you feel visible and accepted with your good and bad traits.

If you have a few people in your life with this type of connection, you will have a healthy social foundation and rely less on belonging to a group.

Groups are valuable in that the human experience is complicated. The best source of information comes directly from other humans and their experiences overcoming complexity.

However, I do agree with the author where certain groups can be problematic, particularly engaging in things like tribalism.

Establishing good self esteem by keeping a few people close to you who see you and accept you as a flawed human is key. The other part is to immerse yourself in activities where you develop mastery and maintain a connection to the activities that are intrinsically motivating and satisfying without distraction from external signals.

I learned this by studying the science of self actualization, from the research done by Scott Barry Kaufman and his book Transcend. He’s a humanistic psychologist who was inspired by Abraham Maslow, one of the founders of humanistic psychology.


did warren buffets success help america or just his shareholders? How does the $350 billion cash concentration affect the economy and his businesses? I would think it would be better served redistributing it back into the economy to create opportunities for working class, so they can make a living wage, vs having multiple jobs and no savings or safety net.

Ways it helps America:

* His shareholders include many Americans. The cash pile is concentrated as far as corporate governance is concerned, but lots of people own a piece of it in their retirement accounts. (Often indirectly via index funds.)

* Many employees work for businesses owned by Berkshire Hathaway. I'd guess they're pretty stable businesses, not about to get bought out and shut down? Maybe decent places to work? (Not sure about the railroad, though.)

* Many people are customers of these businesses. Mostly a good thing?

Admittedly there are lots of Americans that don't participate in that, though.


Learning to adapt to stress without unhealthy vices is a skill.

Finding healthy ways to blow off steam maintains continuity with your most important goals.

Well done!


Cynicism is something I was curious about, particularly what drives people to feel this way. It’s such a powerful emotion that tilts someone’s worldview.

It comes from negative experience and not dealing with the effects of that experience. The weight of that negativity is what tilts your views to be overtly cynical. You don’t just choose to be cynical, it’s part of your predisposition.

Idealism comes when you are younger in age and as you get older, you become more cynical about things, because you have been through many more experiences.

Young people should not be overtly cynical. They should look at the world with bright eyes and try to change things for the better. A young person who is overly cynical is a tragedy.

In regard to playing politics, I think that is just intellectual laziness. Getting people on board with your ideas requires thoughtfulness. Try to find common ground on something you both believe in is a challenge, that requires effort. Treating people as individuals, showing genuine curiosity in their beliefs and exchanging direct feedback in a respectful way, is how you get people on board with your ideas.


Cynicism in younger people is merely operationalizing the adage “prepare for the worst and hope for the best”. Hope just isn’t a great foundation for basing decisions on if you have any other foundation available.

Hope is a great foundation, especially if you’re in a negative environment. It’s important to deal directly with reality, but sometimes reality is harsh and you need a break from the present moment. Hope is that break.

If you have nothing else sure. Otherwise build your life on reality not on sandcastles in the clouds.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: