Hacker Newsnew | past | comments | ask | show | jobs | submit | est's commentslogin

China had many >1M KM electric taxies. Usually they degrade after 200,000 km, and they are like 2 generations behind the latest ones.

Now someone please revive Microsoft Encarta ...


Wow. I’m curious as to why so much spam.

installed since last HN post. So Bonsai (1-bit) and Ternary-Bonsai are different?

Can it be run on browsers with WASM/WebGPU?



> I look at the list of contributors

Specifically if those avatars are cute animie girls.


> Specifically if those avatars are cute anime girls.

I know you are half joking/not joking, but this is definitely a golden signal.


Positive or negative to you? Whenever I see more than one anime-adjacent profile picture I duck out.

Positive ofc, most of them a top-tier Rust devs.

like 1973 oil crisis? End of V8 engines (pun intended)

Yeah, like that. Modern engines eclipsed pre-1970s engines in performance, efficiency, and even (yes I said it) in reliability.

At a cost of simplicity and beauty. And two lost decades of mediocre performance. Sigh


> how homogenous the internet has become in terms of design

I think it's because Steve Jobs killed Flash.


> bootstrap an Artifacts repo from an existing git repository

wow that's cool. I used to hack CF Worker to operate .git using isomorphic-git, it's a PITA.

> ArtifactFS runs a blobless clone of a git repository: it fetches the file tree and refs, but not the file contents. It can do that during sandbox startup, which then allows your agent harness to get to work.

That's insanely useful. How combining with only committing the file changed we'd have a single-blob editing capability against any .git


I really want to know what does M, K, XL XS mean in this context and how to choose.

I searched all unsloth doc and there seems no explaination at all.


Q4_K is a type of quantization. It means that all weights will be at a minimum 4bits using the K method.

But if you're willing to give more bits to only certain important weights, you get to preserve a lot more quality for not that much more space.

The S/M/L/XL is what tells you how many tensors get to use more bits.

The difference between S and M is generally noticeable (on benchmarks). The difference between M and L/XL is less so, let alone in real use (ymmv).

Here's an example of the contents of a Q4_K_:

    S
    llama_model_loader: - type  f32:  392 tensors
    llama_model_loader: - type q4_K:  136 tensors
    llama_model_loader: - type q5_0:   43 tensors
    llama_model_loader: - type q5_1:   17 tensors
    llama_model_loader: - type q6_K:   15 tensors
    llama_model_loader: - type q8_0:   55 tensors
    M
    llama_model_loader: - type  f32:  392 tensors
    llama_model_loader: - type q4_K:  106 tensors
    llama_model_loader: - type q5_0:   32 tensors
    llama_model_loader: - type q5_K:   30 tensors
    llama_model_loader: - type q6_K:   15 tensors
    llama_model_loader: - type q8_0:   83 tensors
    L
    llama_model_loader: - type  f32:  392 tensors
    llama_model_loader: - type q4_K:  106 tensors
    llama_model_loader: - type q5_0:   32 tensors
    llama_model_loader: - type q5_K:   30 tensors
    llama_model_loader: - type q6_K:   14 tensors
    llama_model_loader: - type q8_0:   84 tensors

They are different quantization types, you can read more here https://huggingface.co/docs/hub/gguf#quantization-types

Just start with q4_k_m and figure out the rest later.

hey you can do a bit research yourself and tell your results to us!

Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: