Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I haven't done benchmarking yet (plan to do them), but it should be similar to our post on DeepSeek-V3.1 Dynamic GGUFs: https://unsloth.ai/docs/basics/unsloth-dynamic-2.0-ggufs

> certainly not bubblewrap,

Eh, it might be bubblewrap given it's what flatpak uses.


The post you are responding to discusses Stablecoin regulation multiple times.

yeah, I agree. This is a very cool idea but the visual design of the cards needs some tightening up

> But I think that specifically the way Musk is trying to position it, the moon would be an even harder sell.

I agree. I would be quite a moonshot.


What is not performant about tokio? Do you know a better async runtime? I also heard tokio's "multi-thread" scheduler had some issues.

That’s not an Anthropic problem, that’s a problem with whomever you work for.

What is absurd about that?


After ACH, which I’m assuming you got for free, US to Poland on Wise is still multiple orders of magnitude more expensive.

I'm really amazed at czr exchange functionality. It's really great.

Nice! Yes Q8_0 is similar - the others are different since they use a calibration dataset.

Someone will have to file suit, as is tradition under this administration.

Despite still not really showing any utility these tech companies want so so so much for cryptocurrency to catch on.

It feels like the entirety of cryptocurrency, outside of being a thing people used to buy drugs, has been an example of Chesterton's Fence, with half of Silicon Valley in denial of this fact.


Running llama.cpp rather than vLLM, it's happy enough to run the FP8 variant with 200k+ context using about 90GB vram

And this, ladies and gentlemen, is why SaaS investors don’t understand how to invest in deeptech/hardtech, despite current trends. Like this guy, they have no clue about the differences in business model, except they’re not founders so they don’t go through the pain and they mostly don’t learn.

Hats off to the author for making it through! What a start to the journey!


The ISS is designed to emit 126kW of heat radiation between the active cooking systems and the solar array cooling system.

Probably collecting application fees from people interested in renting it.

Except it's not because it's constantly ambiguous in computing.

E.g. Macs measure file sizes in powers of 10 and call them KB, MB, GB. Windows measures file sizes in powers of 2 and calls them KB, MB, GB instead of KiB, MiB, GiB. Advertised hard drives come in powers of 10.


Steady, strong, and reliable.

Yeah :c I feel the same way. They’ve made a variant with more traditional poker deck look but the same rank/suits of the ever deck that I’m excited to try one day

Engine bays have a lot of design go into where to keep heat and where to get rid of it. You can look up thermal coatings and ceramics etc.

Can you smoke ketamine?

Europe? Uh, I think you mean the whole world.

this is almost certainly not being done on cerebras

is this the beginning of the end?

Right. The alternative is not to send materials from Earth for processing in space, that would be stupid. We send finished stuff, which were manufactured on the ground. But you don’t mine finished widgets from asteroids. You mine ore that needs refining and processing before being used to manufacture things. This ore is orders of magnitude heavier than the finished products, never mind all that’s required to do anything useful with it.

RNNs have two huge issues: - long context. Recurrence degrades the signal for the same reason that 'deep' nn architectures don't go much past 3-4 layers before you need residual connections and the like - (this is the big one) training performance is terrible since you can't parallelize them across a sequence like you can with causal masked attn in transformers

On the huge benefit side though you get: - guaranteed state size so perfect batch packing, perfect memory use, easy load/unload from a batch, O(1) of token gen so generally massive performance gains in inference. - unlimited context (well, no need for a concept of a position embedding or similar system)

Taking the best of both worlds is definitely where it is at for the future. An architecture that can train parallelized, has a fixed state size so you can load/unload and patch batches perfectly, unlimited context (with perfect recall), etc etc. That is the real architecture to go for.


Czr is nice and really work for me

> I don't know how sleep/wake works from a text mode TTY machine if that's what you mean.

Usually, I expect, by just issuing a command... e.g. `systemctl suspend`


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: