mstdn.io is one of the many independent Mastodon servers you can use to participate in the fediverse.

Administered by:

Server stats:

382
active users

To be completely fair, thread safety and atomics are advanced topics.

Several humans I have interviewed for engineering positions would also have a lot of trouble answering these questions. I couldn't write this code on a whiteboard without looking at the Rust library docs.

The main problem here is that the model is making up poor excuses to justify Arc<AtomicUsize>, showing poor reasoning skills.

Larger models like should do better with my questions (haven't tried yet).

Google's Gemini Pro performs even worse than the opensource models running on my modest Linux desktop:

g.co/gemini/share/cdec7f5a6c5c

Missing from the chat log, is the last response in the image below 🤦‍♂️

I don't have Gemini Advanced / Ultra. Is it a bit smarter than this?

#google#gemini#llm

Today I tried running Codestral, a 22B parameter LLM tuned for coding by Mistral AI.

With my Rust mock interview questions, it performed better than all other offline models I tried so far.

paste.benpro.fr/?4eb8f2e158416

#coding#rust#llm

My AMD GPU running Codestral, a 22B parameter LLM.

The gaps in resource usage occur when the model is waiting for the next prompt from me.

With this setup, a response of 543 tokens took about 14.3 seconds (38 tokens/s).

@codewiz I turned #Gemini on, on my pixel. Got in the car today and asked Google "navigate to an address" Gemini told me I can use a map app on my phone to do that. I promptly uninstalled Gemini

@Jfillian lol 🙃

Also in my experience, the new Gemini assistant is super dumb.

@codewiz What's the memory size of that amd gpu?

@codewiz where is the console_log macro coming from in that wasm hello world? I think it's using rustwasm.github.io/wasm-bindge as a base, but without defining the macro.

@codewiz Damn, that is good. The way it offers the type annotated version of the example when you asked for the type of one variable is also really nice.

@penguin42 Check my last post on Codestral.

Hard to tell if it's better or worse than GPT-4o...

...it typically takes twenty to thirty cross-referenced questions to detect a Nexus-60B model 😆

@penguin42

@codewiz Anything that can remember Rust macro syntax from memory must be a replicant....

@codewiz Heck, it's not bad though! Atomics isn't an easy subject (although that's an easy case). Curious it's decided to do the compare-exchange trick on one side and a fetch_sub on the other.