Email or username:

Password:

Forgot your password?
rayslava

During the weekend I've been fighting with ChatGPT and Claude in order to create a simple TUI using ratatui for "windows" and sm for keeping state. And neither of them made something useful, even both of them correcting one another could not keep the logic construction big enough to track the state, update it appropriately and reflect the transitions on the screen.

At last I tired and wrote it manually.

So #LLM is a good tool but not a universal one even in the hype field. Guess, it's still a little bit early for retirement if you do something more complex than generic web-service.

However the boilerplate code and tests were created fast and without much effort (not tests for the state machine itself though), so the tools are useful and help you to get rid of some routine.

TLDR: we still have the most weird programming for us, my fellow developers!

8 comments
a1ba@$INSTANCE$host$
@rayslava have you tried that new deepseek?

My favorite test with them is sending them decompiled code and asking to explain what does it do. So far, it at least figured out what it does.
rayslava

@a1ba tried an 14B version locally—it worked much better than codellamas for simple cases. Do they have a cloud version with large model too?

rayslava

@a1ba Okay 😀
Will check a bit later then, sounds interesting.

Alexey Skobkin

@rayslava @a1ba
It's no surprise since Codellama is based on LLaMA 2. It's so old now. Even LLaMA 3.0 looks bad if compared to LLaMA 3.1 or 3.3 🤷‍♂️

burbilog

@rayslava @a1ba No, you didn't. 14b means that you have tried Qwen 14b, distilled by Deepseek R1 (the reasoning one). It became much better than original Qwen, but still that's not Deepseek.

Unless you have an insane amount of VRAM you can't run Deepseek V3 or R1.

They really scewed model naming, pubilishing other distilled models as subversions of Deepseek itself.

rayslava

@burbilog @a1ba Okay, I tried the chat.deepseek.com and it's the whole new level.
It still wasn't able to give me working code but it was much (like MUCH) closer to what I wanted initially.

Given this model is released to open source, I can only join to all the people mentioning the new stage in LLM race.

burbilog

@rayslava @a1ba Yeah, it is much better now. But please, don't call it open source. They did not publish the SOURCE dataset, only the BINARY model. We don't call freeware .exe files open source, do we?

Go Up