Hacker Newsnew | past | comments | ask | show | jobs | submit | dingocat's commentslogin

> “Blog” stands for “web log”. If it’s on the web, it’s digital, there was never a period when blogs were hand written.

Did you use AI to write this...? Because it does not follow from the post you're replying to.


Read it again. I explicitly quoted the relevant bit. It’s the first sentence in their last paragraph.


What do you mean there is no such thing as R1-1.5b? DeepSeek released a distilled version based on a 1.5B Qwen model with the full name DeepSeek-R1-Distill-Qwen-1.5B, see chapter 3.2 on page 14 of their research article [0].

[0] https://arxiv.org/abs/2501.12948


Which is not the same model, it's not R1 it's R1-Distill-Qwen-1.5B....


A distinction they make clear and write extensively about on the model page, yes?


wheres that made clear in "ollama run deepseek-r1” the command to download/run the model?


Which you have to go to the model page to find.


ollama labels the qwen models R1, while the "R1" moniker standing on its own in deepseek world means the full model that has nothing to do with qwen.

https://ollama.com/library/deepseek-r1

That may have been ok if it was just same model at different sizes but they're completely different things here & it's created confusion out of thin air for absolutely no reason other than ollama being careless.


And their documentation makes that distinction clear, having dedicated a section specifically to the distilled models.


It hasn't been great. Applying for jobs in this market while having the post-MSc burnout has been tough. I finished an MSC in AI back in June, but there are so few AI-relevant jobs where I live it's depressing. I've interview for data science and analyst roles, but barely get in the door.

I'm getting a good amount of interviews for developer and data engineering positions, but the competition is tough. Many positions have seen a 5x increase in candidates since the same time last year, according to my interviewers.

However, I'm hopefully getting an offer as a data platform engineer soon. The department leader has ranked me as their first choice, so unless the higher-ups complain... Knock on wood.


I have multiple questions regarding the methods of this test.

The biggest one is that, well... The test doesn't aim to see what GPT-4 can do and how well it does it, only whether the participant can guess the (possibly cherry-picked) answer the author decided on. In short, we don't know if he sampled answers and decided on the most probable answer (akin to consensus voting/self-consistency[1]), or if he asked a question and chose the first one.

Maybe GPT-4 guesses the correct answer for a question 80% of the time, but he got unlucky? You don't know, the author doesn't tell you. The answers are generated ahead of time and are the same every time you go through the test.

[1] https://doi.org/10.48550/arXiv.2203.11171


> the [ ... ] answer the author decided on

The questions mostly have correct or incorrect answers, and where there is some leeway, the author provides a fairly detailed explanation of what they would consider correct in each case. Do you have some specific criticism of an answer that you believe the author gets wrong?


> only whether the participant can guess the (possibly cherry-picked) answer the author decided on

My understanding is that the quiz samples a new GPT-4 answer every time you use it. That's why you put a confidence rather than a 0%/100% answer. There's always a chance it'll fail by freak accident.


If you're basing this on the animation used when revealing the answer, that's a fake effect. The source code[0] reveals that there's a typewriter effect that plays out when you select to answer the question.

Also, the commentary on the answers refers to specific parts of the answers. For it to be as in-depth as it is, it would have to be either pre-written or the commentary also generated by GPT on the fly. (And of course it wouldn't make sense to do that given the nature of the quiz.)

[0] https://nicholas.carlini.com/writing/llm-forecast/static/que...


A few pages? You barely need a single sentence. "Build lasting habits by recognizing how triggers lead to action."


Yeah, one of the problems with being too abstract is that it's open to many interpretations, most of them incorrect. Also, a few pages or paragraphs won't stick; I need examples. I'm not a computer; installing new ideas in my brain isn't like installing apps. My brain prefers storytelling to better grasp ideas and make them its own.


Agreed, but many self-help books either waste their time by making dubious claims from half-sourced research, using useless anecdotes, or repeating the problem in different ways. The best ones actually tell stories of different ways of using its method to solve a variety of problems.


Absolutely, that contextualization can be so useful even if some deem it fluffy. Also no one is obligated to read every single word of a book, if you don't want the anecdotes and case studies, you can skip them! They are helpful for the rest of us.


Ok, but a few sentences explaining what each of those terms mean would be ideal.


Ditto, I can do it as well. I can best describe it as looking at an object and shifting it in/out of focus.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: