• 0 Posts
  • 45 Comments
Joined 2 years ago
cake
Cake day: June 16th, 2023

help-circle









  • It’s strongly dependent on how you use it. Personally, I started out as a skeptic but by now I’m quite won over by LLM-aided search. For example, I was recently looking for an academic that had published some result I could describe in rough terms, but whose name and affiliation I was drawing a blank on. Several regular web searches yielded nothing, but Deepseek’s web search gave the result first try.

    (Though, Google’s own AI search is strangely bad compared to others, so I don’t use that.)

    The flip side is that for a lot of routine info that I previously used Google to find, like getting a quick and basic recipe for apple pie crust, the normal search results are now enshittified by ad-optimized slop. So in many cases I find it better to use a non-web-search LLM instead. If it matters, I always have the option of verifying the LLM’s output with a manual search.



  • Pretty much inevitable. Nowadays there are so many robot vacuum cleaners from different brands, and everyone has more or less figured out the tech so they all work pretty well. (I have a Roborock, and have nothing to say about it other than it keeps the floors clean and doesn’t cause me any grief.) There’s no moat, so consumer market success is purely a matter of manufacturing and cost efficiency, and iRobot obviously would have a huge upfill fight against Samsung, Xiaomi, and a thousand other light consumer goods makers.






  • Aww come on. There’s plenty to be mad at Zuckerberg about, but releasing Llama under a semi-permissive license was a massive gift to the world. It gave independent researchers access to a working LLM for the first time. For example, Deepseek got their start messing around with Llama derivatives back in the day (though, to be clear, their MIT-licensed V3 and R1 models are not Llama derivatives).

    As for open training data, its a good ideal but I don’t think it’s a realistic possibility for any organization that wants to build a workable LLM. These things use trillions of documents in training, and no matter how hard you try to clean the data, there’s definitely going to be something lawyers can find to sue you over. No organization is going to open themselves up to the liability. And if you gimp your data set, you get a dumb AI that nobody wants to use.


  • It’s definitely a trend. More and more top Chinese students are also opting to stay in China for university, rather than going to the US or Europe to study. It’s in part due to a good thing, i.e. the improving quality of China’s universities and top companies. But I think it’s a troubling development for China overall. One of China’s strengths over the past few decades has been their people’s eagerness to engage with the outside world, and turning inward will not be beneficial for them in the long run.



  • Base models are general purpose language models, mainly useful for AI researchers and people who want to build on top of them.

    Instruct or chat models are chatbots. They are made by fine-tuning base models.

    The V3 models linked by OP are Deepseek’s non-reasoning models, similar to Claude or ChatGPT4o. These are the “normal” chatbots that reply with whatever comes to their mind. Deepseek also has a reasoning model, R1. Such models take time to “think” before supplying their final answer; they tend to give better performance for stuff like math problems, at the cost of being slower to get the answer.

    It should be mentioned that you probably won’t be able to run these models yourself unless you have a data center style rig with 4-5 GPUs. The Deepseek V3 and R1 models are chonky beasts. There are smaller “distilled” forms of R1 that are possible to run locally, though.