WheelsAtLarge 3 days ago

The LLMs in general are already having an impact. The price per line of code has and is falling. Anyone who won't use them to their advantage will be left behind. Deepseek is just one tool that coders will need to use to write code.

Here's a recent example of a similar situation. There was a time when all web pages were written by hand using a text editor. As time went on, tools began to appear that would automate much of the writing. Yes, the code they provided was ugly, but it did the job. Eventually, most developers started to use the tools simply because they could get more done.

I suspect the same will happen with software engineers.

  • muzani 2 days ago

    Prices per line of code have always been falling. Before AI, it was JS frameworks. But it led to an exponential increase in the number of software engineers.

    You still see a lot of frustration in working with AI today, and eventually people will just start hiring for this the way they're hiring React Native and Cordova developers.

PaulHoule 3 days ago

At this kind of workplace, yes

https://news.ycombinator.com/item?id=42833695

It will affect the ability of that kind of workplace to deliver... anything.

  • meiraleal 3 days ago

    Wouldn't it also shift the spending from hardware back to software? As they say, software eats the world

    • PaulHoule 3 days ago

      When ChatGPT-3 came out my first take was that we had no idea how those things worked, but once the research happened, people were going to learn to reduce the cost of the models dramatically and that's what happened.

      The approach where "the model has a huge amount of world knowledge and can speak off the cuff about any subject" would inevitably fall to "it can use a search engine, read the results, and then reason based on the content of the results", not least because people are always going to want to talk about, say, the Bills vs Chiefs game that happened last night and we can't retrain the whole model every day.

      Myself I like to train models that are so simple that I can train 20 in 3 minutes and take the best. I can quickly do experiments to understand how the models. If I had a model that took 30 minutes to train and had a reliable training process I could use that for my purpose, but as a side project I don't have the time to do large numbers of experiments to get there if my model takes that long to train.

      So for all those reasons it seems we should have a 'minimal' model that is good at reasoning. The faster the model is to train and do inference on the faster people can do experiments and make it better.

      The outcry people have that AI models need to steal increasing amounts of content for training might diminish (I hope so!) but people might have the realization that, if you want to research a topic, you could send a webcrawler off to look at 1 million documents and have an AI process them for you (I'm living that dream!)