Hacker Newsnew | past | comments | ask | show | jobs | submit | Artgor's commentslogin

https://andlukyane.com/ - I write paper reviews, share experience of working in ML and my language learning journey.


I sympathise. I did ~200k reviews in Anki and no idea how many in Renshuu in 2025 for language learning: German, Spanish and Japanese... and I think I got into Anki hell. For a long time, Anki was really useful for me, it pushed my Spanish and German forward, but now I plan to decrease the number of reviews significantly. I hope to spend no more than 30 minutes per day of flashcards, and the rest of time on immersion.


How do you decide what to immerse yourself in? Do you just search for things independently or do you have a way of selecting content based on your level?


I prefer to start immersion at ~B1 level. I know some people want to do comprehensive input right from the start, but I prefer to build foundations. I start with popular books that I have already read before - The Little Prince, Harry Potter, etc. Then I take books that are interesting for me and work through them. There are graded readers, but the stories are usually very boring. I prefer to read what I like.


You know, sometimes I feel that all this discourse about AI for coding reflects the difference between software engineers and data scientists / machine learning engineers.

Both often work with unclear requirements, and sometimes may face floating bugs which are hard to fix, but in most cases, SWE create software that is expected to always behave in a certain way. It is reproducible, can pass tests, and the tooling is more established.

MLE work with models that are stochastic in nature. The usual tests aren't about models producing a certain output - they are about metrics, that, for example, the models produce the correct output in 90% cases (evaluation). The tooling isn't as developed as for SWE - it changes more often.

So, for MLE, working with AI that isn't always reliable, is a norm. They are accustomed to thinking in terms of probabilities, distributions, and acceptable levels of error. Applying this mindset to a coding assistant that might produce incorrect or unexpected code feels more natural. They might evaluate it like a model: "It gets the code right 80% of the time, saving me effort, and I can catch the 20%."


This has been about 50% of the time my experience as well. There are very good SWE who know how to use ML in real systems, and then there are the others who believe through and through it will replace well understood systems developed by subdomain experts.

As a concrete example, when I worked at Amazon, there were several really good ML-based solutions for very real problems that didn't have classical approaches to lean on. Motion prediction from grid maps, for example, or classification from imagery or grid maps in general. Very useful and well integrated in a classical estimation and control pipeline to produce meaningful results.

OTOH, when I worked at a startup I won't name, I was berated over and over by a low-level manager for daring to question a learning-based approach for, of all things, estimating orientation of a stationary plane over time. The entire control pipeline for the vehicle was being fed flickering, jumping, adhoc rotating estimate for a stationary object because the entire team had never learned anything fundamental about mapping or filtering, and was just assuming more data would solve the problem.

This divide is very real, and I wish there was a way to tease it out better in interviewing.


The lack of knowledge of and application of fundamental engineering principles is a huge issue in the Software world. While its great that people can pick up programming and learn and get a job, I have noticed this is often correlated with people not having a background in Hard Science and Mathematics. Even amongst CS graduates there are a lot of who seem to get through without any mathematical or engineering maturity. Having a couple people in a team with Physics, Mathematics, Mechanical or Electrical Engineering backgrounds, etc. can really be a big asset as they can fight back and offer a classical solution that will work nearly 100% of the time. Whereas someone who just did a Bootcamp and no formal scientific training seems less likely to be able to grasp or have prior knowledge of classical approaches.

I think that this is one reason Software has such a flavor of the month approach to development.


Your stationary plane example highlights a divide I've seen across my work experience in different domains; teams defaulting to ML when fundamental engineering would work better.

I'm curious: do you think there's any amount of high-quality data that could make the learning-based approach viable for orientation estimation? Or would it always be solving the wrong problem, regardless of data volume and delivery speed?

My sense is that effective solutions need the right confluence of problem understanding, techniques, data, and infrastructure. Missing any one piece makes things suboptimal, though not necessarily unsolvable.


> I've seen across my work experience in different domains; teams defaulting to ML when fundamental engineering would work better.

In my current field (predictive maintenance), there are (in)famous examples and papers using multi-layer deep networks for solving anomaly detection problems, where a "single" line of basic Matlab code (standard deviations, etc.) performs better than the proposed AI solution. Publish or perish, I guess...


One opportunity for human-designed systems to excel over machine learning is the case where treating ML as a black box has caused the designers to pose an impossible problem. From the parent comment, it sounds like each additional measurement was being related to a new estimate by the ML system, while the standard technique could integrate measurements over time (that's called filtering).


> So, for MLE, working with AI that isn't always reliable, is a norm. They are accustomed to thinking in terms of probabilities, distributions, and acceptable levels of error. Applying this mindset to a coding assistant that might produce incorrect or unexpected code feels more natural. They might evaluate it like a model: "It gets the code right 80% of the time, saving me effort, and I can catch the 20%."

And given the current climate, the MLE's feel empowered for force their mindset onto others groups where it doesn't fit. I once heard a senior architect at my company ranting about that after a meeting: my employer sells products where accuracy and correctness have always been a huge selling point, and the ML people (in a different office) didn't seem to get that and thought 80-90% correct should be good enough for customers.

I'm reminded of the arguments about whether a 1% fatality rate for a pandemic disease was small or large. 1 is the smallest integer, but 1% of 300 million is 3 million people.


This is where I find having a disconnect between an ML team and product team is so broken. Same for SE to be fair.

Accuracy rates, F1, anything, they're all just rough guides. The company cares about making money and some errors are much bigger than others.

We'd manually review changes for updates to our algos and models. Even with a golden set, breaking one case to fix five could be awesome or terrible.

I've given talks about this, my classic example is this somewhat imagined scenario (because it's unfair of me to accuse people of not checking at all):

It's 2015. You get an update to your classification model. Accuracy rates go up on a classic dataset, hooray! Let's deploy.

Your boss's, boss's, boss gets a call at 2am because you're in the news.

https://www.bbc.co.uk/news/technology-33347866

Ah. Turns out improving classifications of types of dogs improved but... that wasn't as important as this.

Issues and errors must be understood in context of the business. If your ML team is chucking models over the fence you're going to at best move slowly. At worst you're leaving yourself open to this kind of problem.


[flagged]


The GP wasn't making a political argument just pointing out statistics.


[flagged]


> oh yes he was, it was just thinly veiled and apparently the audience doesn't like it being pointed out.

You don't know what you're talking about.

It appears I hit one of your triggers, and you don't seem to have the self-control to not read into my comment stuff that isn't there.

> If he wanted to say something about statistics he could've picked anything besides COVID.

I picked the handiest example, that's it. Do you think I should walk on eggshells because of you, internet stranger?

> Don't gaslight me about gross virtue signaling masked as intellectualism. I'm a lot of things but stupid isn't one of them.

If you talk like that, I'm not so sure.


You're talking about deterministic behavior vs. probabilistic behavior and yes some discourse lines up with what you describe.

I don't think it's the case with this article. It focuses on the meta-concerns of people doing software engineering and how AI fits into that. I think he hits it on the head when he talks about Program Entropy.

A huge part of building a software product is managing entropy. Specifically, how you can add more code and more people while maintaining a reasonable forward velocity. More specifically, you have to maintain a system so you make it so all of those people understand how all the pieces fit together and how to add more of those pieces. Yes, I can see AI one day making this easier but right now, it oftentimes makes entropy worse.


There are so many use cases where 90% correct answers are absolutely not enough. Nobody would have much of a problem with that, if a flurry of people with vested interests wouldn't try to convince us all that that is not the case, and AI is good to go for absolutely everything. The absurdity of this assumption is so outrageous that it becomes even hard to counter it with logic. It's just a belief-based narrative whose delivery has been highly successful so far in commanding insane investments, and as a travestee for profit-oriented workforce optimizations.


Who is actually saying that AI is always 100 percent right?

There are disclaimers everywhere.

Sure there are usecases AI can't handle, but doesn't mean it is not massively valuable. There is not single thing in the World that can handle all usecases.


SWE use probability all the time. Rearchitect around that race condition or reduce its footprint? How long will this database call make, p99? A/B tests. Etc.


The bigger the system, the more bigger the probability aspect gets too. What are the chances of losing at the data copies at the same time, what are the chances is all slots being full and the new connection dropped, what are the chances of data corruption from bitrot? You can mostly ignore that in toy examples, but at scale you just have chaos and money you can throw at it to reduce it somewhat.


I’m currently getting my masters in AI (lots of ML) and as a SWE, it’s definitely a new muscle I’m growing. At the same time, I can think about MLE in isolation and how it fits within the larger discipline of SWE. How I can build robust pipelines, integrate models into applications, deploy models within larger clusters, etc. I think there are many individuals which are pure MLE and lack the SWE perspective. Most critically, lots of ML people in my program aren’t computer people. They are math people or scientists first. They can grok the ML but grokking SWE without computer affinity is difficult. I see true full-stack being an understanding of low-level systems, back-front architecture, deployment, and now MLE. Just need to find someone who will compensate me for bringing all that to the table. Most postings are still either for SWE or PhD in MLE. Give me money!! I know it all


Yea, the problem is that most people expect things to be absolute and correct outside of engineering.

I love the gray areas and probabilities and creativity of software...but not everyone does.

So the real danger is in everyone assuming the AI model is, must be, and always will be correct. They misunderstand the tool they are using (or directing others to use).

Hmm. It's like autopilot on the Tesla. You aren't supposed to take your hands off the wheel. You're supposed to pay attention. But people use it incorrectly. If they get into an accident, then people want to blame the machine. It's not. It's the fault of person who didn't read the instructions.


I often wonder if society will readjust its expectation of programs or even devices. Historically, machines of all kinds were difficult to design and manufacture.. the structure was hard set (hence the name) but at the same time, society fantasize about adaptive machines, hyper adaptive, multipurpose, context-aware.. which if pushed high, is not far from the noisy flexibility of ML.


Yup. All the people I've worked with through my career post 2020 (AI/ML types) have been AI first and AI native. They're the first users of cursor - a year before it went mainstream.

Sorry not sorry that the rest of the world has to look over their shoulders.


Based on my experience as an MLE I would never use any of the current 'AI' offerings, so whatever bias you are suggesting is very recent.


i agree; but perhaps also it is the difference between managers and SWE? The former (SWE team leaders included) can see that engineers aren't perfect. The latter are often highly focused on determinism (this works/doesn't) and struggle with conflicting goals.

Through a career SWEs start rigid and overly focused on the immediate problem and become flexible/error-tolerant[1] as they become system (mechanical or meat) managers. this maps to an observation that managers like AI solutions - because they compare favourably to the new hire - and because they have the context to make this observation.

[1] https://grugbrain.dev/#:~:text=grug%20note%20humourous%20gra...


OpenIA's Codex-1 isn't so cool anymore. If it was ever cool.

And Claude Code used Opus 4 now!


Claude Code. And... Junie in Jetbrains IDE. It appeared recently and I'm really impressed by its quality. I think it is on the level of Claude Code.


I think it uses claude code by default, it is literally the same thing, with different (better) interface.


Really interesting. Source?


Junie in Ask mode:

> Which LLM are you?

> I am Claude, an AI assistant created by Anthropic. In this interface, I'm operating as "Junie," a helpful assistant designed to explore codebases and answer questions about projects. I'm built on Anthropic's large language model technology, specifically the Claude model family.

Jetbrains wider AI tools let you choose the model that gets used but as far as I can tell Junie doesn't. That said, it works great.


That just means it's using Sonnet, not that it's using Claude Code.


Even that doesn't have to be true, LLMs often impersonate other popular models.


usually OpenAI, but yes


> How you spend that time is one of the most important moral decisions of your life.

What if the work itself isn't "the most important" thing in the person's life?


Then you get to join the 99% of the rest of humanity that views work as something necessary to enable the things they enjoy and find purpose in when they're not working.

Finding purpose, fulfillment or joy in your work is nice, but as you grow as a human, or as the field you're in changes, or as the work dries up... well, you're left thoroughly adrift.


I work in education, a field famous for attracting people based on their own willingness to do good work for good reasons.

And at the end of the day, a job is a job. I do it because it allows me to live a lifestyle close to what I want, while not being soul crushingly boring most of the time.

I came to terms with the fact that I'm not going to change the world. The best I can do is not fuck it up anymore than when I got here. That's about as good as most of us can expect, since most of us are average in many aspects. Without stunning amounts of genius or resources, I think that hoping just to fade into obscurity is the best you can do, really.


I feel this. I do think though, he mentions that this is the exact trap that is laid down for you when you enter society. You’re led to believe resources/genius is what separates do’ers, but I think he wants us to believe in the blind faith doing = progress. The other is option is doing nothing at all and succumbing to obscurity like you’re saying.


Why is a usual STT (speech-to-text) called AI?. (fixed typo)


STT (driven by machine learning models) has been part of the academic research field called Artificial Intelligence for decades.


Probably because AI is a misnomer. Machine learning is a better name, but even than it is pretty lacking.

What is generally called AI in common speech is pretty much a class of non-linear statistical models which require some training to generate weights which are than used to fit the model. Most people that know anything about statistics knows this, so it is fine actually. Misnomers exists in all industries and all science, and we just deal with them.


I believe this article is about STT, not TTS


AI does a better job of STT than the other techs.


Well, if the users ask frequent/common questions to ChatGPT and get acceptable answers, is this even a problem? If the volume of duplicate questions decreases, there should be no bad influence on the training data, right?


They spoke to this point in the abstract. They observe a similar drop in less common and more advanced questions.


I have queried gemini about a specific issue and when it failed to give a decent answer I narrowed down the issue with hints about the answer, it responded with details from an old post by me on reddit.


When I was interviewing at my last company, during the raising bar interview, the interviewer said that he looked at my personal website, liked it, and was really impressed by it. Then we spent one hour chatting about our experiences. So, I can say that it definitely helped me get hired. https://andlukyane.com/blog/


Last year, I started playing DnD with 5e rules and lots of homebrew content for the first time. It is an online game with my friends on Discord. We have character sheets, roll a dice using a bot, and so on. This is really fun, we had some crazy adventures and this campaign is just at early stages :)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: