I taught an intro course last semester. It was intended for non-CS majors, but it ended up with one module having all CS majors after all. They were very pessimistic about their job opportunities at graduation.
I explained that the fundamentals are still very much necessary for now, even if you end up only reviewing AI code. Honestly, computational thinking is as important as ever, although how persuasive I was about this is up for debate.
We used some tools AI models just aren't good at (visual languages are not a strength of language models, and I explained that they couldn't help from day one), but it meant some weaker students still tried to use AI and were confidently told incorrect instructions. They often ended up stuck because the newest group we've gotten is very adverse to office hours when ChatGPT exists (out of ~75 students, only one ever showed up, although I did meet with many right after class).
I'm very concerned for these students, using AI as a crutch was definitely not helping them succeed, but the ability to get easy answers (even if totally wrong) is too appealing. In the classroom they seemed interested, but when they get to a chatbot, they don't want to put it in the "learning" mode, they want to be done with the assignment, and they aren't taught enough "AI literacy" to know to think critically about the outputs or their use of it in general.