Uncritical support for these AI bros refusing to learn CS and thereby making the CS nerds that actually know stuff more employable.
Unfortunately, I am experiencing the opposite effect. I am an IT contractor tasked with writing scripts, and I keep applying for dev and IT jobs with a 2 year degree with no success for the past few years while I see old dinosaur fucks at my job not even knowing what functions are. They use ChatGPT to write scripts for them without any modifications to work with our specific clusterfuck of an environment that they created, and the scripts are broken and run in production because we don’t even have a testing environment. Meanwhile I have to clean their mess and get paid much less than them, let alone not have any PTO or benefits. It’s absolutely maddening to be moderately skilled in programming and witnessing some of the dumbest people on the planet get CS jobs through nepotism or impressing HR, moreso the former. It seems like my workplace will only hire someone if and only if they are incredibly incompetent.
And now the team of fucking morons are taking care of packaging, so the one thing in my job that gave me control to fix things and helped with my sanity is being stripped away from me, which means I get to image laptops unsuccessfully because their scripts will not work. I might as well bring a personal laptop to work and practice programming since I am going to be sitting around waiting shit to get fixed for weeks at a time. Can’t wait to get the fuck out of the IT field and pursue electrical engineering someday. I’m not wasting 10-20 years of my life just to get a single promotion in a field dominated by cop-worshipping, white supremacist libertarians.
Wish I could use a real programming language in a sane environment where I am hired to actually be a developer and not an IT person with 3 odd jobs that don’t help me build any transferrable experience and get paid for less than one for once.
Honestly, how is it possible to unionize white-ass techbros and old techfucks when they don’t hold themselves accountable, work against each other, refuse to keep up to date in their field or learning anything new, and scream fire for non-issues so that other teams have to struggle, let alone most of techbros being against unions or anything remotely left of Hitler?
Absolutely, IT and tech stuff is full of failing up. And it is exacerbsted by generational knowledge gaps that put vaguely millennial-aged people in the “sweet spot” for competent debugging and CS knowledge. The older generations that did not need to update their skills retained positions where those skills had been needed, passing them to a younger generation so they could join a massively incompetent managerial class. And zoomers etc never had to fight much with technical issues so they went straight to frameworks and web stuff.
Obviously there are exceptions in all cases, but this tendency has lead to a situation where the Boomer/Gen X managerial class does the incompetent things you mention, with everything working only because one or two people under them put out the fires (usually but not always in the millennial range) and when younger still people get hired to replace them (I.e. pay people less) the whole thing explodes and management does its best to fail upwards again while everyone else gets to be unemployed for a while. This can be put off for a while by accidentally hiring competent zoomers that inevitably get sick of that shot and get a different job.
My comment is really just joking about a silly extreme of this, which is CS students failing to learn basic CS because they think that AI solves everything for them. AI basically becomes the nerd they pay to do their homework poorly and then they will get a job, gear someone sat, “this is all built on linked lists” and have no idea what anyone is talking about despite dropping $130k on a degree that covered that topic a bunch of times.
PS sorry you have to be in that environment, it sounds very frustrating. Nothing worse than being responsible to fix something that the higher-up decision-makers break. In my experience they even blame you for not fixing it fast enough or preventing the problem in the first place even though they forced the decision despite your objections. Scapegoating is rewarded in these corporate environments, it is how the nepotism fail-children protect their petty fiefdoms.
How do old techbros not know what a function is do they code entirely in Hex or something?
My workplace for some reason wants 20 years of IT manager experience for dev jobs, and many of the incompetent higher-ups were promoted from IT despite having no experience, especially outside experience from any other company, which is why they make the stupidest decisions. Management is also equally incompetent. The particular software team that handles patches is led by a shitty manager that creates a terrible culture that seems to make even decent people become incompetent, like some brain worm. I worked with another contractor in my team that was hired for IT Networking with literally no experience. I haven’t had the same luck, probably because I don’t fit in with their culture.
“INTEGER SIZE DEPENDS ON ARCHITECTURE!”
idk why boomers decided to call integers int/short/word/double word/long/long long
you ever heard of numbers? you think maybe a number might be a little more descriptive than playing “guess how wide i am?”
yet another thing boomers ruined
This is simply revolutionary. I think once OpenAI adopts this in their own codebase and all queries to ChatGPT cause millions of recursive queries to ChatGPT, we will finally reach the singularity.
There was a paper about improving llm arithmetic a while back (spoiler: its accuracy outside of the training set is… less than 100%) and I was giggling at the thought of AI getting worse for the unexpected reason that it uses an llm for matrix multiplication.
Yeah lol this is a weakness of LLMs that’s been very apparent since their inception. I have to wonder how different they’d be if they did have the capacity to stop using the LLM as the output for a second, switched to a deterministic algorithm to handle anything logical or arithmetical, then fed that back to the LLM.
I’m pretty sure some of the newer ChatGPT-like products (the consumer-facing interface, not the raw LLM) do in fact do this. They try to detect certain types of inputs (i.e. math problems or requesting the current weather) and convert it to an API request to some other service and return the result instead of a LLM output. Frankly it comes across to me as an attempt to make the “AI” seem smarter than it really is by covering up its weaknesses.
modern CS is taking a perfectly functional algorithm and making it a million times slower for no reason
mallocPlusAI
That made me think of… molochPlusAI("Load human sacrifice baby for tokens")
I’d just like to interject for a moment. What you’re refering to as molochPlusAI, is in fact, GNU/molochPlusAI, or as I’ve recently taken to calling it, GNUplusMolochPlusAI.
GNUplusMolochPlusAI!
GNUplusMolochPlusAI!
GNUplusMolochPlusAI!
Now is the time to dance!
GNUplusMolochPlusAI!
GNUplusMolochPlusAI!
GNUplusMolochPlusAI!
-–
To a drum and bass beat