Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "data literacy"
-
I am so mentally drained from having to deal with the intern who I have to literally spoon-feed every single thing. My previous posts illustrate the situation...
The language and cultural barriers are too much, and I am too afraid to open my mouth because of the sensitive nature of my country's history and I'll get labelled as some horrible person.
I told my manager today that I'll stick it out until end of January (thankfully I am on vacation for most of December and January), but I cannot work with her. She was supposed to move to the data team end of December, but my manager told me if she can't even properly grasp this HTML and CSS stuff, then she will not be able to do the other tasks they have for her.
This was a disaster of an experiment and I'm somewhat traumatised ( I am sure the intern is too) and I never want another intern again, nor do I want to manage people. I never said I want to be a people manager, I just want to quietly code at my desk.
This company sells MBTI psychometric assessments and they damn well know my preference, so I'm seriously annoyed that they threw this horrendous surprise on me and kept ignoring my requests for revisiting this intern's role, because I noticed a long time ago that she was struggling with basic concepts and all they did was make her do Udemy courses.
I told them multiple times that she seriously needs computer literacy training because she will not survive in this industry if she still struggles to understand how files and folders work. Other employers would have fired her a long time ago.
She's just too slow for this job. I feel sorry for her, but I do not have the capacity to do this anymore. I'm tired, it's been a long year.6 -
Data Disinformation: the Next Big Problem
Automatic code generation LLMs like ChatGPT are capable of producing SQL snippets. Regardless of quality, those are capable of retrieving data (from prepared datasets) based on user prompts.
That data may, however, be garbage. This will lead to garbage decisions by lowly literate stakeholders.
Like with network neutrality and pii/psi ownership, we must act now to avoid yet another calamity.
Imagine a scenario where a middle-manager level illiterate barks some prompts to the corporate AI and it writes and runs an SQL query in company databases.
The AI outputs some interactive charts that show that the average worker spends 92.4 minutes on lunch daily.
The middle manager gets furious and enacts an Orwellian policy of facial recognition punch clock in the office.
Two months and millions of dollars in contractors later, and the middle manager checks the same prompt again... and the average lunch time is now 107.2 minutes!
Finally the middle manager gets a literate person to check the data... and the piece of shit SQL behind the number is sourcing from the "off-site scheduled meetings" database.
Why? because the dataset that does have the data for lunch breaks is labeled "labour board compliance 3", and the LLM thought that the metadata for the wrong dataset better matched the user's prompt.
This, given the very real world scenario of mislabeled data and LLMs' inability to understand what they are saying or accessing, and the average manager's complete data illiteracy, we might have to wrangle some actions to prepare for this type of tomfoolery.
I don't think that access restriction will save our souls here, decision-flumberers usually have the authority to overrule RACI/ACL restrictions anyway.
Making "data analysis" an AI-GMO-Free zone is laughable, that is simply not how the tech market works. Auto tools are coming to make our jobs harder and less productive, tech people!
I thought about detecting new automation-enhanced data access and visualization, and enacting awareness policies. But it would be of poor help, after a shithead middle manager gets hooked on a surreal indicator value it is nigh impossible to yank them out of it.
Gotta get this snowball rolling, we must have some idea of future AI housetraining best practices if we are to avoid a complete social-media style meltdown of data-driven processes.
Someone cares to pitch in?14