• 2 Posts
  • 30 Comments
Joined 14 天前
cake
Cake day: 2025年12月30日

help-circle
  • This is a subject that people (understandably) have strong opinions on. Debates get heated sometimes and yes, some individuals go on the attack. I never post anything with the expectation that no one is going to have bad feelings about it and everyone is just going to hold hands and sing a song.

    There are hard conversations that need to be had regardless. All sides of an argument need to be open enough to have it and not just retreat to their own cushy little safe zones. This is the Fediverse, FFS.


  • I have never once said that AI is bad. Literally everything I’ve argued pertains to the ethics and application of AI. It’s reductive to call all arguments critical of how AI is being implemented “AI bad”.

    It’s not even about it being disruptive, though I do think discussions about that are absolutely warranted. Experts have pointed to potentially catastrophic “disruptions” if AI isn’t dealt with responsibly, and we are currently anything but responsible in our handling of it. It’s unregulated, running rampant and free everywhere claiming to be all things for all people, leaving a mass of problems in its wake.

    If a specific individual or company is committed to behaving ethically, I’m not condemning them. A major point to understand is that those small, ethical actors are the extreme minority. The major players, like those you mentioned, are titans. The problems they create are real.






  • From what I’ve heard, the influx of AI data is one of the reasons actual human data is becoming increasingly sought after. AI training AI has the potential to become a sort of digital inbreeding that suffers in areas like originality and other ineffable human qualities that AI still hasn’t quite mastered.

    I’ve also heard that this particular approach to poisoning AI is newer and thought to be quite effective, though I can’t personally speak to its efficacy.



  • Is the only imaginable system for AI to exist one in which every website operator, or musician, artist, writer, etc has no say in how their data is used? Is it possible to have a more consensual arrangement?

    As far as the question about ethics, there is a lot of ground to cover on that. A lot of it is being discussed. I’ll basically reiterate what I said that pertains to data rights. I believe they are pretty fundamental to human rights, for a lot of reasons. AI is killing open source, and claiming the whole of human experience for its own training purposes. I find that unethical.








  • I do agree with your point that we need to educate people on how to use AI in responsible ways. You also mention the cautious approach taken by your kids school, which sounds commendable.

    As far as the idea of preparing kids for an AI future in which employers might fire AI illiterate staff, this sounds to me more like a problem of preparing people to enter the workforce, which is generally what college and vocational courses are meant to handle. I doubt many of us would have any issue if they had approached AI education this way. This is very different than the current move to include it broadly in virtually all classrooms without consistent guidelines.

    (I believe I read the same post about the CEO, BTW. It sounds like the CEO’s claim may likely have been AI-washing, misrepresenting the actual reason for firing them.)

    [Edit to emphasize that I believe any AI education we do to prepare for employment purposes should be approached as vocational education which is optional, confined to those specific relevant courses, rather than broadly applied]


  • While there are some linked sources, the author fails to specify what kind of AI is being discussed or how it is being used in the classroom.

    One of the important points is that there are no consistent standards or approaches toward AI in the classroom. There are almost as many variations as there are classrooms. It isn’t reasonable to expect a comprehensive list of all of them, and it’s neither the point nor the scope of the discussion.

    I welcome specific and informed counterarguments to anything presented in this discussion, I believe many of us would. I frankly find it ironic how lacking in “nuance or level-headed discussion” your own comment seems.


  • I appreciated this comment, I think you made some excellent points. There is absolutely a broader, complex and longstanding problem. I feel like that makes the point that we need to consider seriously what we introduce into that vulnerable situation even more crucial. A bad fix is often worse than no fix at all.

    AI is a crutch for a broken system. Kicking the crutch out doesn’t fix the system.

    A crutch is a very simple and straightforward piece of tech. It can even just be a stick. What I’m concerned about is that AI is no stick, it’s the most complex technology we’ve yet developed. I’m reminded of that saying “the devil is in the details”. There are a great many details in AI.



  • I get where he’s coming from… I do… but it also sounds a lot like letting the dark side of the force win. The world is just better with more talent in open source. If only there was some recourse against letting LLM barons strip mine open source for all it’s worth and only leave behind ruin.

    Some open source contributors are basically saints. Not everyone can be, but it still makes things look more bleak when the those fighting for the decent and good of the digital world abandon it and pick up the red sabre.