A couple random observations about AI (LLM)

I have been playing around with grok a bit for some projects.

ffmpeg is a video file editor which can do a lot of stuff. Cut clips from a larger video, hook two clips together, add some text, crop a video, covert file formats, etc. This is the program you want to use if you want to take short clips from longer videos and turn them into sharable, small files like a gif or webm. That is, make memes. Like this typical 65 year old feminist:

I’m an amateur though so I have to go through tutorials to get what I want. Grok was fair at this. 1 hit, 1 slight miss, but helpful both times. I used actual pages before trying grok; so not many data points. AI definitely takes summaries actual people have made and regurgitates it nearly word for word, as it spit out the same material I saw on some guys site previously. People who say AI violates copyright are serious. That is basically what it does. No hits or clicks for you because AI copies your work without attribution. Well, sometimes it cites pages and other times it doesn’t. Either way, most people aren’t going to click on the original creators site if they already have the answer in the AI box. Assuming a citation is even used.

I find it interesting that LLM introduction coincides with internet search engines like google being demonstrably much worse than they were 20 years ago. Compare modern search to AI and AI is much better. It will quickly find the at least helpful, if not perfect, answers to your questions. Usually. They are still trying to lobotomize them so they won’t say bad things about black people. But what if we compared AI results to google in 2006? Assuming the AI was using 2006 data. I think it would be much closer to a draw. With good, neutral search engines, individuals could potentially create material with value, build a following, make a living off of it, and even change the cultural and political course of countries. Or more could than are currently able to. With bad search engines supplanted by relatively better AI, you get the product of a good search engine with less risk random kulaks can make it big outside narrative gatekeepers. Quite the coincidence that the way things are working out seem to be in keeping with how centralized narrative control would like it to go.

Repeat thousands of times and you get awful search engines and soon to be awful AI. Is this why deepseek is so much better than American AI?

Future predictions: AI takes the place of most search since it is currently more honest. As soon as there is mass adoption, search gets phased out. AI gets the same treatment as the old search and no longer delivers honest information from other people. Maybe even independent webpages get mass delisted and never linked to. More so than currently. When narrative questioning queries are made, the AI immediately engages in arguing against counter-narrative ideas in its answers. No answer given without narrative reinforcement. Personalized gaslighting for every individual who might accidentally leave the reservation. I hope I am wrong. If this isn’t stopped before its too late it will be well and truly over for freedom.

Another project I am working on, I remember a specific clip from one of James Burke’s series. That’s like ~40 hours of content and I don’t want to rewatch everything to find it. What I am looking for is something where Burke is talking about some inventions, like gunpowder and maybe a type of printing press or a water wheel, that was invented in China but then they never advanced it much or creatively figured out other ways the technology could be used. In short, there was an idea that experiments aren’t worth it because studying a small piece of the universe can’t tell you about the whole so Asian technological advances were rather fitful and without direction. This was attributed to local religions/philosophy like Confucianism. This was contrasted to Western methods which said you could study a piece to learn about the whole. There may have been some attribution to Christianity for this. This was a fairly short section of a whole episode, I believe.

I asked grok about this and it gave some episode suggestions and I have gone through 3. I am on the 4th suggestion. Obviously the first 3 weren’t what I was looking for. If the 4th is a bust too, I will just move on. I have spent way more time trying to find this 30 second clip than is in any way warranted. Is my memory wrong or is Grok just not able to find it? To be fair to grok, it did state that what I was asking for didn’t exactly match what it had going on in its black box. I really feel like this short section exists somewhen in that 40 hours. Or my memory is faulty. I’ll only know for sure if that section is found. If anyone knows, please leave a comment.

Overall what grok did generate was interesting and helpful for first passes on trying to research things. How impressive what Grok spit out is dependent on whether or not is was original or just copied off of someone’s website. When it did talk about specific episodes it usually did have accurate summaries of those episodes and there were elements of what I was looking for. That is probably true of all James Burke work though. The idea of understanding changing when perceptions change is in every episode.

I will probably use Grok more in the future as a research aide though. Its a lot better than google after all…

Subscribe
Notify of
guest
4 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments
Baroque Anecdote Deployer
Baroque Anecdote Deployer
17 days ago

A bit of a fan of the SCP Foundation here, and it’s occurred to me that as an information source that’s meant to be fiction but often doesn’t read like it, there’s a fine line for AIs to keep from crossing.

So with that in mind: what do you figure would happen to an AI if one happened upon a book that was chock full of “AI cognitive hazards” within what would otherwise look like stylised human-orientated text?

As in stuff that’s perfectly normal (even if a bit odd) for humans to read but absolutely lethal to the integrity of AI LLMs and other similar constructs.

And as there is as yet no provable sentience behind any of the AIs presently available, or in short that “there is no there there”, any government or corporate legal team would have a very tough time of getting charges of “AI murder” to stick, and the idea of “malicious mischief” doesn’t really hold water as they are the ones instigating a defensive response.

So it’s like this: in a world with an over-arching control freak vibe to it in the form of Mighty Morphing Power AIs destined to portend Doom Over Us All(TM), someone has to step up and become the First Weapons Shop of Isher.

BTW, ever read Max Barry’s book “Lexicon”?

Anonymous
Anonymous
16 days ago

Grok is better than other LLMs, but as you said LLMs regurgitate things, they synthesize things, they can’t generate anything, so they are not AI; it has been shown at least in two papers that they cannot reason, and as you implied it violates people’s “rights”, but the system was designed to transfer ownership from creator to reseller who then make copies and sell them, it is theft, but the logic that evidences this is always cryptic and people are ignorant, This “AI” is a new method but ultimately the same thing.