With the DeepSeek breakthrough reviving the hype around large language models I've just had a thought: what if we trained an LLM using the contents of the Geminispace to measure how much knowledge is added to as time progresses?
I think it would be an interesting experiment considering the Geminispace is in its early stages. It may give some interesting insights on how knowledge accumulates overtime.
=> Posted in: s/AI | 🗿 argenkiwi
Jan 30 · 2 days ago · 🙁 2
=> ☕️ protoc0l · Jan 30 at 01:01:
no.
I wish that Gemini space remained for humans to discover. A general- purpose LLM may be fun though
=> 🌻 softwarepagan · Jan 30 at 02:29:
No. If "AI" ever touches this, we need to burn it all to the ground. Leave that unholy abomination as far away from Geminispace as possible.
=> 👻 nightliife · Jan 30 at 03:18:
I'm here to talk to humans, not bots. I could just do that on X or Facebook if I wanted to (I don't).
=> 👤 jdcard · Jan 30 at 04:49:
Ummm... aren't we talking about using Gemini content to train the model, rather having the model interact with users directly in Gemini-space? Once a month or once a year the model would be queried to summarize (or produce detailed reports of) the changes that occurred in Gemini content available during that period. A research tool, perhaps not even accessible on the Gemini network, rather than chat-bot play toy.
=> 👤 jdcard · Jan 30 at 04:52:
I suppose there is nothing stopping anyone from doing that now, if they were so inclined and had the compute and bandwidth resources to dedicate to it.
=> 🦋 CarloMonte · Jan 30 at 06:00:
i think, the amount of contents in gemini space is several orders magnitude below what is required to train an LLM.
=> 🐐 drh3xx · Jan 30 at 09:25:
@softwarepagan haven't you heard we NEED AI in everything? Just waiting for the AI trainers with integrated cameras on the toe peg that can text me when my laces have come undone and record a log on a nice web dashboard.
=> 🎮 lucss21a · Jan 30 at 13:08:
no. just no. i don't want my little corner of the internet to be crawled by the tenacles of big ai. it's horrid to imagine that. please reconsider your life choices and go back to the world wide web again. please. touching grass is also a better choice. dumbass.
=> 🎮 lucss21a · Jan 30 at 13:09:
sorry for being rude but please don't. it's a net negative for us here.
=> 🐸 HanzBrix · Jan 30 at 13:22:
For the ones who don't want to be crawled, scraped and the like, aggressive rate limiting would solve the problem. 😁
=> 🎮 lucss21a · Jan 30 at 13:25:
i heard some old web oriented platforms such as nekoweb and poyoweb includes scraper blockers
@vi, totally with you!
We need to start thinking of how to build antitank weapons against these things. Pollution is not a bad idea.
=> 🐸 HanzBrix · Jan 30 at 15:04:
@lucss21a You can set up ufw or iptables to outright ban connection spam, basically killing all scrapers.
why whould anyone want this. the purpose of gemini is to stay minimal and clean, not to spread slop generated by bullshitting machine trained on stolen data. there's no good gen-AI, the whole concept is rotten to the core. ew. yikes, even.
=> 🗿 argenkiwi [OP] · Jan 30 at 18:36:
Thanks @jdcard. That is exactly the thought experiment I had in mind. I wasn't thinking about filling the Geminispace with bots, but I guess it is a reasonable fear. It seems it will not be easy to prevent that. How do we know one of us is not a bot right now?
But as @CarloMonte says there is only enough content for an SLM at this stage. Although I'm sure eventually they will develop techniques to learn from a small knowledge base and make an expert SLM for a specific use case.
Now I am thinking of making a parody of the LLM I was imagining when I started this post. It would be a self-hating one. What should I call it? RATMGPT, DeepRATM...
text/gemini; charset=utf-8
This content has been proxied by September (3851b).