this post was submitted on 18 Aug 2025
1138 points (99.0% liked)

Technology

74439 readers
2208 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] prole@lemmy.blahaj.zone 84 points 6 days ago (5 children)

Tech bros just actively making the internet worse for everyone.

[–] ShaggySnacks@lemmy.myserv.one 65 points 6 days ago

Tech bros just actively making ~~the internet~~ society worse for everyone.

FTFY.

load more comments (4 replies)
[–] nialv7@lemmy.world 34 points 6 days ago* (last edited 5 days ago) (1 children)

We had a trust based system for so long. No one is forced to honor robots.txt, but most big players did. Almost restores my faith in humanity a little bit. And then AI companies came and destroyed everything. This is why we can't have nice things.

[–] Shapillon@lemmy.world 19 points 6 days ago

Big players are the ones behind most AIs though.

[–] thatonecoder@lemmy.ca 42 points 6 days ago (1 children)

I know this is the most ridiculous idea, but we need to pack our bags and make a new internet protocol, to separate us from the rest, at least for a while. Either way, most “modern” internet things (looking at you, JavaScript) are not modern at all, and starting over might help more than any of us could imagine.

[–] Pro@programming.dev 44 points 6 days ago* (last edited 6 days ago) (10 children)

Like Gemini?

From official Website:

Gemini is a new internet technology supporting an electronic library of interconnected text documents. That's not a new idea, but it's not old fashioned either. It's timeless, and deserves tools which treat it as a first class concept, not a vestigial corner case. Gemini isn't about innovation or disruption, it's about providing some respite for those who feel the internet has been disrupted enough already. We're not out to change the world or destroy other technologies. We are out to build a lightweight online space where documents are just documents, in the interests of every reader's privacy, attention and bandwidth.

[–] thatonecoder@lemmy.ca 12 points 6 days ago (2 children)

Yep! That was exactly the protocol on my mind. One thing, though, is that the Fediverse would need to be ported to Gemini, or at least for a new protocol to be created for Gemini.

[–] echodot@feddit.uk 10 points 6 days ago (1 children)

If it becomes popular enough that it's used by a lot of people then the bots will move over there too.

They are after data, so they will go where it is.

One of the reasons that all of the bots are suddenly interested in this site is that everyone's moving away from GitHub, suddenly there's lots of appealing tasty data for them to gobble up.

This is how you get bots, Lana

load more comments (1 replies)
load more comments (1 replies)
[–] cwista@lemmy.world 9 points 6 days ago

Won't the bots just adapt and move there too?

[–] 0x0@lemmy.zip 4 points 6 days ago

It's not the most well thought-out, from a technical perspective, but it's pretty damn cool. Gemini pods are a freakin' rabbi hole.

load more comments (7 replies)

reminder to donate to codeberg and forgejo :)

[–] bizza@lemmy.zip 14 points 5 days ago

I use Anubis on my personal website, not because I think anything I’ve written is important enough that companies would want to scrape it, but as a “fuck you” to those companies regardless

That the bots are learning to get around it is disheartening, Anubis was a pain to setup and get running

[–] zbyte64@awful.systems 30 points 6 days ago (6 children)

Is there nightshade but for text and code? Maybe my source headers should include a bunch of special characters that then give a prompt injection. And sprinkle some nonsensical code comments before the real code comment.

load more comments (6 replies)
[–] mfed1122@discuss.tchncs.de 16 points 6 days ago* (last edited 6 days ago) (5 children)

Okay what about...what about uhhh... Static site builders that render the whole page out as an image map, making it visible for humans but useless for crawlers 🤔🤔🤔

[–] lapping6596@lemmy.world 25 points 6 days ago (1 children)

Accessibility gets throw out the window?

[–] mfed1122@discuss.tchncs.de 15 points 6 days ago (1 children)

I wasn't being totally serious, but also, I do think that while accessibility concerns come from a good place, there is some practical limitation that must be accepted when building fringe and counter-cultural things. Like, my hidden rebel base can't have a wheelchair accessible ramp at the entrance, because then my base isn't hidden anymore. It sucks that some solutions can't work for everyone, but if we just throw them out because it won't work for 5% of people, we end up with nothing. I'd rather have a solution that works for 95% of people than no solution at all. I'm not saying that people who use screen readers are second-class citizens. If crawlers were vision-based then I might suggest matching text to background colors so that only screen readers work to understand the site. Because something that works for 5% of people is also better than no solution at all. We need to tolerate having imperfect first attempts and understand that more sophisticated infrastructure comes later.

But yes my image map idea is pretty much a joke nonetheless

load more comments (1 replies)
[–] echodot@feddit.uk 7 points 6 days ago (1 children)

AI is pretty good at OCR now. I think that would just make it worse for humans while making very little difference to the AI.

[–] mfed1122@discuss.tchncs.de 5 points 6 days ago (3 children)

The crawlers are likely not AI though, but yes OCR could be done effectively without AI anyways. This idea ultimately boils down to the same hope Anubis had of making the processing costs large enough to not be worth it.

[–] nymnympseudonym@lemmy.world 6 points 6 days ago (2 children)

OCR could be done effectively without AI

OCR has been neural nets even before convolutional networks emerged in the 2010s

load more comments (2 replies)
load more comments (2 replies)
[–] iopq@lemmy.world 4 points 5 days ago

AI these days reads text from images better than humans can

load more comments (2 replies)
[–] StopSpazzing@lemmy.world 18 points 6 days ago* (last edited 5 days ago) (2 children)

Is there a migration tool? If not would be awesome to migrate everything including issues and stuff. Bet even more people would move.

[–] BlameTheAntifa@lemmy.world 17 points 6 days ago

Codeberg has very good migration tools built in. You need to do one repo at a time, but it can move issues, releases, and everything.

load more comments (1 replies)
[–] Monument@lemmy.sdf.org 10 points 6 days ago

Increasingly, I’m reminded of this: Paul Bunyan vs. the spam bot (or how Paul Bunyan triggered the singularity to win a bet). It’s a medium-length read from the old internet, but fun.

load more comments
view more: next ›