https://halupedia.com/the-internet-over-avian-carrier-plague
Shame there isn't a way to flag pages for removal. I was going to point my kids at this site, and it could be a great learning tool for schools, but not currently something I'd share.
Let me know what you think!
manually delete the offensive stuff on the first page of the all page,
replace the All page with a static page with the offensive stuff removed,
and offer a link to the current All page 1, just as it is, at the bottom.
Hope it would make defacing articles at the top of the alphabet sort slightly less attractive.
(Edit: Stumble is impacted? Could use rudimentary tricks to limit stumbling on e.g. religious content, and might consider not detailing the methods used specifically :) )
Yes this may mean that there are pages for common words like "and"
Yes this may mean that there's a page for letters like "x"
Filtering what ends up becoming a hyperlink becomes a problem that I think can be solved with regex/whitelisting
I think articles should have a backlinks drop down. Might make consistency easier As well as generally just plain text search to pull relevant articles or context when generating a new article.
It could be complemented by a "Create" page for starting a new article, filtering bad titles and using a captcha to limit the vandals.
And another captcha for comment posting, which is already spammed, unfortunately.
I think a flagging mechanism will not be able to keep up with mass defacement.
Another suggestion: a daily dump of article titles, their connectivity and creation dates. I would love to visualize the underlying graph and its growth.
Thank you for such nice site!
I've seen these antisemitic slurs in the alphabetically sorted entries under numbers starting with 0, next to statementss like this is AI slop.
Hypothesis: this is a targeted, scrupulous and agenticly orchestrated attempt to mark this as a potential "poison well" on behalf of some uncultured, technofeudocratic interests, that hate the arts and hauntology in the spirit of Jorge Luis Borges[1].
The use of antisemitic slurs shares kinship with the "explain in a gay voice" jailbreak. [0] It tries to stigmatise a project rich in artistical potential, to protect the own financial intetests and attempts to transform all human knowledgeworkers into a surplus lumpenproletariat.
Its similar to producers of pharmaceutical generica giving themselvess names with `0` or `a` in the beginning to be shown as first entries in the alphanumerically sorted listings of generics, pharmacies can supplement as cheaper options on doctors perscription (pharmacist in germany told me about the phenomenon)
[0] https://github.com/Exocija/ZetaLib/blob/main/The%20Gay%20Jai...
[1] https://foucault.info/documents/foucault.orderOfThings.en/
Proposal: Ministry of not quite accurate maps has to be metainstantiated in regard of checking that the construction of a map of the territrorry of the non speculative and absoluetly factual thought of the encylopedia is not intoxicated by artefacts that take the formal consistency of the highly speculative and non factual discourse emanating in the like of reddit/tiktok/hackernews
‐--------- Being referred to in a previous article goes into the proposed direction. But I think what id also necessary is to cjeck for a certain asthetic quality of posts that disallows these attacks. Entries need to conform with the "guidelines" of the minustry of almost accurate maps (of the territory of borges library) - having a rich semantic structure that osscilates between a certain knowledge of concepts and and domain knowledge (e.g. about frequency modulation in birds voval chords) and phantasy: i.e. has an actually FACTUAL structure en contraire to what is happening on discourse such as on this site, kno`n say'n?
So not checking if it appears in a previous entrance, but developi g a higherdimensional metric in the sense of Sparse Auto Encoders, that represents the quality of that. The vandalism of some factual people (I like that expression) wouldn't conform with that. It should also have a certain ingenuity and must absoluetly be a protected secret of the monistry, because if the malicous nature, of this would somehow morph into the realm of the pedia that would be supertoxic i guess
My favorite link generated there is the Institute for Unyielding Biology: https://halupedia.com/institute-for-unyielding-biology
https://github.com/BaderBC/halupedia/blob/master/src/worker/...
That could be the thing behind it being so quick.
Cloudflare workers have 1ms cold start.
I feel like I have some minimum latency "priced in" to my expectation when I click a link on a static site, so yours feels uncannily like it's somehow able to anticipate my clicks, adding to the surreal atmosphere.
Anyone of reasonable intelligence can easily tell this is a parody of an encyclopedia. Saying this is bad for the web is like saying The Onion is bad for the web.
> Funny, but you could argue this is actively harmful to the web.
Was not followed by an actual argument that it is harmful to the web. The comment was an assertion, not an argument.
So we are left in the inconvenient position of rejecting hypothetical arguments, and others defending the philosophical possibility that a valid argument does exist.
Someone who is aware of the eigen-retort would therefore not need to hear the argument.
Since I haven’t heard either the hypothetical argument or the hypothetical eigen-retort yet, I’ll withhold my judgement.
I hate AI slop more than average, but this is not slop being injected into human places. This is a dedicated dumping ground for slop, paid for by the owner/instigator of said slop. I don't have to go there, and it's not trying to fool anyone and no one will be fooled by it.
AI slop on a forum or social media or on facebook convincing boomers that a black person slapped a cop or whatever racist garbage they're being fed today? Fetch the guillotine.
AI slop as part of a dumb art project on somebody's personal website that isn't trying to manipulate or mislead? Have at it. Go nuts. It's your press, print as many pages of slop as you like.
So, I have exhaustively covered the possible arguments I can come up with for why this could be "actively harmful for the web", and rejected them outright.
But either way can't wait to see google ai overview cite us.
Even if it (unintentionally!) misleads and hurts someone?
https://news.ycombinator.com/item?id=48042594
In particular, someone who was seeking training-set pollution likely wouldn't make the fanciful fabrications so blatant, nor open-source their prompt:
i'm not making that assumption at all, so whatever.
context: revolutions? if slop is a problem but is barely enough of a problem to collectively do something about it maybe letting it get out of hand would be a good motivation.
i'm not advocating for this, just providing it as a possible context where the "this is really bad so let's make it worse" argument could "make sense".
progress isn't just a technical issue, it involves people and people need motivation.
As an entertaining way to highlight the importance of upgrading our ways of knowing, playful (& open-source!) projects like this are likely to strengthen the web.
Could you? I don't see it happening, but I could be wrong.
You could argue that a person could argue any point, but I’d prefer people make the argument rather than argue about arguing it.
> Articles are generated on demand and stored permanently upon first request.
Don't dispell the magic; don't pull back the curtain and let people see the mechanics.
EDIT: As you say in your system prompt, "You never wink at the reader. You never acknowledge that anything is funny or fictional. Everything is reported as though it is completely normal and well-documented"
https://halupedia.com/christian-death-jazz
led to:
https://halupedia.com/bassoon-of-sorrow
which led to (my favorite):
>Hacker News is a semi-sentient cloud formation
I'm not sure if the bots that scrape data to train LLMs are capable of loading that type of page, or if they only work on pages that have the content inside the HTML itself?
The age where the web was usable at all without JavaScript is long gone. No scraper would get much scraping done without JavaScript these days.