Post

Avatar
Avatar
JSTOR when Aaron Swartz values their 'corpus': boo! evil! one million years dungeon! JSTOR when AI brotech values their 'corpus': hell yeah! this rules! Yeet yourself into this corpus, fam!
Avatar
I would like JSTOR to clarify what it means with the following words, because they keep using them, and I don't think they know what they mean: collaboration, community, interactive, trusted, corpus, empower, deepen, and expand.
Avatar
Ai, generative, research
Avatar
I'm actually really curious what they think people are.
Avatar
What. The. Actual. Fuck?
Avatar
Avatar
I feel like they don't know what "generative" means. Or they don't care. It doesn't mean "go fetch this stuff for me".
Avatar
Well, an LLM is generative AI, and it looks like they're using one. This is a thing you will see being used in many places as we figure out how to do them better. I can explain further how they work and avoid hallucinations if you'd like.
Avatar
Avatar
Nobody wants you to explain this. People want you and the rest of these AI cultists to stop shoving it down our throats. Nobody wants this shit. Any of it.
Avatar
Does this one not use a fuckton of water?
Avatar
The backend is gpt-3.5-turbo so... no. It only costs 1.7% of what GPT-4 does to run. You don't need to run the massive models for tasks like this which is a major upshot.
Avatar
It should be in not existing-a.
Avatar
Avatar
can’t wait for it to make up shit and students to fail their papers for not reading their citations
Avatar
That's not a huge problem when dealing with a limited corpus. The LLM can piece together words which you can then tokenize and vectorize and then compare to the same info found in the limited dataset it is supposed to talk about. If the LLM goes off into la-la land have it try again.
Avatar
[nodding sagely] i understand what these words mean
Avatar
So you’re saying the trick is to use the lie machine but not trust it any farther than you can throw it?
Avatar
YES! That is a nice way of putting it. Checking the output isn't easy but we're coming up with ways of doing it. Kind of exciting!
Avatar
That’s where I get lost; the part where it lies and you have to say “no you’re lying stop that and try again” which requires that you know when it is lying If I already know when it’s lying, why did I need to ask in the first place?
Avatar
so it can lie and make stuff up? which means students can just believe it and not check their sources? soooo…
Avatar
Ah, that's part of the computer's job, not the user. FIguring out how close you are to "ground truth" is hard in a generic context but easier when dealing with a limited corpus.
Avatar
I don't even need to publish anymore. Just get JSTOR's AI to say I did.
Avatar
they let robots have jstor but not aaron swartz
Avatar
It’s legal, even encouraged, to let the lie robot have all the documents, but if someone wants to let us have all the documents without putting them in the idiot blender it is bad
Avatar
Avatar
Interesting, my work appears in JSTOR, but I don't recall being collaborated with on this.
Avatar
Avatar