this post was submitted on 26 Oct 2024
668 points (99.6% liked)

Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ

54462 readers
320 users here now

⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.

Rules • Full Version

1. Posts must be related to the discussion of digital piracy

2. Don't request invites, trade, sell, or self-promote

3. Don't request or link to specific pirated titles, including DMs

4. Don't submit low-quality posts, be entitled, or harass others



Loot, Pillage, & Plunder

📜 c/Piracy Wiki (Community Edition):


💰 Please help cover server costs.

Ko-Fi Liberapay
Ko-fi Liberapay

founded 1 year ago
MODERATORS
 
top 50 comments
sorted by: hot top controversial new old
[–] Iunnrais@lemm.ee 101 points 1 week ago (2 children)

Just let anyone scrape it all for any reason. It’s science. Let it be free.

[–] chicken@lemmy.dbzer0.com 12 points 1 week ago (4 children)

The OP tweet seems to be leaning pretty hard on the "AI bad" sentiment. If LLMs make academic knowledge more accessible to people that's a good thing for the same reason what Aaron Swartz was doing was a good thing.

[–] Ashelyn@lemmy.blahaj.zone 27 points 1 week ago* (last edited 1 week ago) (5 children)

On the whole, maybe LLMs do make these subjects more accessible in a way that's a net-positive, but there are a lot of monied interests that make positive, transparent design choices unlikely. The companies that create and tweak these generalized models want to make a return in the long run. Consequently, they have deliberately made their products speak in authoritative, neutral tones to make them seem more correct, unbiased and trustworthy to people.

The problem is that LLMs 'hallucinate' details as an unavoidable consequence of their design. People can tell untruths as well, but if a person lies or misspeaks about a scientific study, they can be called out on it. An LLM cannot be held accountable in the same way, as it's essentially a complex statistical prediction algorithm. Non-savvy users can easily be fed misinfo straight from the tap, and bad actors can easily generate correct-sounding misinformation to deliberately try and sway others.

ChatGPT completely fabricating authors, titles, and even (fake) links to studies is a known problem. Far too often, unsuspecting users take its output at face value and believe it to be correct because it sounds correct. This is bad, and part of the issue is marketing these models as though they're intelligent. They're very good at generating plausible responses, but this should never be construed as them being good at generating correct ones.

load more comments (5 replies)
[–] umbrella@lemmy.ml 8 points 1 week ago

i agree, my problem is that it wont

[–] funkless_eck@sh.itjust.works 5 points 1 week ago* (last edited 1 week ago) (2 children)

That would be good if they did that but that is not the intent of the org, the purpose of the tool, the expected or even available outcome.

It's important to remember this data is not being scraped to make it available or presentable but to make a machine that echos human authography convincingly more convincingly.

On an extremely simplified level, it doesn't want to answer 1+1=? with "2", it wants to appear like a human confidently answering an arithmetic question, even if the exchange is "1+1=?" "yes, 2+3 does equal 9"

Obviously it can handle simple sums, this is an illustrative example

load more comments (2 replies)
[–] Auli@lemmy.ca 5 points 1 week ago

Except it won’t. And AI we’ll be pay to play

[–] sunzu2@thebrainbin.org 2 points 1 week ago (1 children)

Aint jstor a private enterprise?

[–] Natanox@discuss.tchncs.de 21 points 1 week ago (2 children)

It's a US "non-profit". One that demands 19$ per article which they merely provide as aggregator, they don't own shit.

Utterly absurd.

[–] sunzu2@thebrainbin.org 9 points 1 week ago

Non profit here merely means they are exemot from US income taxes so they are grifting even hardrr on us.

MIT is grifting in a similar but bigger manner.

[–] Flocklesscrow@lemm.ee 3 points 1 week ago

Which means they're adding profit margin to the otherwise zero marginal cost of said information good.

[–] CosmicTurtle0@lemmy.dbzer0.com 44 points 1 week ago

To paraphrase Nixon:

"When you're a company, it's not illegal."

To paraphrase Trump:

"When you're a company, they just let you do it."

[–] Facebones@reddthat.com 35 points 1 week ago (3 children)

All is legal in the eyes of capital.

[–] DarkDarkHouse@lemmy.sdf.org 11 points 1 week ago

The real golden rule

load more comments (2 replies)
[–] rasakaf679@lemmy.ml 33 points 1 week ago
[–] sunzu2@thebrainbin.org 27 points 1 week ago (5 children)

Yes.. but it was MIT that pushed the feds to prosecute.

Never forge to name the proper perp.

Disgusting. And we subsidize their existence 🤡

[–] TheReturnOfPEB@reddthat.com 7 points 1 week ago* (last edited 1 week ago)

https://en.wikipedia.org/wiki/Carmen_Ortiz

Ortiz said "Stealing is stealing whether you use a computer command or a crowbar, and whether you take documents, data or dollars. It is equally harmful to the victim whether you sell what you have stolen or give it away."

So that was some bullshit, huh ?

[–] Flocklesscrow@lemm.ee 5 points 1 week ago

MIT releases financials and endowment figures for 2024:

The Institute’s pooled investments returned 8.9 percent last year; endowment stands at $24.6 billion

load more comments (3 replies)
[–] PanArab@lemm.ee 24 points 1 week ago* (last edited 1 week ago) (1 children)

Who writes the laws? There's your answer.

I'm curious why https://www.falconfinance.ae/ cares about this though.

The hell they are selling? https://www.falconfinance.ae/falcon-securities/

[–] TheOakTree@lemm.ee 25 points 1 week ago

I did some digging. It's a parody finance website that makes it seem like you can invest in falcons and make a blockchain (flockchain) with them. Dig a little further, go to the linked forum, and you'll see it's just a community of people shitposting (mostly).

[–] electricprism@lemmy.ml 24 points 1 week ago

Remember what you learned in school: Working as a team to solve a test or problem is unacceptable!!! Unless you are a company town.

[–] ayyy@sh.itjust.works 15 points 1 week ago
[–] What_Religion_R_They@hexbear.net 12 points 1 week ago (6 children)

double standards are capitalism's lifeblood

load more comments (6 replies)
[–] crmsnbleyd@sopuli.xyz 11 points 1 week ago

Anything the rich and powerful do retroactively becomes okay

[–] EmbarrassedDrum@lemmy.dbzer0.com 10 points 1 week ago (1 children)

and in due time, we'll hack OpenAI and get the sources from the chat module..

I've seen a few glitches before that made ChatGPT just drop entire articles in varying languages.

[–] FaceDeer@fedia.io 6 points 1 week ago (1 children)

AI models don't actually contain the text they were trained on, except in very rare circumstances when they've been overfit on a particular text (this is considered an error in training and much work has been put into coming up with ways to prevent it. It usually happens when a great many identical copies of the same data appears in the training set). An AI model is far too small for it, there's no way that data can be compressed that much.

thanks! it actually makes much sense.

welp guess I was wrong. so back to .edu scraping!

[–] xiao@sh.itjust.works 9 points 1 week ago

I'm still blaming the MIT for that !

[–] umami_wasbi@lemmy.ml 7 points 1 week ago (6 children)
[–] Albbi@lemmy.ca 10 points 1 week ago (2 children)
[–] CHKMRK@programming.dev 5 points 1 week ago

Never really was

[–] ddplf@szmer.info 4 points 1 week ago (8 children)

Wait, since when it had not been? Or are you telling me that vastly the fastest growing platform in history with multiple payment gates (subscriptions, pay per token, licensing etc.) was not profitable for some reason?

[–] winterayars@sh.itjust.works 5 points 1 week ago

It's following the Amazon monopolization model.

[–] catloaf@lemm.ee 4 points 1 week ago (1 children)

Running those datacenters is extremely expensive.

[–] Flocklesscrow@lemm.ee 3 points 1 week ago* (last edited 1 week ago) (2 children)

The cost is to the whole world, because they consume enormous amounts of energy and produce essentially nothing. Like bitcoin miners.

[–] TheDoctor@hexbear.net 3 points 1 week ago

Worse than Bitcoin miners, AI seems to have the wholethroated support of capital (rather than a single faction), who see it as the next big form of automation

load more comments (1 replies)
[–] facow@hexbear.net 4 points 1 week ago

Or are you telling me that vastly the fastest growing platform in history with multiple payment gates (subscriptions, pay per token, licensing etc.) was not profitable

Are you not aware that 99 times out of 100 if you see a tech company rapidly growing it's completely unprofitable and not even attempting to be profitable yet? It's called blitzscaling and is pretty clearly what openai is attempting. Like if you see a tech company quickly growing you should be assuming it's unprofitable until proven otherwise not the opposite lol.

[–] sp3tr4l@lemmy.zip 3 points 1 week ago (4 children)

Not sure if you are joking but... it does not appear to be making anywhere near the amount of money that has been invested in it.

It costs a stupendous amount of money to develop the models, to train them, to rent out or just buy the hardware needed to do this, to pay for the electrical power to do this.

load more comments (4 replies)
load more comments (4 replies)
[–] ProgrammingSocks@pawb.social 4 points 1 week ago

No and AI almost never will be. However, investor money keeps coming, so it doesn't matter.

load more comments (3 replies)
load more comments
view more: next ›