From their release notes:
[...] Note that this feature is completely optional and no AI related code is even loaded until you configure an AI provider.
I'm glad for that part.
A community for everything relating to the GNU/Linux operating system (except the memes!)
Also, check out:
Original icon base courtesy of lewing@isc.tamu.edu and The GIMP
From their release notes:
[...] Note that this feature is completely optional and no AI related code is even loaded until you configure an AI provider.
I'm glad for that part.
I... kinda don't care about this. It does nothing if not enabled and its super barebones as far as I can tell.
This is why I hate how everything is AI now.
Generative AI is absolute trash.
A fancy suped up search engine using AI to check files and answer questions about the files, whatever.
What would it be doing if I enabled it?
And can it be configured to use local(ly installed) AI?
Probably comprehend sections of the text.
Regarding your 2nd question: Yes, you can e.g. use a local Ollama instance.
I thought it would be useful for filling and finding metadata, but I don't think it can do that
Why the fuck would i want that?
Users can highlight any text and ask AI models questions about it, receiving explanations, context, or summaries on the spot.
I can see this being particularly useful for autistic people who don't understand a very poetic section, or for people reading a text which is not their first language.
Then don't enable it, I hate the AI but there are people who want it and calibre implemented it the right way. Its opt in, you have to configure AI provider first before that no code related to ai is loaded. That's best way to implement AI.
I don’t really see a use case for it on calibre. Maybe others are more creative than I am and can tell me what they would use it for.
As far as I see it on the release notes it is completely optional and nothing gets loaded until the Provider is defined and activated so it’s peachy for me.
Well, have you ever read something and went "what the hell is this?", "I don't get this", "what is an abubemaneton?". Well, now you'll be able to quickly ask AI about it.
I would always prefer a dictionary or a Wikipedia over a fucking LLM.
But you can't always use a dictionary to understand the use of words or phrases in novel contexts. For instance, maybe a phrase contains meaning not in its literal text, but in its resemblance to a quotation from another work, or to a previous quotation within the same work.
I've taken a classic Swedish poem ("Bron" by Erik Lindorm), and I can't seem to find an explanation of what it means via traditional searching. But when I paste it into ChatGPT and ask what it means, it gives a detailed interpretation.
A human could do the same, but it's unreasonable to expect every learner has a human on standby to cater to their every educational whim at all hours of the day.
LLMs will tell you anything, and if you can't fact check it, you will never know if it's true or not.
If you're fine with believing in random words put together nicely, I can't stop you, but I won't cheer you on, either.
Your own reference explains how the AI responses are more than just random words put together nicely, and I can and have fact-checked it. It's not a trustworthy tool for many things, but it is useful for language-based pursuits, because language is precisely what it's designed to work with.
For example, I've recently been watching a Chinese period drama and asked ChatGPT what various words were by transcribing how I heard them and explaining the context. It gave me accurate hanzi, pinyin, and definitions as confirmed by dictionary sites. It's been a very valuable tool to me for language learning.
It's a token-based model. Not exactly language.
I'm happy that you feel like it's helping you.
But never trust it.
I remember having an Oxford Dictionary CD as a child (got it with the physical copy).
Unfortunately, it stopped working long ago (and I didn't rip it), but while it did work, I had quite a lot of fun reading up on word-origins, synonyms/antonyms, pronunciations and whatnot.
I'd honestly rather be able to connect something like that to Calibre (and other programs) with DBus, rather than use AI for a definition. And that was just a single CD (I can be sure, because I didn't have a DVD reader).
So, perhaps some other use case?
AI is not only capable of definitions. In fact... You wouldn't use it for that. But It's terribly good at context. So it can interpret a whole phrase, or paragraph. Maybe calibre even passes the book metadata so it can infer characters, places and broader context.
Yeah, that won't really be doable just by an extended dictionary.
I myself tend to use Google sometimes, to look for stuff like "one word for the phrase ..." and most of the times the AI is the one giving the answer.
QuickDic's default databases are compiled from Wiktionary entries, and Wiktionary seems like the most reliable part of Wikipedia currently. Wonder then if that couldn't be used also. On QuickDic, having all databases installed takes a bit over 1 GB, not much for desktop standards afaik.
And then the AI can give you some made up and incorrect answer. Hooray!
Sure, I guess, but I could just as easily ask chatgpt in the chatgpt app, I feel.
This just makes it faster, and convenient, you don't have to get out of your book.
AI slop has no place in books.
fuck that shit!
So if it only lets the llm see highlighted text, whats the point of even adding it into calibre. It takes 0 extra seconds to paste that text into google or chatgpt or duck.ai or whatever
You know it doesn't take 0 seconds. Record yourself. If you are studying or something (and need AI comments somehow), it's a good feature.
There’s also a small mental cost to switching contexts, from reader to browser.
This is actually one of the better uses of AI I’ve seen because it is literally asking a large language model about language.
I probably still won’t use it, but at least it makes sense!
Just what nobody wanted in their eReading software.