andrew0

joined 1 year ago
[–] andrew0@lemmy.dbzer0.com 5 points 2 weeks ago (1 children)

If I am not mistaken, the difference was that the Internet Archive was distributing books with a DRM that would make the PDF unusable after a certain time. You could relate it to how a physical library offers books for a limited time, for free. Now, of course, one could bypass the DRM or copy the contents differently, but so can another person photocopy a book they borrowed physically. Meanwhile, other physical libraries are allowed to distribute e-books, but I'm not sure if that's made possible due to licensing fees.

I'm not saying that they approached this well, especially given the copyright laws in the US, but it was indeed a good thing for the normal person at the time. Too bad that the judicial system in the US is biased towards leeching companies. I really can't wait to see the AI vs publishers fight, though. Let's see who has deeper pockets and better plants in the courts :D

[–] andrew0@lemmy.dbzer0.com 13 points 3 months ago* (last edited 3 months ago)

Organic Maps. Make sure you download the areas beforehand. The resolution for walking paths is much nicer, and you can clearly see the routes of temples and other points of interest.

It also shows data about drinking faucets, toilets, and other things provided by users.

[–] andrew0@lemmy.dbzer0.com 3 points 4 months ago (1 children)

Good luck! You can try the huggingface-chat repo, or ollama with this web-ui. Both should be decent, as they have instructions to set up a docker container.

I believe the Llama 3 models are out there in a torrent somewhere, but I didn't dig to find it. For the 70B model, you'll probably need around 64GB of RAM available, but the 7B one should run fine with just 8GB. It will be somewhat slow though, compared to the ChatGPT experience. The self-attention mechanism can be parallelized, which is why you will see much better results on a GPU. According to some others that tested it, if you offload some stuff to RAM, you could see ~10-12 tokens per second on an RTX 3090 for certain 70B models. But more capable ones will be at less than 1 token per second, all depending on the context window you use.

If you don't have a GPU available, just give the Phi-3 model a try :D If you quantize it to 4 bits, it can apparently get 12 tokens per second on an iPhone haha. It should play nice with pooling information from a search engine, or a vector database like milvus, qdrant or chroma.

[–] andrew0@lemmy.dbzer0.com 4 points 4 months ago (3 children)

What db2 already said. Microsoft just released Phi-3 mini, which could, allegedly, run locally on newer smartphones.

If I understood correctly, the Rabbit thingy just captures your information locally and then forwards it to their server. So, if you want more power, you could probably do the same by submitting the same info to a bigger open source model than Phi-3, like Llama 3, hosted on your homelab. I believe you can set it up with huggingface/gradio, which sort of provides an API that you could use.

That way, you don't need a shitty orange box, and can always get the latest open source models with a few lines of code. There are plenty of open source frameworks in the works at the moment, and I believe that we're not far off from having multi-modal LLMs running on homelab-level hardware (if you don't mind a bit of lag).

[–] andrew0@lemmy.dbzer0.com 3 points 4 months ago

That is good to know. Tried the free version of Roll20 before, and it definitely felt lacking in certain areas. Oh, and thanks for letting me know about the sale! I'll definitely keep an eye out for that one :)

[–] andrew0@lemmy.dbzer0.com 26 points 5 months ago* (last edited 5 months ago) (2 children)

How will you move to WhatsApp if everyone else uses iMessage? Europe has the same issue, but reversed. Everyone uses WhatsApp and can't jump to Signal/Telegram because they're not as popular.

[–] andrew0@lemmy.dbzer0.com 1 points 5 months ago

I got NFS Most Wanted (2005) working in Wine, and was somewhat impressed how easy it was at the time. Game worked quite well, and would only crash once in a while with some cryptic errors that I don't remember. Made me hopeful for the future of linux gaming :)

[–] andrew0@lemmy.dbzer0.com 30 points 8 months ago (1 children)

Wow, some of the comments on that article saying Google should have made Android closed source are mindboggling. They realize they never would have had their current worldwide marketshare if they did that, no?

But maybe if they did, we would have had more people working on true linux phones 🤔 I'm a bit torn on this one haha.

[–] andrew0@lemmy.dbzer0.com 7 points 8 months ago (1 children)

The Framework 13 inch model should be plenty, especially if you want to dev on the go. Much more lightweight and smaller, and you can connect it to external monitors if the screen size is not big enough. Also, you shouldn't have issues running Linux on either laptops.

Instead of going for the 16 version, I would use the extra 900-1000 euros (that's the amount I saw I could save between the two almost maxed-out models) to make a dedicated server or mini-cluster to run your workloads. Deploy Kubernetes or Proxmox on it, and you'll also get some more practice on it outside work if you want to run stuff for your home lab. That is only if you don't want to game on your laptop, but I'd still put that money aside to make a desktop.

 

Hello everyone! I've been playing around with Wayland for a bit and was hoping to start learning some more about it. For example, I would be interested in making a lock screen, similar to Swaylock, as a toy project.

What GUI toolkit would you use to develop apps on Wayland? I've added a little poll below with some of the popular choices I've seen thrown around. Feel free to add your own suggestions and maybe leave a comment as to why you'd use that!

Link to poll

[–] andrew0@lemmy.dbzer0.com 58 points 9 months ago (1 children)

It's amazing that Linux gaming is becoming a thing that's better sometimes than Windows gaming (minus the getting banned part in some games). I also like that AMD is making some big pushes on open source drivers, plus their ROCm open-source alternative to CUDA.

This is a great time for Linux users! :)

[–] andrew0@lemmy.dbzer0.com 7 points 9 months ago

Same. It sucks that most banks wouldn't jump on this train :(

[–] andrew0@lemmy.dbzer0.com 83 points 10 months ago (1 children)

What a stupid article. It's like saying "stop using electric vehicles because you can't use gas stations". I don't understand why he's so adamant about this? It's not like Wayland had about 20 years of extra time to develop like X11. People keep working on it, and it takes time to polish things.

44
Jump from Arch to NixOS? (lemmy.dbzer0.com)
submitted 1 year ago* (last edited 1 year ago) by andrew0@lemmy.dbzer0.com to c/linux@lemmy.ml
 

As the title implies, should I do it? I love Arch so far, and I can fix most issues that pop out. However, I sometimes wish to start fresh without too much hassle, but I get a feeling NixOS isn't as mature as Arch.

Have any of you used both, and if so, what do you miss from Arch? What are you grateful for in NixOS?

1
submitted 1 year ago* (last edited 1 year ago) by andrew0@lemmy.dbzer0.com to c/piracy@lemmy.dbzer0.com
 

Hi everyone! I'll soon take the DP-100 exam for Microsoft Azure, and I was interested in finding more leaked exam questions. At the moment, I was using examtopics for this, but it sucks because it basically cuts you off halfway through.

I heard there are some private trackers that specialize in exam questions, such as LearnFlakes, but I do not have anyone that can invite me to them. Therefore, I was wondering if there is another way to find the information I need for this exam.

Do you know any other sources that are fully free?

view more: next ›