TheHobbyist

joined 1 year ago
[–] TheHobbyist@lemmy.zip 8 points 1 day ago (1 children)

Indeed, totally an Apple approach to modularity: it is a proprietary Apple SSD...

[–] TheHobbyist@lemmy.zip 11 points 3 days ago

I see, so there is indeed a broader context to the burning alone, it was also with additional verbal hatred and then possibly the location, and the overall intention. I think this makes it clearer. Thanks

[–] TheHobbyist@lemmy.zip 28 points 3 days ago (14 children)

Not familiar with the guy himself who maybe does deserve criticism and prison, but about the Quran burning, is it genuinely fair to sentence someone to prison for that? Is it equivalent to burning the cross? The Swedish flag? I might be mission a broader context, but I don't feel like someone burning my symbol or flag should be punished with prison. Am I alone? I would hate it, don't get me wrong, but I still feel it goes in freedom of expression.

[–] TheHobbyist@lemmy.zip 1 points 4 days ago* (last edited 4 days ago)

The proud dad's name ends with Unis and the kid remembers the X first digits, hence Unix, hence Linux!

[–] TheHobbyist@lemmy.zip 1 points 4 days ago

They do mention compatibility a lot, if it's hardware, I agree with you. But perhaps they mean something else?

 

cross-posted from: https://lemmy.world/post/21664063

Linux Mint and Framework Laptops Join Forces

The October 2024 edition of Linux Mint’s Monthly News brings exciting updates, including a significant announcement about collaboration with Framework Laptops, having potential to advance Mint’s compatibility with hardware designed with flexibility, repairability, and sustainability in mind.

For those unfamiliar, unlike most traditional laptops, which are often difficult or impossible to repair or upgrade, Framework laptops are built to be user-friendly, making it easy to replace or upgrade components. This modular approach extends the laptop’s lifespan and promotes sustainability by reducing e-waste.

[–] TheHobbyist@lemmy.zip 51 points 4 days ago

This. I will resume my recommendation of Bitwarden.

[–] TheHobbyist@lemmy.zip 3 points 4 days ago (1 children)

I didn't say it can't. But I'm not sure how well it is optimized for it. From my initial testing it queues queries and submits them one after another to the model, I have not seen it batch compute the queries, but maybe it's a setup thing on my side. vLLM on the other hand is designed specifically for the multi co current user use case and has multiple optimizations for it.

[–] TheHobbyist@lemmy.zip 22 points 4 days ago* (last edited 3 days ago) (5 children)

I run the Mistral-Nemo(12B) and Mistral-Small (22B) on my GPU and they are pretty code. As others have said, the GPU memory is one of the most limiting factors. 8B models are decent, 15-25B models are good and 70B+ models are excellent (solely based on my own experience). Go for q4_K models, as they will run many times faster than higher quantization with little performance degradation. They typically come in S (Small), M (Medium) and (Large) and take the largest which fits in your GPU memory. If you go below q4, you may see more severe and noticeable performance degradation.

If you need to serve only one user at the time, ollama +Webui works great. If you need multiple users at the same time, check out vLLM.

Edit: I'm simplifying it very much, but hopefully should it is simple and actionable as a starting point. I've also seen great stuff from Gemma2-27B

Edit2: added links

Edit3: a decent GPU regarding bang for buck IMO is the RTX 3060 with 12GB. It may be available on the used market for a decent price and offers a good amount of VRAM and GPU performance for the cost. I would like to propose AMD GPUs as they offer much more GPU mem for their price but they are not all as supported with ROCm and I'm not sure about the compatibility for these tools, so perhaps others can chime in.

Edit4: you can also use openwebui with vscode with the continue.dev extension such that you can have a copilot type LLM in your editor.

[–] TheHobbyist@lemmy.zip 3 points 4 days ago* (last edited 4 days ago)

As you probably know, an LLM works iteratively: you give it instructions and it "auto-completes", one token at a time. Every time you want to generate the next token, you have to perform the whole inference task, which is expensive.

However, verifying if a next token is the correct one, can be cheap because you can do it in parallel. For instance, take the sentence " The answer to your query is that the sky is blue due to some physical concept". If you wanted to check whether your model would output each one of those tokens, you would split the sentence after every token and you could batch verify the next token for every split and see whether the next token matches the sentence.

Speculative decoding is the process where a cheap and efficient draft model is used to generate a tentative output, which is then verified in parallel by the expensive model. Because the cheap draft model is many times quicker, you can get a sample output very fast and batch verify the output with the expensive model. This saves a lot of computational time because all the parallel verifications require a single forward pass. And the best part is that it has zero effect on the output quality of the expensive model. The cost is that you know have to run two models, but the smaller one may be a tenth of the size, so runs possibly 10x faster. The closer the draft model output matches the expensive model output, the higher the inference speed gain potential.

[–] TheHobbyist@lemmy.zip 1 points 4 days ago* (last edited 3 days ago) (2 children)

This is interesting. Need to check if this is implemented in Open-WebUI.

But I think the thing which I'm hoping for most (in open-webui), is the support of draft models for speculative decoding. This would be really nice!

Edit: it seems it's not implemented in ollama yet

[–] TheHobbyist@lemmy.zip 4 points 4 days ago

Can't you return the laptop within 30 days if you don't like it? If that's the case, why don't you just go ahead, buy it and give it a reasonable shot? Nobody else's opinion will change how the laptop works for you :)

 

I landed on this article today about how to configure vLLM for AMD GPUs and it contained this specific snippet:

Meta recently announced they’re running 100% of their live Llama 3.1 405B model traffic on AMD MI300X GPUs [...]

I thought that was an interesting piece of trivia and shows both how important it is for industrial partners to find alternatives to Nvidia and also how much AMD has improved its software suite to enable these use-cases.

30
submitted 1 month ago* (last edited 1 month ago) by TheHobbyist@lemmy.zip to c/linux@lemmy.ml
 

Hi folks,

I have Alpine Linux installed in an encrypted LUKS partition. I came across this tutorial which shows how to setup a key in a USB drive and when the drive is inserted and the computer booted, the LUKS partition auto-unlocks with the key on the USB drive.

https://askubuntu.com/questions/1414617/configure-ubuntu-22-04-zfs-for-automatic-luks-unlock-on-boot-via-usb-drive

I would like to setup the same thing but I do not have Alpine linux installed on ZFS, so I'm looking for ways to adapt the instructions.

So far, what I've done is:

  1. I've setup the key on the usb stick and I can unlock the LUKS partition with that key.
  2. create a /etc/mkinitfs/features.d/usb-unlock.sh script with the following content:

(the echo to /dev/kmesg was to check whether the script did indeed run at boot by trying to print to the kernel messages but I can't find anything in the kernel messages).

#!/bin/sh

echo "usb-unlock script starting..." > /dev/kmsg

USB_MOUNT="/mnt/my-usb-key" # The USB stick mounting point
LUKS_KEY_FILE="awesome.key"  # The name of your keyfile on the USB stick

# Search for the USB stick with the key
for device in $(ls /dev/disk/by-uuid/*); do
    mount $device $USB_MOUNT 2>/dev/null
    if [ -f "$USB_MOUNT/$LUKS_KEY_FILE" ]; then
        # Unlock the LUKS partition
        cryptsetup luksOpen /dev/sda3 cryptroot \
            --key-file "$USB_MOUNT/$LUKS_KEY_FILE" && exit 0
    fi
    umount $USB_MOUNT
done
echo "No USB key found, falling back to password prompt." # this message never appears, despite not having found the key on the usb stick

echo "usb-unlock script ending." > /dev/kmsg
  1. I added usb-unlock to the features in mkinitfs.conf:
mytestalpine:~# cat /etc/mkinitfs/mkinitfs.conf 
features="ata base ide scsi usb virtio ext4 cryptsetup keymap usb-unlock"
  1. run mkinitfs to rebuild the initramfs. Then reboot to test the implementation, which was unsuccessful.

What am I missing / doing wrong? Thank you for your help!

Edit: forgot to add step 4

 

Hi folks,

I'm seeing there are multiple services which externalise the task of "identity provider" (e.g. login with Facebook, google or what not).

In my case, I am curious about Tailscale, a VPN service which allows one to chose an identity provider/SSO between Google, Microsoft, Github, Apple and OIDC.

How can I find out what data is actually communicates to the identity provider? Their task should simply be to decide whether I am who I claim to be, nothing more. But I'm guessing there may be some subtleties.

In the case of Tailscale, would the identity provider know where I'm trying to connect? Or more?

Answers and insights much appreciated! The topic does not seem to have much information online.

 

Hi folks, I'm considering setting up an offsite backup server and am seeking recommendations for a smallish form factor PC. Mainly, are there some suitable popular second hand PCs which meet the following requirements:

  • fits 4x 3.5" HDD
  • Smaller than a regular tower (e.g. mATX or ITX)
  • Equipped with a 6th of 7th gen Intel CPU at least (for power efficiency and transcoding, in case I want it to actually to some transcoding) with video output.
  • Ideally with upgradeable RAM

Do you know of something which meets those specs and is rather common on the second hand market?

Thanks!

Edit: I'm looking for a prebuilt system, such as a dell optiplex or similar.

 

Yesterday, there was a live scheduled by Louis Grossman, titled "Addressing futo license drama! Let's see if I get fired...". I was unable to watch it live, but now the stream seems to be gone from YouTube.

Did it air and was later removed? Or did it never happen in the first place?

Here's the link to where it was meant to happen: https://www.youtube.com/watch?v=HTBYMobWQzk

Cheers

Edit: a new video was recently posted at the following link: https://www.youtube.com/watch?v=lCjy2CHP7zU

I do not know if this was the supposedly edited and reuploaded video or if this is unrelated.

 

DeepComputing is preparing a RISC-V based motherboard to be used in existing Framework Laptop 13s!

Some snippets from the Framework blog post (the link to which is provided below):

The DeepComputing RISC-V Mainboard uses a JH7110 processor from StarFive which has four U74 RISC-V cores from SiFive.

This Mainboard is extremely compelling, but we want to be clear that in this generation, it is focused primarily on enabling developers, tinkerers, and hobbyists to start testing and creating on RISC-V.

DeepComputing is also working closely with the teams at Canonical and Red Hat to ensure Linux support is solid through Ubuntu and Fedora.

DeepComputing is demoing an early prototype of this Mainboard in a Framework Laptop 13 at the RISC-V Summit Europe next week.

Announcement: https://frame.work/blog/introducing-a-new-risc-v-mainboard-from-deepcomputing

The upcoming product page (no price/availability yet): https://frame.work/products/deep-computing-risc-v-mainboard

Edit: Adding link the the announcement by DeepComputing: https://deepcomputing.io/a-risc-v-world-first-independently-developed-risc-v-mainboard-for-a-framework-laptop-from-deepcomputing/

28
submitted 7 months ago* (last edited 7 months ago) by TheHobbyist@lemmy.zip to c/localllama@sh.itjust.works
 

From Simon Willison: "Mistral tweet a link to a 281GB magnet BitTorrent of Mixtral 8x22B—their latest openly licensed model release, significantly larger than their previous best open model Mixtral 8x7B. I’ve not seen anyone get this running yet but it’s likely to perform extremely well, given how good the original Mixtral was."

 

Hi all,

I think around 1 or 2 years ago, I stumbled upon a personal blog of an asian woman (I think) working at OpenAI. She had numerous extensive fascinating blog posts on a black themed blog, going into the technical details of embeddings of language models and such.

I can no longer find that blog and have no other information to go by. Would anyone possibly know which blog I'm referring to? It would be very much appreciated.

 

Hi folks,

I seem to be having some internet connectivity issues lately and I would like to monitor my access to the internet. I have a homelab and was wondering whether someone had perhaps something like a docker container which pings a custom website every so often and plots a timescale of when the connection was successful and when it was not.

Or perhaps you have another suggestion? I know of dashboards like grafana but I don't know whether they can be configured to actually generate that data or whether they rely on a third party to feed them. Thanks!

 

Just wanted to share my appreciation of the game.

I grabbed a copy of this game a year ago, taking advantage of a sale and ahead of the massive update. Then forgot about it, never touched it.

Fast forward a year later, and now I got a steam deck and decided to dive into the game. I love it. I'm just a few hours in but I can already say this is among my favorite games. The broad openness of the world, the level of detail, the characters, the interactive dialogs, the items, the strategies, the game mechanics. It's a very involved game. It really is up there. Thank you CDPR for this game and this remake.

 

I was exploring the fps and refresh rate slider and I realized that when setting the framerate limiter to 25, the refresh rate was incorrectly set to 50Hz on the OLED version, when the 75 Hz setting would be a more appropriate setting, for the same reason 30 fps is at 90 Hz and not 60 Hz. Anyone else seeing the same behavior? Is there an explanation I'm missing here?

view more: next ›