[-] rutrum@lm.paradisus.day 1 points 5 days ago

I've been using micromamba/mamba and not had solving issues like I did with conda. Im glad conda integrated libmamba.

Question: why were docker containers deemed security risks?

15

I just setup my first automated and encrypted backup with borg. It's got me thinking about other chaotic events, and how to respond accordingly. I figured now is a good time to document my infrastructure: hardware, network, a files. This way if something bad happens, like my house burns down, I or a family member has instructions for how to quickly recover data and services. Examples:

  1. If my website goes down, with my nextcloud on it, what steps do I need to take to recover the data and restore service?
  2. If my harddrive fails, how do I access lost data and reimplement redundancy after a replacement is stood up?
  3. If someone important to me needs to access encrypted files, how can that access that data and get access to the passwords/encryption keys?
  4. If my phone bricks, how to recover 2fa codes?

So I'd like to have a physical printing copy that tries to cover these emergency scenarios. Of course, I'll have digital copy around as well.

I'm focusing more on digital assets, like encryption keys, personal files and media, cloud service access, accessing inaccessible machines, how to restart/recover from self hosted service if its down, etc. I understand how much wider this document can be to include physical assets, so to start I want to start with digital infrastructure.

So my big questions: what scenarios should be documented in this disaster recovery document? What should I prepare for? The nice correlary of this is that documenting a recovery plan will force me to actually stand up the backups/redundancy needed to recover.

45
submitted 1 month ago by rutrum@lm.paradisus.day to c/linux@lemmy.ml

I just got a drawing tablet, and have been wanting some software that would allow me to work out math problems, draw architecture diagrams, etc. I've seen some tools like Excalidraw, which look handy for the sharing capabilities. I also have just used plain krita, which has great feedback for the pen sensitivity, but obviously is overkill for whiteboarding.

Are there any tools you use or recommend for handwriting or picture drawing? Pen or mouse?

[-] rutrum@lm.paradisus.day 34 points 2 months ago

Big fan of the reader mode changes. I'll probabky start using it more often, not just on sites with horrendous popups.

[-] rutrum@lm.paradisus.day 25 points 2 months ago* (last edited 2 months ago)

I've used this extension for krita, which lets you select part of the image and have an AI draw in your selection based on a prompt. It can work for outpainting, and inpainting, like removing a feature from an image (or adding one). You may have to do some prompt engineering to get the right outcome: https://github.com/Acly/krita-ai-diffusion?tab=readme-ov-file

[-] rutrum@lm.paradisus.day 25 points 2 months ago

Mull browser != mullvad browser, for those who were curious like I was. Mull Browser Source

35

I'm in desparate need of setting up borgmatic for borg backup. I would like to encrypt my backups. (I suppose, an unencrypted backup is better than none in my case, so I should get it done today regardless.)

How do I save those keys? Is there a directory structure I follow? Do you backup the keys as well? Are there keys that I need to write down by hand? Should I use a cloud service like bitwarden secrets manager? Could I host something?

Im ignorant on this matter. The most I've done is add ssh keys to git forges and use ssh-copyid. But I've always been able to access what I need to without keeping those (I login to the web interface.) Can you share with me best practices or what you do to manage non-password secrets?

17
submitted 4 months ago* (last edited 4 months ago) by rutrum@lm.paradisus.day to c/linux@lemmy.ml

TabbyML is a self-hosted code assistant. I have been unsuccessful at running it using my Nvidia GPU. There's two ways I've tried to deploy this.

As a docker container

Following the docs, it states I run the following docker run command. Below is what I run, modified to use the correct port:

docker run -it --gpus all \
  -p 11029:8080 -v $HOME/.tabby:/data \
  tabbyml/tabby serve --model StarCoder-1B --device cuda

Then I get the following error:

docker: Error response from daemon: could not select device driver "" with capabilities: [[gpu]].

So this would appear that I don't have the "nvidia-container-toolkit" installed on my machine. So I go ahead and enable this in nixos:

hardware.nvidia-container-toolkit.enable = true;

To validate that this works, I should be able to run nvidia-smi from within a container. I can run this from the host without issue:

$ nvidia-smi
Wed Jun  5 08:14:50 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.78                 Driver Version: 550.78         CUDA Version: 12.4     |
|-----------------------------------------+------------------------+----------------------+
...and so on

But if test this from a container, as the nvidia docs suggest as follows, I unable to access it from within the container.

$ sudo docker run --rm --runtime=nvidia --gpus all ubuntu nvidia-smi
docker: Error response from daemon: unknown or invalid runtime name: nvidia.

Okay, so I go and read the instructions further. Install instructions state that after installation, I need to configure the runtime like so:

$ sudo nvidia-ctk runtime configure --runtime=docker
sudo: nvidia-ctk: command not found

Ah nuts. That's a bug in nixos. I made a PR for this here: https://github.com/NixOS/nixpkgs/pull/317199 Still awaiting results from this. I don't know if this is a bug that will be backported to 24.05. Regardless, I wouldn't expect this ad-hoc configuration when I enable the nvidia-container-toolkit option in NixOS. Anyway, this option could still work but with some more time. If you have advice doing this let me know.

FOUND Docker method solution

So looking closer at people with the error message "no such runtime nvidia" I found this thread. It specifies that what nvidia-ctk is supposed to do is add a "runtime" that points to the nvidia-container-runtime executable. So I tried manually adding that my nixos configuration by using the virtualisation.docker.daemon.settings options. I was having trouble getting that working, because I needed to find the exact path to the nvidia-container-runtime executable. If you know Nix, you know that it isn't just in /usr/bin/.

But that's still not a satisfying solution anyway...I shouldn't have to this. I went in deeper and looked at module for nvidia-container-toolkit. This module calls a script called cdi-generate.nix. It outputs the results of nvidia-ctk to a file called nvidia-container-toolkit.json.

Let's go look for that file...can't find it. I do more searching...anyway, I found the solution.

The nvidia-container-toolkit is a new option in NixOS 24.05. It explicitly states in the release notes that it is supposed to replace the now deprecated virtualisation.{docker, podman}.enableNvidia options. Well, when you go look at the module that defines docker.enableNvidia you see it there at the bottom! This file actually defines the nvidia runtime!

And yes, it works. Using the now "deprecated" option is the one that actually works. I guess this is another bug to file to NixOS.

This seems to work so far, but I don't know why the solution using a NixOS module doesn't work either.

As a NixOS module

Let's just do it the full NixOS module way (which is what I tried first). That should be easy. Let's enable the feature and set some options:

services.tabby = {
    enable = true;
    port = 11029;
    acceleration = "cuda";
  };
  networking.firewall.allowedTCPPorts = [ 11029 ];

It appears to be working! VSCodium extension sees the server and prompts for a authentication token. I add the token. I type some code and set for a manual trigger...then tabby dies. Let''s look at the systemd logs.

tabby[76786]: ๐Ÿ“„ Version 0.11.1
tabby[76786]: ๐Ÿš€ Listening at 0.0.0.0:11029
tabby[76786]:   JWT secret is not set
tabby[76786]:   Tabby server will generate a one-time (non-persisted) JWT secret for the current process.
tabby[76786]:   Please set the TABBY_WEBSERVER_JWT_TOKEN_SECRET environment variable for production usage.
systemd[1]: tabby.service: Main process exited, code=exited, status=1/FAILURE
systemd[1]: tabby.service: Failed with result 'exit-code'.
systemd[1]: tabby.service: Consumed 2.285s CPU time, received 121.0K IP traffic, sent 1.6M IP traffic

That's it. It's not very descriptive about what happened. I've had success running it this way using the "cpu" option for acceleration (no GPU) but that's too slow to be useful.

GPU specs

I am running a Nvidia RTX 2060 and using the proprietary drivers version 550.

Thanks for the read, if you have any input on what to do next let me know what I can try. Ideally, I'd like to have both options work, since I think the docker implementation may have the same problem as the NixOS module option.

[-] rutrum@lm.paradisus.day 21 points 4 months ago

I think it overinflates the click rate, which means despite having more clicks on an ad, that doesnt mean that more people bought some product. This devalues click rate which might make the ad service less valuble to advertisers, so they dont spend as much on Google's ad service.

And in general I think makes any training data for a model more muddy, since adnauseum isnt behaving like a human. So it could make it more difficult to train models that do targeted advertising.

[-] rutrum@lm.paradisus.day 23 points 4 months ago

As a community, I do think we get hungup on distros. Most of them, as you mentioned, are just different defaults of the same packages.

But at the maintainer level, I do think theres a lot of work distributions do at making sure the software they choose as defaults are up to date, secure, and work with one another. I dont enounter it often, but relying on maintainers to prevent mismatched depencies ending up in the day-to-day linux user has to be worth something. And every set of defaults needs that level of assurance, I would think. Im not a maintainer, I could be off here.

24
"No code" databases (lm.paradisus.day)
submitted 4 months ago* (last edited 4 months ago) by rutrum@lm.paradisus.day to c/selfhosted@lemmy.world

I've been seeing easy ways to store and view tabular data. I'm aware of tools like nocodb, baserow, and mathesar. I'm currently playtesting nocodb. But I wanted to start a discussion on what everyone uses for easily storing tabular data, and if anyone uses these tools.

I've also tried nextcloud tables but it still is very early in development from what I can tell.

[-] rutrum@lm.paradisus.day 21 points 5 months ago

Do you self host or are you running a nextcloud-managed instance?

[-] rutrum@lm.paradisus.day 58 points 5 months ago

Sometimes the app just shows a barcode that they scan. I always screenshotted the barcode and deleted the app. Better yet, save the barcode in catima https://catima.app/

[-] rutrum@lm.paradisus.day 33 points 7 months ago

To be fair, you're taking on a lot of new things at once. You can spin up docker containers on windows too, all while using a UI. I think it's great your exposing yourself to self hosting, linux, command line interface, and containerization all at once, but don't beat yourself up for it taking longer than expected. A lot of it takes time. I encourage you to keep trying and playing. Good luck!

[-] rutrum@lm.paradisus.day 35 points 9 months ago

Theres so many. Check out the awesome list: https://github.com/awesome-selfhosted/awesome-selfhosted

I think your stategy should be one service at a time. Do everything in docker, and start by tackling a simpler service. For example, you should try paperless-ngx. Absolute game changer. I didnt realize how much managing ny own directory structure sucked until I used this. Then, grow your service list more and more!

229
submitted 10 months ago* (last edited 10 months ago) by rutrum@lm.paradisus.day to c/linux@lemmy.ml

You know, ZFS, ButterFS (btrfs...its actually "better" right?), and I'm sure more.

I think I have ext4 on my home computer I installed ubuntu on 5 years ago. How does the choice of file system play a role? Is that old hat now? Surely something like ext4 has its place.

I see a lot of talk around filesystems but Ive never found a great resource that distiguishes them at a level that assumes I dont know much. Can anyone give some insight on how file systems work and why these new filesystems, that appear to be highlights and selling points in most distros, are better than older ones?

Edit: and since we are talking about filesystems, it might be nice to describe or mention how concepts like RAID or LUKS are related.

[-] rutrum@lm.paradisus.day 21 points 11 months ago

Actually cd isnt a program. Your present working directory is managed entirely by the shell. If you type "type cat" in the terminal it will tell you its a program, but "type cd" says its a shell builtin. So yes, cd depends on the shell and zsh has some awesome quality of life features. This is not something you can do in bash.

250
submitted 11 months ago by rutrum@lm.paradisus.day to c/linux@lemmy.ml

Dust is a rewrite of du (in rust obviously) that visualizes your directory tree and what percentage each file takes up. But it only prints as many files fit in your terminal height, so you see only the largest files. It's been a better experience that du, which isn't always easy to navigate to find big files (or atleast I'm not good at it.)

Anyway, found a log file at .local/state/nvim/log that was 70gb. I deleted it. Hope it doesn't bite me. Been pushing around 95% of disk space for a while so this was a huge win ๐Ÿ‘

97
submitted 11 months ago by rutrum@lm.paradisus.day to c/privacy@lemmy.ml

I came across privacy.com, a service that generates virtual credit cards, like aliases for your real credit card that can be paused or discarded at any moment.

My own credit card company has this feature. But it requires a browser plugin that so obviously is there to track my spending habits, so I've not wanted to consider it. Privacy.com looks like a great alternative.

But is it even worth it? It may be a hastle, but I can also cancel my actual credit card at any moment and they will send me a new number immediately and a card a few days later. From a privacy prospective, how much can a company use my credit card credentials to track me? Maybe a third-party virtual card provider even masks my own purchases so not even my credit card company knows? Not sure about that one.

Please share if you use one, who its with, and if its worth it.

[-] rutrum@lm.paradisus.day 77 points 1 year ago

You'd have to explain how gimp doesnt suit your needs, because in the open source world its best in class for photo editing.

4

I've bought Pine64 products before (pinephone and recently the pinetime) but I've heard Zach Freedman (void star labs) mention this before and rave about it. I wanted to hear from the ergo community what they thought about it before I bought one to start my keyboard building journey.

In addition, if I should buy one, what tips do you recommend be purchased? It looks like I can buy long/short versions of fine/gross tips. What's best for keyboard building? Here's a link to their pinecil products.

view more: next โ€บ

rutrum

joined 1 year ago