I just checked myself and it’s card1 too, now I am curious why it’s not card0. 🤷
I just checked myself and it’s card1 too, now I am curious why it’s not card0. 🤷
I believe it’s cat /sys/class/drm/card0/device/pp_power_profile_mode
.
There’s also the power_dpm_force_performance_level
.
As a small homelabber I agree with this. I started with a baremetal and using Docker, and switched to Proxmox, and now over to Incus, actually currently I am using Debian with cockpit + cockpit-machines. I do like Incus, I keep hopping back and forth between cockpit, I need to settle on one.
Yeah, Debian has older firmware found in the firmware-amd-graphics
package which doesn’t include the firmware. You’ll need to download it from https://git.kernel.org/pub/scm/linux/kernel/git/firmware/linux-firmware.git/ I believe anything from linux-firmware-20231030 and newer should work.
20231030 tag: https://git.kernel.org/pub/scm/linux/kernel/git/firmware/linux-firmware.git/tree/amdgpu?h=20231030 or newest(20240410) tag: https://git.kernel.org/pub/scm/linux/kernel/git/firmware/linux-firmware.git/tree/amdgpu?h=20240410
These files need to be placed in /lib/firmware/amdgpu/
It’s a relatively low performance hit and it benefits me when having to replace a failing/old disk. I can just toss the drive without having to erase the data first, that is as long as the key is a secure length.
grass
is on the AUR… https://aur.archlinux.org/packages/grass
~53 W
Server:
Mini PC: Beelink S12 N95
8 port unmanaged TP Link switch
I would like to expand my storage, however I don’t have any available SATA ports and I believe adding an HBA would increase the idle draw about 8 W. I might just upgrade the SSDs and split the storage between the HDDs and SSDs.
I recently switched from Proxmox to Debian Bookworm with Incus(LXD fork) as my primary setup, it’s been a pleasant experience. I also like the idea of using something like Cockpit to manage VMs though haven’t come to a need yet for a VM over a container. I’ll also point out that Incus can handle VMs as well.
Stéphane Graber, Project leader of Linux Containers is also on the fediverse and responds to questions often.
That’s a good recommendation, thanks!
There’s even an RSS feed for that page: https://openwrt.org/feed.php?mode=list&ns=advisory
You’re welcome! Also thanks for asking this question, I hadn’t seen ShotShare before, it looks useful.
No, since you are using the bind mount, you do not need to use the volume.
I just did another test.
You should be able to create the directories manually. I cheated by simply cloning the repo and copying them to the bind mount location like so. You can use the bind mount method like you wanted.
git clone https://github.com/mdshack/shotshare
cp -r shotshare/storage/* /srv/dev-disk-by-uuid-7fe66601-5ca0-4c09-bc13-a015025fe53a/Files/Shotshare/shotshare_data/
chown 82:82 -R /srv/dev-disk-by-uuid-7fe66601-5ca0-4c09-bc13-a015025fe53a/Files/Shotshare/shotshare_data
It will be stored in /var/lib/docker/volumes
, you can find the exact location by inspecting the volume. Use docker volume ls
to list the volumes, and do docker volume inspect <volume_name>
replacing <volume_name> with the one from the list. Look for “Mountpoint”, that is the exact location. You could try copying that to bind mount location, though I can’t be sure if it will continue to work.
This appears to be the exact same problem as https://github.com/mdshack/shotshare/issues/31
For testing I just spun up a VM with Docker, I tried the same compose file as you. I found I had to use the volume instead of a bind mount for /app/storage
.
This compose file should work.
version: "3.3"
services:
shotshare:
ports:
- 2000:80
environment:
- HOST=:80
- ALLOW_REGISTRATION=false
volumes:
- shotshare_data:/app/storage
- /srv/dev-disk-by-uuid-7fe66601-5ca0-4c09-bc13-a015025fe53a/Files/Shotshare/database.sqlite:/app/database/database.sqlite
- /srv/dev-disk-by-uuid-7fe66601-5ca0-4c09-bc13-a015025fe53a/Files/Shotshare/.env:/app/.env
restart: unless-stopped
container_name: shotshare
image: mdshack/shotshare:latest
volumes:
shotshare_data:
networks: {}
That error message says it the permissions of the /home/user/Documents/Docker/LinguaCafe/logs
directory. You can try changing it full r/w temporarily to test.
I like it, then it’s even harder to know that it was encrypted in the first place. Thanks for that suggestion.
You will need to create a network bridge: https://www.truenas.com/docs/scale/23.10/scaletutorials/virtualization/accessingnasfromvm/ Also more information in the Setting up NFS for Portainer section of this guide: https://forum.level1techs.com/t/truenas-scale-ultimate-home-setup-incl-tailscale/186444
While this wouldn’t work for you now, something to think about is encrypting new drives going forward so that you don’t have to worry about erasing/zeroing them, just toss the encryption key and your good to go.
It looks like they are working on fixing that with this pull request.