you can stop and start it via systemctl and systemd is going to make mounts for fstab entries automatically, I just put local drives in my fstab so that way I can copy mount files between machines
With these systemd mount files I don’t need to touch the fstab, I can use ansible to copy the file, enable the service then start it. I can also have other services like Docker, Jellyfin or whatever to depend on that service. If the nfs share can’t be mounted then systemd won’t try to start docker.
With these systemd mount files I don’t need to touch the fstab, I can use ansible to copy the file, enable the service then start it. I can also have other services like Docker, Jellyfin or whatever to depend on that service. If the nfs share can’t be mounted then systemd won’t try to start docker.
So you can easily start and stop it as a service and you get your logging easily accessible via journalctl as a unit. But practically speaking there’s not much difference.
Yeap! You can even make an automount unit too! That way it’s mounted on demand! Makes life sooo much easier. I even do it for my external drives I use for backups
I’ll probably switch to simple script, since I don’t like the idea of my laptop shouting my NAS access credentials into any available random network on startup.
Here’s how to mount an nfs share:
#cat /etc/systemd/system/mnt.data.mount [Unit] Description=nfs mount script [Mount] What=192.168.0.30:/mnt/tank/Media Where=/mnt/data Type=nfs4 [Install] WantedBy=remote-fs.target
I’ve always mounted network shares in fstab, what’s the benefit to doing it with systemd?
(Also, for those of you learning, this method only works on systemd-based distros)
Removed by mod
you can stop and start it via systemctl and systemd is going to make mounts for fstab entries automatically, I just put local drives in my fstab so that way I can copy mount files between machines
With these systemd mount files I don’t need to touch the fstab, I can use ansible to copy the file, enable the service then start it. I can also have other services like Docker, Jellyfin or whatever to depend on that service. If the nfs share can’t be mounted then systemd won’t try to start docker.
Kinda interested now, why would you use systemd script for this instead of fstab ?
With these systemd mount files I don’t need to touch the fstab, I can use ansible to copy the file, enable the service then start it. I can also have other services like Docker, Jellyfin or whatever to depend on that service. If the nfs share can’t be mounted then systemd won’t try to start docker.
Oh that’s actually really good reason with docker containers that rely on the NFS share. Thanks, I’m gonna steal this
So you can easily start and stop it as a service and you get your logging easily accessible via journalctl as a unit. But practically speaking there’s not much difference.
Yeap! You can even make an automount unit too! That way it’s mounted on demand! Makes life sooo much easier. I even do it for my external drives I use for backups
Meanwhile I found a solution using fstab.
What’s the advantage of using a systemd script?
I’ll probably switch to simple script, since I don’t like the idea of my laptop shouting my NAS access credentials into any available random network on startup.
How would you do this with fstab? (Working with an smb share which I’m assuming is standard)
I described what I did here.
An now explain how to setup Kerberos
No idea, I don’t use Kerberos.