jowe

Members
  • Posts

    122
  • Joined

  • Last visited

About jowe

  • Birthday 11/11/1978

Converted

  • Gender
    Male
  • Location
    Sweden
  • ICQ
    174543

Recent Profile Visitors

1993 profile views

jowe's Achievements

Apprentice

Apprentice (3/14)

3

Reputation

  1. No problem, just glad i could help!
  2. I'm not entirely sure what you are trying to do. But unraid dont need to have an IP on that VLAN. If you dont want it to. If unraid needs to talk to CCTV VLAN, it should be done over the firewall. As long as the switch port you are connecting the NVR to are configured on the correct vlan, and the device have the correct IP/mask. It should work. A good start is to ping the firewall and se if you get a block in the pfsense log. Then atleast you have a conection to the firewall.
  3. I get the errors instead if i remove the plugin, so its more of an unraid thing, it tries to spin down the drive again and again... Not a problem for me as i have the disks to never spin down now.
  4. Yes I understand that the message should display, when timer reaches for example 15min. Or hitting spin down button. I did not push the button at all, just waited 15min, and the message loops every second. I could have provided a really long list with same messages. Yes I know, just wanted to give it a go with the latest FW, but unfortunately it didn't work. Nevertheless it's a great project!
  5. Really good work @doron! Just tried the commands against my SAS drives after upgrading firmware of the SAS controller. But "sg_start --readonly --pc=3 /dev/sgX" disables the drive, and i need a reboot to get it back. However, the plugin loops the "drive not supported" messages see below. Nov 18 20:13:45 Tower kernel: mdcmd (931): spindown 3 Nov 18 20:13:45 Tower kernel: mdcmd (932): spindown 4 Nov 18 20:13:45 Tower SAS Assist v0.7: disk 3 (/dev/sdh) not supported by SAS spindown plugin (excluded), not spun down Nov 18 20:13:46 Tower SAS Assist v0.7: disk 4 (/dev/sdi) not supported by SAS spindown plugin (excluded), not spun down Nov 18 20:13:46 Tower kernel: mdcmd (933): spindown 3 Nov 18 20:13:46 Tower kernel: mdcmd (934): spindown 4 Nov 18 20:13:47 Tower SAS Assist v0.7: disk 3 (/dev/sdh) not supported by SAS spindown plugin (excluded), not spun down Nov 18 20:13:47 Tower SAS Assist v0.7: disk 4 (/dev/sdi) not supported by SAS spindown plugin (excluded), not spun down Nov 18 20:13:47 Tower kernel: mdcmd (935): spindown 3 Nov 18 20:13:48 Tower kernel: mdcmd (936): spindown 4 Nov 18 20:13:48 Tower SAS Assist v0.7: disk 3 (/dev/sdh) not supported by SAS spindown plugin (excluded), not spun down Nov 18 20:13:48 Tower SAS Assist v0.7: disk 4 (/dev/sdi) not supported by SAS spindown plugin (excluded), not spun down
  6. Thanks @NNate that was it! Good catch!
  7. Hi, I posted in the b30 thread, that i have a performance impact when going from 6.8.3 to 6.9 beta. I did not have that with the first beta (was it 22 or something?) Post from b30 thread. I have the same problem now with beta 35, have tried different settings on the VMs but cant seem to resolve it. Does anyone have a similar problem? Diagnostics zip attached. tower-diagnostics-20201116-2040.zip
  8. A new version has been posted, ls62 and is working (at least for me) So linuxserver/emby:latest Is the one to use if you want to get the latest updates. @jcato @mbc0
  9. With the latest update i cant scan new media. It finds the media, but can't read what kind it is. And it wont play on Shield / Web. Media previously scaned works. I went back to linuxserver/emby:4.5.2.0-ls59 And all is working like before again.
  10. I did some more testing, reformated my cache pool with 1MB alignment. After that i tested with beta30 and the 6.8.3 It's a WS2019 no other difference than. On b30 i have Virtio-net anf i440fx-5.2 on 6.8.3 virtio and i440fx 4.2 Disk mark 7259 17187 CPU Mark 933 3135 Memorymark 813 1879 Quite a difference! (forgot to pull diagnostics file while on b30)
  11. Hi, i modified the script, and got it working. It's VERY simplified. And cusom to my setup. I modified 99-spinsasdown.conf to trigger "spindown 3" For my SAS disk 3. :msg,contains,"spindown 3" ^/usr/local/bin/unraid-spinsasdown And in your script unraid-spinsasdown I removed all intelligence and just spin down the correct disk. sg_start --readonly --pc=3 /dev/sg7 Unraid log after spin up / spin down in GUI Oct 21 10:27:41 Tower kernel: mdcmd (99): spinup 3 Oct 21 10:27:41 Tower kernel: md: do_drive_cmd: disk3: ATA_OP e3 ioctl error: -5 Oct 21 10:28:03 Tower kernel: mdcmd (100): spindown 3 Oct 21 10:28:03 Tower emhttpd: error: mdcmd, 2723: Input/output error (5): write Oct 21 10:28:03 Tower kernel: md: do_drive_cmd: disk3: ATA_OP e0 ioctl error: -5 And checking status root@Tower:/# sdparm --command=sense /dev/sdh /dev/sdh: HP MB8000JFECQ HPD7 Additional sense: Standby condition activated by command Edit: Tried again, waited 15 min and got this, and disk not spun down.. Oct 21 10:45:52 Tower kernel: mdcmd (103): spinup 3 Oct 21 10:45:52 Tower kernel: md: do_drive_cmd: disk3: ATA_OP e3 ioctl error: -5 Oct 21 11:00:53 Tower kernel: mdcmd (104): spindown 3 Oct 21 11:00:53 Tower emhttpd: error: mdcmd, 2723: Input/output error (5): write Oct 21 11:00:53 Tower kernel: md: do_drive_cmd: disk3: ATA_OP e0 ioctl error: -5 Oct 21 11:00:54 Tower kernel: mdcmd (105): spindown 3 Oct 21 11:01:10 Tower rsyslogd: program '/usr/local/bin/unraid-spinsasdown' (pid 23020) exited with status 2 [v8.1908.0] Edit 2 If i run /usr/local/bin/unraid-spinsasdown from CLI, the disk spins down root@Tower:/mnt/disk1/media/org.rel# sdparm --command=sense /dev/sdh /dev/sdh: HP MB8000JFECQ HPD7 root@Tower:/mnt/disk1/media/org.rel# /usr/local/bin/unraid-spinsasdown root@Tower:/mnt/disk1/media/org.rel# sdparm --command=sense /dev/sdh /dev/sdh: HP MB8000JFECQ HPD7 Additional sense: Standby condition activated by command
  12. Hi, I have tried to read up, but don't know if anyone got the same error as I have. I have a major performance decrease after beta 25 i think it was. Took like 10h to update a Windows 10 VM. And my Nextcloud docker is unbearable! I have looked at the excessive SSD write But i don't have any of that. Here's 6.8.3 Cache write. Normal operation. graph looks like this most of the time. Compared to 6.9b30, it's about the same. Different timeline though. Here is where i changed from beta30 to 6.8.3 unpacking a large file from windows server., first at 16MB/s. then at 55+MB/s My cache is a RAID10 BTRFS pool. I'm on 6.8.3 now, so a diagnostic may not help much. But i can go back to 6.9b30 to create one if it's of any help. Or is it the same known problem with btrfs? And i need to reformat the cache disks in pool to make it work? Thanks
  13. How can I check that? I have disabled my hddtemp script, but i do have "cache dirs" running, can disable that for now.
  14. Hi, the log is empty as soon as i restart rc.rsyslogd. I used that file for an hour or so yesterday, and didn't get a single line in the log. I also went back to 6.8.3 yesterday due to performance problems. But i have the same issue with the log. And no spindown of SAS. I also have the same problem with 3 WD RED SATA drives. On the same controller i have 4 drives that go to standby as they should. They are actually on different channels. So the 4 drives working are on one, and 4 not working are on another. Maybe something there?! I had the same problem years ago with my WD RED not spinning down, but if i remember correctly i solved it somehow. Dont remember how... I have sent the file.
  15. Got some read errors when the disk spun up again, no red X thank god! Nothing in the log unfortunately.