Show Posts

This section allows you to view all posts made by this member. Note that you can only see posts made in areas you currently have access to.


Messages - bloudraak

Pages: [1]
1
Makes sense. Since I’m running doing this in Jenkins, it seems wise to have a dedicated account and change the password frequently.

2
Thanks for all the help. Pointing out the path for mapper helped a ton.

It seems the admin user I created didn't have sufficient permissions. When I SSH'ed into BMC with the root account, I was able to shut down the device and then turn it back on.

3
It's there.  :)

I created a custom admin user account, and /usr/sbin/ is not in the path. That's why obmcutil couldn't find it. Once I ran as root, the command worked as expected.

4
The issue I'm getting is that the command always fails with mapper
Code: [Select]
obmcutil state
/usr/bin/obmcutil: line 160: mapper: command not found

5
Operating Systems and Porting / Mananging Talos II remotely via BMC.
« on: April 17, 2022, 10:07:17 am »
I'm trying to run BMC commands via SSH to check the power status of the host, turn it on, and turn it off.

I can successfully SSH into the host and run some commands, for example:

Code: [Select]
$ cat /etc/os-release
ID="openbmc-phosphor"
NAME="Phosphor OpenBMC (Phosphor OpenBMC Project Reference Distro)"
VERSION="2.7.0-dev"
VERSION_ID="2.7.0-dev-571-g67efd9872"
PRETTY_NAME="Phosphor OpenBMC (Phosphor OpenBMC Project Reference Distro) 2.7.0-dev"
BUILD_ID="2.7.0-dev"
OPENBMC_TARGET_MACHINE="talos"


I know the BMC versioning is probably out of date, but Raptor isn't making it easy to stay current (neither does Dell and other vendors).

Looking at the documentation, it seems obmcutil would be the right tool for the job, and it gives me some help:
Code: [Select]
obmcutil -h
But when I run one of the commands, I always get an error.

Code: [Select]
$ obmcutil osstate
/usr/bin/obmcutil: line 181: mapper: command not found

Are there any other commands I can run via SSH to check the status of power and power on a device when it's powered off?

For background, I'm using my Talos II for release engineering and running Ubuntu 20.04 and other operating systems under KVM (whichever I can get my hands on). I'm using Jenkins to automate tasks through SSH, especially patching and powering off devices. Power is expensive in California, with the most expensive time to run machines being the evening. Running homelab sets me back around $200+ extra a month. That's a lot of wasted beers -- or 4+ trips to the Zoo with my daughter. So I turn servers off when not in use. My Talos II consumes 85W+, so it doesn't run cheap. I need to figure out ways to power the host on when Jenkins needs it, such as building stuff for POWER and then gracefully power it down when not in use. I can shut down the operating systems via Jenkins but have no way to turn it back on.

I'm exploring how to automatically build and test BMC, which would involve Jenkins uploading and updating it. Since Jenkins (and other CI/CD tools I'm playing with) natively support SSH, I think it may be worthwhile to explore what else I can do when I SSH into the BMC.

6
I'm curious whether anyone has been able to run FreeBSD 12.2 PPC64 as a guest in KVM, running on Ubuntu 20.04 ppc64le (or anything relevant), and how they managed to get it to work. I'm hoping I looked at this problem too long and I'm missing something really simple.

I'm building out a lab of different operating systems and configurations in order to test some software. My Talos II is running Ubuntu 20.04 ppc64le, with KVM installed and configured. I'm already running a few ppc64le and ppc64 virtual machines on the host, including FreeBSD 13 (both ppc64le and ppc64). The issue at hand is specific to FreeBSD 12.02 ppc64. It has 256GB RAM and 24TB storage, which I think is sufficient to build a lab to test various PPC64 and PPC64LE configurations.

I downloaded the FreeBSD 12.2 ppc64 ISO and tried to create the virtual machine as follows:

Code: [Select]
virt-install \
    --name freebsd12-ppc64-01 \
    --memory 2048 \
    --virt-type=kvm \
    --disk pool=default,size=64,format=qcow2 \
    --vcpus=1 \
    --network network=br0 \
    --os-variant freebsd12.2 \
    --graphics none \
    --cdrom /var/lib/libvirt/boot/FreeBSD-12.2-RELEASE-powerpc-powerpc64-dvd1.iso \
    --debug \
    --force

which failed with

Code: [Select]
Hit [Enter] to boot immediately, or any other key for command prompt.
Booting [/boot/kernel/kernel]...               
Kernel entry at 0x102390 ...
panic: Standard large pages (SLB[L] = 1, PTE[LP] = 0) not supported by this system. Please enable huge page backing if running under PowerKVM.
cpuid = 0
time = 1
KDB: stack backtrace:
#0 0xc000000000729964 at ??+0
#1 0xc0000000006c1254 at ??+0
#2 0xc0000000006c1320 at ??+0
#3 0xc000000000b1a764 at ??+0
#4 0xc000000000b0b2ac at ??+0
#5 0xc000000000b04a0c at ??+0
#6 0xc000000000102440 at ??+0
Uptime: 1s

I configured huge pages on Ubuntu using  https://help.ubuntu.com/community/KVM%20-%20Using%20Hugepages. And the output of

Code: [Select]
cat /proc/meminfo | grep Huge
Looks something as follows:

Code: [Select]
AnonHugePages:   2904064 kB
ShmemHugePages:        0 kB
FileHugePages:         0 kB
HugePages_Total:   16384
HugePages_Free:    15360
HugePages_Rsvd:        0
HugePages_Surp:        0
Hugepagesize:       2048 kB
Hugetlb:        33554432 kB

I then tried to create the VM using the following after adding memorybacking argument.

Code: [Select]
virt-install \
    --name freebsd12-ppc64-01 \
    --memory 2048 \
    --virt-type=kvm \
    --disk pool=default,size=64,format=qcow2 \
    --vcpus=1 \
    --memorybacking hugepages=yes \
    --network network=br0 \
    --os-variant freebsd12.2 \
    --graphics none \
    --cdrom /var/lib/libvirt/boot/FreeBSD-12.2-RELEASE-powerpc-powerpc64-dvd1.iso \
    --debug \
    --force

It still failed with the same error.  I requested some help from the FreeBSD forums https://forums.freebsd.org/threads/installing-freebsd-12-2-ppc64-in-kvm-on-power9.83253/, on which there's an interesting comment:

Quote
freebsd kernel reads available page sizes from openfirmware device tree @ibm,segment-page-sizes
if the key is present but it does not report 16MB pages it bombs

The default huge page size is 2M, and no matter what I have tried (kernel args and whatnot), that remains the default. Hugeadm lists 2M and 1G page sizes.

Code: [Select]
$ hugeadm --page-sizes-all
2097152
1073741824

and if I try to create a 16M pool, it barfs:

Code: [Select]
$hugeadm --pool-pages-min 16M:64
hugeadm:ERROR: 16M: unknown page size

What am I missing?


7
Talos II / Re: Resetting Talos II to "Factory Settings"
« on: April 18, 2020, 03:48:16 pm »
I wasn't able to use it. Just getting "command not found". That said, I think the issue is something different. The error message I ended up getting never displays on the monitor, only the serial cable.

When attempting to install the OS, I get the following error after the reboot.

Code: [Select]
cpu 0x3b: Vector: 380 (Data Access Out of Range) at [c000201fdd29b620]
    pc: c0000000001d05f0: __free_pages+0x10/0x50
    lr: c000000000123c24: dma_direct_free_pages+0x54/0x90
    sp: c000201fdd29b8b0
   msr: 900000000280b033
   dar: c04240000000dcb4
  current = 0xc000201fdd244200
  paca    = 0xc000201fff704f00   irqmask: 0x03   irq_happened: 0x01
    pid   = 1002, comm = init
Linux version 5.5.0-openpower1 (root@raptor-build-public-staging-01) (gcc version 6.5.0 (Buildroot 2019.05.3-06769-g7bdd570165)) #2 SMP Thu Feb 20 02:19:47 UTC 2020
enter ? for help
[c000201fdd29b8b0] c000000000123c24 dma_direct_free_pages+0x54/0x90 (unreliable)
[c000201fdd29b8d0] c000000000038728 dma_iommu_free_coherent+0x98/0xc0
[c000201fdd29b920] c000000000123020 dma_free_attrs+0x100/0x110
[c000201fdd29b970] c0000000001d9bf4 dma_pool_destroy+0x174/0x200
[c000201fdd29ba10] c00800000e1417e8 _base_release_memory_pools+0x1e0/0x498 [mpt3sas]
[c000201fdd29baa0] c00800000e14b428 mpt3sas_base_detach+0x40/0x160 [mpt3sas]
[c000201fdd29bb10] c00800000e15bb5c scsih_shutdown+0xc4/0x110 [mpt3sas]
[c000201fdd29bb70] c0000000003cda10 pci_device_shutdown+0x50/0xc0
[c000201fdd29bba0] c00000000064a908 device_shutdown+0x1f8/0x330
[c000201fdd29bc40] c0000000000cfe2c kernel_restart_prepare+0x4c/0x60
[c000201fdd29bc60] c0000000000cff50 kernel_restart+0x20/0xc0
[c000201fdd29bcd0] c0000000000d0370 __do_sys_reboot+0x1b0/0x2c0
[c000201fdd29be20] c00000000000b50c system_call+0x5c/0x68
--- Exception: c01 (System Call) at 00007fff89bd10a4
SP (7ffff94afef0) is in userspace

When restarting the server

Code: [Select]
[520051.754983] configfs-gadget gadget: high-speed config #1: c
[520053.170880] occ-hwmon occ-hwmon.1: OCC found, code level: op_occ_191023a
[520053.414668] occ-hwmon occ-hwmon.2: OCC found, code level: op_occ_191023a
[Disconnected]
[Connected]
 

 

--== Welcome to Hostboot hostboot-a2ddbf3/hbicore.bin ==--
 

  3.10730|secure|SecureROM valid - enabling functionality
  5.51749|Booting from SBE side 0 on master proc=00050000
  5.55317|ISTEP  6. 5 - host_init_fsi
  5.73355|ISTEP  6. 6 - host_set_ipl_parms
  6.03033|ISTEP  6. 7 - host_discover_targets
  7.79252|HWAS|PRESENT> DIMM[03]=A0A0A0A000000000
  7.79254|HWAS|PRESENT> Proc[05]=8800000000000000
  7.79255|HWAS|PRESENT> Core[07]=5645406555000000
  7.91166|ISTEP  6. 8 - host_update_master_tpm
  8.41908|SECURE|Security Access Bit> 0x0000000000000000
  8.41909|SECURE|Secure Mode Disable (via Jumper)> 0xC000000000000000
  8.41928|ISTEP  6. 9 - host_gard
 8.91332|HWAS|Applying GARD record for HUID=0x00030000 (Physical:/Sys0/Node0/DIMM0) due to 0x90000010
  8.91664|HWAS|Applying GARD record for HUID=0x00030000 (Physical:/Sys0/Node0/DIMM0) due to 0x90000010
  8.91676|HWAS|Deconfig HUID 0x00030000, Physical:/Sys0/Node0/DIMM0
  8.91704|HWAS|FUNCTIONAL> DIMM[03]=20A0A0A000000000
  8.91705|HWAS|FUNCTIONAL> Proc[05]=8800000000000000
  8.91707|HWAS|FUNCTIONAL> Core[07]=5645406555000000
  8.92233|ISTEP  6.11 - host_start_occ_xstop_handler
10.14337|ISTEP  6.12 - host_voltage_config
10.27450|ISTEP  7. 1 - mss_attr_cleanup
10.95124|ISTEP  7. 2 - mss_volt
11.19280|ISTEP  7. 3 - mss_freq
11.56893|ISTEP  7. 4 - mss_eff_config
12.21124|ISTEP  7. 5 - mss_attr_update
12.23273|ISTEP  8. 1 - host_slave_sbe_config
12.44698|ISTEP  8. 2 - host_setup_sbe
12.44873|ISTEP  8. 3 - host_cbs_start
12.49990|ISTEP  8. 4 - proc_check_slave_sbe_seeprom_complete
17.55574|ISTEP  8. 5 - host_attnlisten_proc
17.59473|ISTEP  8. 6 - host_p9_fbc_eff_config
17.60271|ISTEP  8. 7 - host_p9_eff_config_links
17.64173|ISTEP  8. 8 - proc_attr_update
17.64332|ISTEP  8. 9 - proc_chiplet_fabric_scominit
17.68790|ISTEP  8.10 - proc_xbus_scominit
18.97508|ISTEP  8.11 - proc_xbus_enable_ridi
18.99465|ISTEP  8.12 - host_set_voltages
19.08378|ISTEP  9. 1 - fabric_erepair
19.25325|ISTEP  9. 2 - fabric_io_dccal
20.01976|ISTEP  9. 3 - fabric_pre_trainadv
20.04370|ISTEP  9. 4 - fabric_io_run_training
20.20031|ISTEP  9. 5 - fabric_post_trainadv
20.20406|ISTEP  9. 6 - proc_smp_link_layer
20.20861|ISTEP  9. 7 - proc_fab_iovalid
20.49515|ISTEP  9. 8 - host_fbc_eff_config_aggregate
20.52556|ISTEP 10. 1 - proc_build_smp
21.60404|ISTEP 10. 2 - host_slave_sbe_update
22.65219|ISTEP 10. 4 - proc_cen_ref_clk_enable
22.70371|ISTEP 10. 5 - proc_enable_osclite
22.70484|ISTEP 10. 6 - proc_chiplet_scominit
22.78700|ISTEP 10. 7 - proc_abus_scominit
22.81221|ISTEP 10. 8 - proc_obus_scominit
22.81445|ISTEP 10. 9 - proc_npu_scominit
22.83694|ISTEP 10.10 - proc_pcie_scominit
22.92795|ISTEP 10.11 - proc_scomoverride_chiplets
22.93975|ISTEP 10.12 - proc_chiplet_enable_ridi
22.97791|ISTEP 10.13 - host_rng_bist
23.03426|ISTEP 10.14 - host_update_redundant_tpm
23.03943|ISTEP 11. 1 - host_prd_hwreconfig
23.30043|ISTEP 11. 2 - cen_tp_chiplet_init1
23.30748|ISTEP 11. 3 - cen_pll_initf
23.31247|ISTEP 11. 4 - cen_pll_setup
23.31704|ISTEP 11. 5 - cen_tp_chiplet_init2
23.34176|ISTEP 11. 6 - cen_tp_arrayinit
23.34678|ISTEP 11. 7 - cen_tp_chiplet_init3
23.35156|ISTEP 11. 8 - cen_chiplet_init
23.35625|ISTEP 11. 9 - cen_arrayinit
23.36093|ISTEP 11.10 - cen_initf
23.36704|ISTEP 11.11 - cen_do_manual_inits
23.37181|ISTEP 11.12 - cen_startclocks
23.37685|ISTEP 11.13 - cen_scominits
23.38151|ISTEP 12. 1 - mss_getecid
24.21273|ISTEP 12. 2 - dmi_attr_update
24.23695|ISTEP 12. 3 - proc_dmi_scominit
24.29667|ISTEP 12. 4 - cen_dmi_scominit
24.30138|ISTEP 12. 5 - dmi_erepair
24.36646|ISTEP 12. 6 - dmi_io_dccal
24.37048|ISTEP 12. 7 - dmi_pre_trainadv
24.37525|ISTEP 12. 8 - dmi_io_run_training
24.39397|ISTEP 12. 9 - dmi_post_trainadv
24.39866|ISTEP 12.10 - proc_cen_framelock
24.40360|ISTEP 12.11 - host_startprd_dmi
24.40765|ISTEP 12.12 - host_attnlisten_memb
24.41171|ISTEP 12.13 - cen_set_inband_addr
24.41584|ISTEP 13. 1 - host_disable_memvolt
24.58138|ISTEP 13. 2 - mem_pll_reset
24.62217|ISTEP 13. 3 - mem_pll_initf
24.66147|ISTEP 13. 4 - mem_pll_setup
24.69804|ISTEP 13. 6 - mem_startclocks
24.71370|ISTEP 13. 7 - host_enable_memvolt
24.73321|ISTEP 13. 8 - mss_scominit
25.30592|ISTEP 13. 9 - mss_ddr_phy_reset
25.40047|ISTEP 13.10 - mss_draminit
25.98234|ISTEP 13.11 - mss_draminit_training
28.04184|ISTEP 13.12 - mss_draminit_trainadv
28.28842|ISTEP 13.13 - mss_draminit_mc
28.32749|ISTEP 14. 1 - mss_memdiag
33.56260|ISTEP 14. 2 - mss_thermal_init
33.62293|ISTEP 14. 3 - proc_pcie_config
33.67115|ISTEP 14. 4 - mss_power_cleanup
33.67716|ISTEP 14. 5 - proc_setup_bars
33.71607|ISTEP 14. 6 - proc_htm_setup
33.72956|ISTEP 14. 7 - proc_exit_cache_contained
33.76993|ISTEP 15. 1 - host_build_stop_image
37.33691|ISTEP 15. 2 - proc_set_pba_homer_bar
37.40133|ISTEP 15. 3 - host_establish_ex_chiplet
37.43321|ISTEP 15. 4 - host_start_stop_engine
37.46371|ISTEP 16. 1 - host_activate_master
38.70238|ISTEP 16. 2 - host_activate_slave_cores
38.89911|ISTEP 16. 3 - host_secure_rng
38.88787|ISTEP 16. 4 - mss_scrub
38.90828|ISTEP 16. 5 - host_load_io_ppe
38.94111|ISTEP 16. 6 - host_ipl_complete
39.32079|ISTEP 18.11 - proc_tod_setup
39.44248|ISTEP 18.12 - proc_tod_init
39.43995|ISTEP 20. 1 - host_load_payload
40.13808|ISTEP 20. 2 - host_load_hdat
41.63064|ISTEP 21. 1 - host_runtime_setup
53.10080|htmgt|OCCs are now running in ACTIVE state
58.33571|ISTEP 21. 2 - host_verify_hdat
58.37178|ISTEP 21. 3 - host_start_payload
[   59.225278060,5] OPAL skiboot-9858186 starting...
[   59.225281070,7] initial console log level: memory 7, driver 5
[   59.225283107,6] CPU: P9 generation processor (max 4 threads/core)
[   59.225284921,7] CPU: Boot CPU PIR is 0x0834 PVR is 0x004e1203
[   59.225287534,7] OPAL table: 0x30103830 .. 0x30103e10, branch table: 0x30002000
[   59.225290544,7] Assigning physical memory map table for nimbus
[   59.225293297,7] Parsing HDAT...
[   59.225294624,7] SPIRA-S found.
[   59.225296926,6] BMC #0: HW version 3, SW version 2, chip DD1.0
[   59.225457609,6] SP Family is ibm,ast2500,openbmc
[   59.225463830,7] LPC: IOPATH chip id = 0
[   59.225465150,7] LPC: FW BAR       = f0000000
[   59.225466678,7] LPC: MEM BAR      = e0000000
[   59.225468142,7] LPC: IO BAR       = d0010000
[   59.225469592,7] LPC: Internal BAR = c0012000
[   59.225482159,7] LPC UART: base addr = 3f8 (3f8) size = 1 clk = 1843200, baud = 115200
[   59.225484833,7] LPC: BT [0, 0] sms_int: 0, bmc_int: 0
[   59.227438048,5] HDAT I2C: found e3p1 - unknown@1c dp:ff (ff:)
[   59.227553256,5] HDAT I2C: found e3p1 - unknown@1d dp:ff (ff:)
[   59.227606613,5] HDAT I2C: found e3p0 - unknown@19 dp:ff (ff:)
[   59.227659260,5] HDAT I2C: found e3p1 - unknown@1e dp:ff (ff:)
[   59.227704460,5] HDAT I2C: found e3p0 - unknown@1b dp:ff (ff:)
[   59.227754386,5] HDAT I2C: found e3p1 - unknown@1f dp:ff (ff:)
[   59.227819475,5] HDAT I2C: found e3p0 - unknown@1a dp:ff (ff:)
[   59.227898145,5] HDAT I2C: found e3p0 - unknown@18 dp:ff (ff:)
[   59.228269121,5] HDAT I2C: found e3p1 - unknown@1c dp:ff (ff:)
[   59.228347500,5] HDAT I2C: found e3p1 - unknown@1d dp:ff (ff:)
[   59.228398443,5] HDAT I2C: found e3p0 - unknown@19 dp:ff (ff:)
Petitboot (0ed84c0-p94177c1)                         T2P9D01 REV 1.00 A1000645
──────────────────────────────────────────────────────────────────────────────
 

  System information
  System configuration
  System status log
  Language
  Rescan devices
  Retrieve config from URL
  Plugins (0)
*Exit to shell         
 

 

 

 

 

 

 

 

 

 

──────────────────────────────────────────────────────────────────────────────
Enter=accept, e=edit, n=new, x=exit, l=language, g=log, h=help
[enP4p1s0f1] Probing from base tftp://192.168.0.1/pxelinux.cfg/

At this point I wait for the Debian installer (in the USB at the back of the device) to show up, and then select "Expert Installation". After that it prints that a SIGTERM is received, and it spits out the following.

Code: [Select]
cpu 0x3b: Vector: 380 (Data Access Out of Range) at [c000201fdd29b620]
    pc: c0000000001d05f0: __free_pages+0x10/0x50
    lr: c000000000123c24: dma_direct_free_pages+0x54/0x90
    sp: c000201fdd29b8b0
   msr: 900000000280b033
   dar: c04240000000dcb4
  current = 0xc000201fdd244200
  paca    = 0xc000201fff704f00   irqmask: 0x03   irq_happened: 0x01
    pid   = 1002, comm = init
Linux version 5.5.0-openpower1 (root@raptor-build-public-staging-01) (gcc version 6.5.0 (Buildroot 2019.05.3-06769-g7bdd570165)) #2 SMP Thu Feb 20 02:19:47 UTC 2020
enter ? for help
[c000201fdd29b8b0] c000000000123c24 dma_direct_free_pages+0x54/0x90 (unreliable)
[c000201fdd29b8d0] c000000000038728 dma_iommu_free_coherent+0x98/0xc0
[c000201fdd29b920] c000000000123020 dma_free_attrs+0x100/0x110
[c000201fdd29b970] c0000000001d9bf4 dma_pool_destroy+0x174/0x200
[c000201fdd29ba10] c00800000e1417e8 _base_release_memory_pools+0x1e0/0x498 [mpt3sas]
[c000201fdd29baa0] c00800000e14b428 mpt3sas_base_detach+0x40/0x160 [mpt3sas]
[c000201fdd29bb10] c00800000e15bb5c scsih_shutdown+0xc4/0x110 [mpt3sas]
[c000201fdd29bb70] c0000000003cda10 pci_device_shutdown+0x50/0xc0
[c000201fdd29bba0] c00000000064a908 device_shutdown+0x1f8/0x330
[c000201fdd29bc40] c0000000000cfe2c kernel_restart_prepare+0x4c/0x60
[c000201fdd29bc60] c0000000000cff50 kernel_restart+0x20/0xc0
[c000201fdd29bcd0] c0000000000d0370 __do_sys_reboot+0x1b0/0x2c0
[c000201fdd29be20] c00000000000b50c system_call+0x5c/0x68
--- Exception: c01 (System Call) at 00007fff89bd10a4
SP (7ffff94afef0) is in userspace

What jumps out is the following:

Code: [Select]
  8.91332|HWAS|Applying GARD record for HUID=0x00030000 (Physical:/Sys0/Node0/DIMM0) due to 0x90000010
  8.91664|HWAS|Applying GARD record for HUID=0x00030000 (Physical:/Sys0/Node0/DIMM0) due to 0x90000010

What is this about?

The box consists of:

  • Talos™ II Mainboard (Board Only)
  • 2U Heatsink Assembly for POWER9 CPUs
  • LSI 9300-8i 8-port Internal SAS 3.0 HBA
  • M393A4K40BB2-CTD - Samsung 1x 32GB DDR4-2666 RDIMM PC4-21300V-R Dual Rank x4 Module (x8 = 256GB Memory)
  • Supermicro CSE-836BE1C-R1K03B Server Chassis 3U Rackmount
  • 3TB SAS3 drives (x16)

From the boot sequence, I can see it detects the SAS3 hard disks as well as the memory.

Any suggestions would be appreciated. I can't find much on the internet about the error and why it may occur. And I can't find any documentation of commands to run when the initial IPL is complete but before attempting to install the operating system. Is there any? Where can I learn more about

I did create a ticket (#367172) with support, but given that they usually take weeks to respond so I don't have much faith that I'll have this resolved anytime soon through that channel. 

8
Talos II / Re: Resetting Talos II to "Factory Settings"
« on: April 14, 2020, 08:56:58 pm »
I haven't installed the OS on it yet. :(

Just got the serial cables via the mail and see what it produces.

9
Talos II / Resetting Talos II to "Factory Settings"
« on: April 05, 2020, 02:20:08 pm »
Is there an easy way to reset an Talos II back to factory settings so I can go through the setup process again?

I changed the passwords per documentation but can't SSH to OpenBMC, nor use the website to login.

10
Talos II / Building a Talos II computer
« on: March 24, 2020, 02:12:31 pm »
Hello,

I have the following equipment

I'm trying to connect the various components, get Debian installed so I can start virtualize some workloads (mostly for development and testing). There seems to be a disconnect between the Supermicro chassis documentation and that of the motherboard, especially how cables are connected from chassis.

A couple of questions:
  • The backplane SAS-836EL1 comes with an 16 port extender. If I connect it to the HBA, would it be sufficient for the 16 drives?
  • Which cable is used to connect the HBA to the backplane? From the backplane documentation, I'm gathering a SFF-8484 is required but the HBA specification states it has 2 Mini-SAS HD SFF8643 internal connectors. I'm a bit lost.
  • I'm trying to figure out which backplane cables need to go where. See attached picture. I'm not seeing any codes on them and thus can't figure out which cable they are. I have not removed the backplane from the chassis.
  • I'm trying to figure out which power supply cable should be connected where on the motherboard (to be done last).
Are there any videos of folks doing a build or going through the process step by step? I really which Raptor Computing Systems can be a bit more proactive, like doing videos similar to the Dell videos, in particular with Supermicro chassis since they claim some compatibility.


Thanks,
Werner

11
Talos II / Re: Rackmount Server Chassis for Talos II
« on: February 14, 2020, 08:37:52 am »
My understanding is that the board and 2U heatsink will fit into a 2U chassis and if I had fans in the chassis to force airflow over the board and CPU, any excess heat will be disbursed and later be extracted by the rack fans.

While there is no aircon where the rack is located, the rack itself (NetShelter CX 24U) has extraction fans. The servers will also be 1U apart which should allow for more airflow and heat disbursement. 

That said, would it be better to invest in a 3U or 4U chassis? It does seem like waisted space, which could be used for a shelf (I have a ton of single board computers I need to fit in too).

12
Talos II / Rackmount Server Chassis for Talos II
« on: February 12, 2020, 09:53:48 am »
Hello,

I just purchased the Talos II board with 2xIBM POWER9 v2 CPU (8-Core) and 2U Heatsink Assemblies.  Looking for a rack mountable chassis. I had a look at https://wiki.raptorcs.com/wiki/Talos_II/Hardware_Compatibility_List, are there any other Rackmount Server Chassis folks used for Talos II?

I have been looking at variants of the Superchassis CSE-825TQ, such as:

  • SuperChassis 825TQ-R740LPB
  • SuperChassis 825TQ-600LPB

Has anyone tried these before? What are the power supply would you recommend for 2xIBM POWER9 v2 CPU (8-Core), with 8 SAS drives and 256GB memory? I only have an 15A/120V power outlet available.

Thanks,
Werner

Pages: [1]