logo

oasis-root

Compiled tree of Oasis Linux based on own branch at <https://hacktivis.me/git/oasis/> git clone https://anongit.hacktivis.me/git/oasis-root.git

zpool.8 (20137B)


  1. .\"
  2. .\" CDDL HEADER START
  3. .\"
  4. .\" The contents of this file are subject to the terms of the
  5. .\" Common Development and Distribution License (the "License").
  6. .\" You may not use this file except in compliance with the License.
  7. .\"
  8. .\" You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
  9. .\" or https://opensource.org/licenses/CDDL-1.0.
  10. .\" See the License for the specific language governing permissions
  11. .\" and limitations under the License.
  12. .\"
  13. .\" When distributing Covered Code, include this CDDL HEADER in each
  14. .\" file and include the License file at usr/src/OPENSOLARIS.LICENSE.
  15. .\" If applicable, add the following below this CDDL HEADER, with the
  16. .\" fields enclosed by brackets "[]" replaced with your own identifying
  17. .\" information: Portions Copyright [yyyy] [name of copyright owner]
  18. .\"
  19. .\" CDDL HEADER END
  20. .\"
  21. .\" Copyright (c) 2007, Sun Microsystems, Inc. All Rights Reserved.
  22. .\" Copyright (c) 2012, 2018 by Delphix. All rights reserved.
  23. .\" Copyright (c) 2012 Cyril Plisko. All Rights Reserved.
  24. .\" Copyright (c) 2017 Datto Inc.
  25. .\" Copyright (c) 2018 George Melikov. All Rights Reserved.
  26. .\" Copyright 2017 Nexenta Systems, Inc.
  27. .\" Copyright (c) 2017 Open-E, Inc. All Rights Reserved.
  28. .\"
  29. .Dd February 14, 2024
  30. .Dt ZPOOL 8
  31. .Os
  32. .
  33. .Sh NAME
  34. .Nm zpool
  35. .Nd configure ZFS storage pools
  36. .Sh SYNOPSIS
  37. .Nm
  38. .Fl ?V
  39. .Nm
  40. .Cm version
  41. .Op Fl j
  42. .Nm
  43. .Cm subcommand
  44. .Op Ar arguments
  45. .
  46. .Sh DESCRIPTION
  47. The
  48. .Nm
  49. command configures ZFS storage pools.
  50. A storage pool is a collection of devices that provides physical storage and
  51. data replication for ZFS datasets.
  52. All datasets within a storage pool share the same space.
  53. See
  54. .Xr zfs 8
  55. for information on managing datasets.
  56. .Pp
  57. For an overview of creating and managing ZFS storage pools see the
  58. .Xr zpoolconcepts 7
  59. manual page.
  60. .
  61. .Sh SUBCOMMANDS
  62. All subcommands that modify state are logged persistently to the pool in their
  63. original form.
  64. .Pp
  65. The
  66. .Nm
  67. command provides subcommands to create and destroy storage pools, add capacity
  68. to storage pools, and provide information about the storage pools.
  69. The following subcommands are supported:
  70. .Bl -tag -width Ds
  71. .It Xo
  72. .Nm
  73. .Fl ?\&
  74. .Xc
  75. Displays a help message.
  76. .It Xo
  77. .Nm
  78. .Fl V , -version
  79. .Xc
  80. .It Xo
  81. .Nm
  82. .Cm version
  83. .Op Fl j
  84. .Xc
  85. Displays the software version of the
  86. .Nm
  87. userland utility and the ZFS kernel module.
  88. Use
  89. .Fl j
  90. option to output in JSON format.
  91. .El
  92. .
  93. .Ss Creation
  94. .Bl -tag -width Ds
  95. .It Xr zpool-create 8
  96. Creates a new storage pool containing the virtual devices specified on the
  97. command line.
  98. .It Xr zpool-initialize 8
  99. Begins initializing by writing to all unallocated regions on the specified
  100. devices, or all eligible devices in the pool if no individual devices are
  101. specified.
  102. .El
  103. .
  104. .Ss Destruction
  105. .Bl -tag -width Ds
  106. .It Xr zpool-destroy 8
  107. Destroys the given pool, freeing up any devices for other use.
  108. .It Xr zpool-labelclear 8
  109. Removes ZFS label information from the specified
  110. .Ar device .
  111. .El
  112. .
  113. .Ss Virtual Devices
  114. .Bl -tag -width Ds
  115. .It Xo
  116. .Xr zpool-attach 8 Ns / Ns Xr zpool-detach 8
  117. .Xc
  118. Converts a non-redundant disk into a mirror, or increases
  119. the redundancy level of an existing mirror
  120. .Cm ( attach Ns ), or performs the inverse operation (
  121. .Cm detach Ns ).
  122. .It Xo
  123. .Xr zpool-add 8 Ns / Ns Xr zpool-remove 8
  124. .Xc
  125. Adds the specified virtual devices to the given pool,
  126. or removes the specified device from the pool.
  127. .It Xr zpool-replace 8
  128. Replaces an existing device (which may be faulted) with a new one.
  129. .It Xr zpool-split 8
  130. Creates a new pool by splitting all mirrors in an existing pool (which decreases
  131. its redundancy).
  132. .El
  133. .
  134. .Ss Properties
  135. Available pool properties listed in the
  136. .Xr zpoolprops 7
  137. manual page.
  138. .Bl -tag -width Ds
  139. .It Xr zpool-list 8
  140. Lists the given pools along with a health status and space usage.
  141. .It Xo
  142. .Xr zpool-get 8 Ns / Ns Xr zpool-set 8
  143. .Xc
  144. Retrieves the given list of properties
  145. .Po
  146. or all properties if
  147. .Sy all
  148. is used
  149. .Pc
  150. for the specified storage pool(s).
  151. .El
  152. .
  153. .Ss Monitoring
  154. .Bl -tag -width Ds
  155. .It Xr zpool-status 8
  156. Displays the detailed health status for the given pools.
  157. .It Xr zpool-iostat 8
  158. Displays logical I/O statistics for the given pools/vdevs.
  159. Physical I/O operations may be observed via
  160. .Xr iostat 1 .
  161. .It Xr zpool-events 8
  162. Lists all recent events generated by the ZFS kernel modules.
  163. These events are consumed by the
  164. .Xr zed 8
  165. and used to automate administrative tasks such as replacing a failed device
  166. with a hot spare.
  167. That manual page also describes the subclasses and event payloads
  168. that can be generated.
  169. .It Xr zpool-history 8
  170. Displays the command history of the specified pool(s) or all pools if no pool is
  171. specified.
  172. .El
  173. .
  174. .Ss Maintenance
  175. .Bl -tag -width Ds
  176. .It Xr zpool-prefetch 8
  177. Prefetches specific types of pool data.
  178. .It Xr zpool-scrub 8
  179. Begins a scrub or resumes a paused scrub.
  180. .It Xr zpool-checkpoint 8
  181. Checkpoints the current state of
  182. .Ar pool ,
  183. which can be later restored by
  184. .Nm zpool Cm import Fl -rewind-to-checkpoint .
  185. .It Xr zpool-trim 8
  186. Initiates an immediate on-demand TRIM operation for all of the free space in a
  187. pool.
  188. This operation informs the underlying storage devices of all blocks
  189. in the pool which are no longer allocated and allows thinly provisioned
  190. devices to reclaim the space.
  191. .It Xr zpool-sync 8
  192. This command forces all in-core dirty data to be written to the primary
  193. pool storage and not the ZIL.
  194. It will also update administrative information including quota reporting.
  195. Without arguments,
  196. .Nm zpool Cm sync
  197. will sync all pools on the system.
  198. Otherwise, it will sync only the specified pool(s).
  199. .It Xr zpool-upgrade 8
  200. Manage the on-disk format version of storage pools.
  201. .It Xr zpool-wait 8
  202. Waits until all background activity of the given types has ceased in the given
  203. pool.
  204. .El
  205. .
  206. .Ss Fault Resolution
  207. .Bl -tag -width Ds
  208. .It Xo
  209. .Xr zpool-offline 8 Ns / Ns Xr zpool-online 8
  210. .Xc
  211. Takes the specified physical device offline or brings it online.
  212. .It Xr zpool-resilver 8
  213. Starts a resilver.
  214. If an existing resilver is already running it will be restarted from the
  215. beginning.
  216. .It Xr zpool-reopen 8
  217. Reopen all the vdevs associated with the pool.
  218. .It Xr zpool-clear 8
  219. Clears device errors in a pool.
  220. .El
  221. .
  222. .Ss Import & Export
  223. .Bl -tag -width Ds
  224. .It Xr zpool-import 8
  225. Make disks containing ZFS storage pools available for use on the system.
  226. .It Xr zpool-export 8
  227. Exports the given pools from the system.
  228. .It Xr zpool-reguid 8
  229. Generates a new unique identifier for the pool.
  230. .El
  231. .
  232. .Sh EXIT STATUS
  233. The following exit values are returned:
  234. .Bl -tag -compact -offset 4n -width "a"
  235. .It Sy 0
  236. Successful completion.
  237. .It Sy 1
  238. An error occurred.
  239. .It Sy 2
  240. Invalid command line options were specified.
  241. .El
  242. .
  243. .Sh EXAMPLES
  244. .\" Examples 1, 2, 3, 4, 12, 13 are shared with zpool-create.8.
  245. .\" Examples 6, 14 are shared with zpool-add.8.
  246. .\" Examples 7, 16 are shared with zpool-list.8.
  247. .\" Examples 8 are shared with zpool-destroy.8.
  248. .\" Examples 9 are shared with zpool-export.8.
  249. .\" Examples 10 are shared with zpool-import.8.
  250. .\" Examples 11 are shared with zpool-upgrade.8.
  251. .\" Examples 15 are shared with zpool-remove.8.
  252. .\" Examples 17 are shared with zpool-status.8.
  253. .\" Examples 14, 17 are also shared with zpool-iostat.8.
  254. .\" Make sure to update them omnidirectionally
  255. .Ss Example 1 : No Creating a RAID-Z Storage Pool
  256. The following command creates a pool with a single raidz root vdev that
  257. consists of six disks:
  258. .Dl # Nm zpool Cm create Ar tank Sy raidz Pa sda sdb sdc sdd sde sdf
  259. .
  260. .Ss Example 2 : No Creating a Mirrored Storage Pool
  261. The following command creates a pool with two mirrors, where each mirror
  262. contains two disks:
  263. .Dl # Nm zpool Cm create Ar tank Sy mirror Pa sda sdb Sy mirror Pa sdc sdd
  264. .
  265. .Ss Example 3 : No Creating a ZFS Storage Pool by Using Partitions
  266. The following command creates a non-redundant pool using two disk partitions:
  267. .Dl # Nm zpool Cm create Ar tank Pa sda1 sdb2
  268. .
  269. .Ss Example 4 : No Creating a ZFS Storage Pool by Using Files
  270. The following command creates a non-redundant pool using files.
  271. While not recommended, a pool based on files can be useful for experimental
  272. purposes.
  273. .Dl # Nm zpool Cm create Ar tank Pa /path/to/file/a /path/to/file/b
  274. .
  275. .Ss Example 5 : No Making a non-mirrored ZFS Storage Pool mirrored
  276. The following command converts an existing single device
  277. .Ar sda
  278. into a mirror by attaching a second device to it,
  279. .Ar sdb .
  280. .Dl # Nm zpool Cm attach Ar tank Pa sda sdb
  281. .
  282. .Ss Example 6 : No Adding a Mirror to a ZFS Storage Pool
  283. The following command adds two mirrored disks to the pool
  284. .Ar tank ,
  285. assuming the pool is already made up of two-way mirrors.
  286. The additional space is immediately available to any datasets within the pool.
  287. .Dl # Nm zpool Cm add Ar tank Sy mirror Pa sda sdb
  288. .
  289. .Ss Example 7 : No Listing Available ZFS Storage Pools
  290. The following command lists all available pools on the system.
  291. In this case, the pool
  292. .Ar zion
  293. is faulted due to a missing device.
  294. The results from this command are similar to the following:
  295. .Bd -literal -compact -offset Ds
  296. .No # Nm zpool Cm list
  297. NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
  298. rpool 19.9G 8.43G 11.4G - 33% 42% 1.00x ONLINE -
  299. tank 61.5G 20.0G 41.5G - 48% 32% 1.00x ONLINE -
  300. zion - - - - - - - FAULTED -
  301. .Ed
  302. .
  303. .Ss Example 8 : No Destroying a ZFS Storage Pool
  304. The following command destroys the pool
  305. .Ar tank
  306. and any datasets contained within:
  307. .Dl # Nm zpool Cm destroy Fl f Ar tank
  308. .
  309. .Ss Example 9 : No Exporting a ZFS Storage Pool
  310. The following command exports the devices in pool
  311. .Ar tank
  312. so that they can be relocated or later imported:
  313. .Dl # Nm zpool Cm export Ar tank
  314. .
  315. .Ss Example 10 : No Importing a ZFS Storage Pool
  316. The following command displays available pools, and then imports the pool
  317. .Ar tank
  318. for use on the system.
  319. The results from this command are similar to the following:
  320. .Bd -literal -compact -offset Ds
  321. .No # Nm zpool Cm import
  322. pool: tank
  323. id: 15451357997522795478
  324. state: ONLINE
  325. action: The pool can be imported using its name or numeric identifier.
  326. config:
  327. tank ONLINE
  328. mirror ONLINE
  329. sda ONLINE
  330. sdb ONLINE
  331. .No # Nm zpool Cm import Ar tank
  332. .Ed
  333. .
  334. .Ss Example 11 : No Upgrading All ZFS Storage Pools to the Current Version
  335. The following command upgrades all ZFS Storage pools to the current version of
  336. the software:
  337. .Bd -literal -compact -offset Ds
  338. .No # Nm zpool Cm upgrade Fl a
  339. This system is currently running ZFS version 2.
  340. .Ed
  341. .
  342. .Ss Example 12 : No Managing Hot Spares
  343. The following command creates a new pool with an available hot spare:
  344. .Dl # Nm zpool Cm create Ar tank Sy mirror Pa sda sdb Sy spare Pa sdc
  345. .Pp
  346. If one of the disks were to fail, the pool would be reduced to the degraded
  347. state.
  348. The failed device can be replaced using the following command:
  349. .Dl # Nm zpool Cm replace Ar tank Pa sda sdd
  350. .Pp
  351. Once the data has been resilvered, the spare is automatically removed and is
  352. made available for use should another device fail.
  353. The hot spare can be permanently removed from the pool using the following
  354. command:
  355. .Dl # Nm zpool Cm remove Ar tank Pa sdc
  356. .
  357. .Ss Example 13 : No Creating a ZFS Pool with Mirrored Separate Intent Logs
  358. The following command creates a ZFS storage pool consisting of two, two-way
  359. mirrors and mirrored log devices:
  360. .Dl # Nm zpool Cm create Ar pool Sy mirror Pa sda sdb Sy mirror Pa sdc sdd Sy log mirror Pa sde sdf
  361. .
  362. .Ss Example 14 : No Adding Cache Devices to a ZFS Pool
  363. The following command adds two disks for use as cache devices to a ZFS storage
  364. pool:
  365. .Dl # Nm zpool Cm add Ar pool Sy cache Pa sdc sdd
  366. .Pp
  367. Once added, the cache devices gradually fill with content from main memory.
  368. Depending on the size of your cache devices, it could take over an hour for
  369. them to fill.
  370. Capacity and reads can be monitored using the
  371. .Cm iostat
  372. subcommand as follows:
  373. .Dl # Nm zpool Cm iostat Fl v Ar pool 5
  374. .
  375. .Ss Example 15 : No Removing a Mirrored top-level (Log or Data) Device
  376. The following commands remove the mirrored log device
  377. .Sy mirror-2
  378. and mirrored top-level data device
  379. .Sy mirror-1 .
  380. .Pp
  381. Given this configuration:
  382. .Bd -literal -compact -offset Ds
  383. pool: tank
  384. state: ONLINE
  385. scrub: none requested
  386. config:
  387. NAME STATE READ WRITE CKSUM
  388. tank ONLINE 0 0 0
  389. mirror-0 ONLINE 0 0 0
  390. sda ONLINE 0 0 0
  391. sdb ONLINE 0 0 0
  392. mirror-1 ONLINE 0 0 0
  393. sdc ONLINE 0 0 0
  394. sdd ONLINE 0 0 0
  395. logs
  396. mirror-2 ONLINE 0 0 0
  397. sde ONLINE 0 0 0
  398. sdf ONLINE 0 0 0
  399. .Ed
  400. .Pp
  401. The command to remove the mirrored log
  402. .Ar mirror-2 No is :
  403. .Dl # Nm zpool Cm remove Ar tank mirror-2
  404. .Pp
  405. At this point, the log device no longer exists
  406. (both sides of the mirror have been removed):
  407. .Bd -literal -compact -offset Ds
  408. pool: tank
  409. state: ONLINE
  410. scan: none requested
  411. config:
  412. NAME STATE READ WRITE CKSUM
  413. tank ONLINE 0 0 0
  414. mirror-0 ONLINE 0 0 0
  415. sda ONLINE 0 0 0
  416. sdb ONLINE 0 0 0
  417. mirror-1 ONLINE 0 0 0
  418. sdc ONLINE 0 0 0
  419. sdd ONLINE 0 0 0
  420. .Ed
  421. .Pp
  422. The command to remove the mirrored data
  423. .Ar mirror-1 No is :
  424. .Dl # Nm zpool Cm remove Ar tank mirror-1
  425. .Pp
  426. After
  427. .Ar mirror-1 No has been evacuated, the pool remains redundant, but
  428. the total amount of space is reduced:
  429. .Bd -literal -compact -offset Ds
  430. pool: tank
  431. state: ONLINE
  432. scan: none requested
  433. config:
  434. NAME STATE READ WRITE CKSUM
  435. tank ONLINE 0 0 0
  436. mirror-0 ONLINE 0 0 0
  437. sda ONLINE 0 0 0
  438. sdb ONLINE 0 0 0
  439. .Ed
  440. .
  441. .Ss Example 16 : No Displaying expanded space on a device
  442. The following command displays the detailed information for the pool
  443. .Ar data .
  444. This pool is comprised of a single raidz vdev where one of its devices
  445. increased its capacity by 10 GiB.
  446. In this example, the pool will not be able to utilize this extra capacity until
  447. all the devices under the raidz vdev have been expanded.
  448. .Bd -literal -compact -offset Ds
  449. .No # Nm zpool Cm list Fl v Ar data
  450. NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
  451. data 23.9G 14.6G 9.30G - 48% 61% 1.00x ONLINE -
  452. raidz1 23.9G 14.6G 9.30G - 48%
  453. sda - - - - -
  454. sdb - - - 10G -
  455. sdc - - - - -
  456. .Ed
  457. .
  458. .Ss Example 17 : No Adding output columns
  459. Additional columns can be added to the
  460. .Nm zpool Cm status No and Nm zpool Cm iostat No output with Fl c .
  461. .Bd -literal -compact -offset Ds
  462. .No # Nm zpool Cm status Fl c Pa vendor , Ns Pa model , Ns Pa size
  463. NAME STATE READ WRITE CKSUM vendor model size
  464. tank ONLINE 0 0 0
  465. mirror-0 ONLINE 0 0 0
  466. U1 ONLINE 0 0 0 SEAGATE ST8000NM0075 7.3T
  467. U10 ONLINE 0 0 0 SEAGATE ST8000NM0075 7.3T
  468. U11 ONLINE 0 0 0 SEAGATE ST8000NM0075 7.3T
  469. U12 ONLINE 0 0 0 SEAGATE ST8000NM0075 7.3T
  470. U13 ONLINE 0 0 0 SEAGATE ST8000NM0075 7.3T
  471. U14 ONLINE 0 0 0 SEAGATE ST8000NM0075 7.3T
  472. .No # Nm zpool Cm iostat Fl vc Pa size
  473. capacity operations bandwidth
  474. pool alloc free read write read write size
  475. ---------- ----- ----- ----- ----- ----- ----- ----
  476. rpool 14.6G 54.9G 4 55 250K 2.69M
  477. sda1 14.6G 54.9G 4 55 250K 2.69M 70G
  478. ---------- ----- ----- ----- ----- ----- ----- ----
  479. .Ed
  480. .
  481. .Sh ENVIRONMENT VARIABLES
  482. .Bl -tag -compact -width "ZPOOL_STATUS_NON_NATIVE_ASHIFT_IGNORE"
  483. .It Sy ZFS_ABORT
  484. Cause
  485. .Nm
  486. to dump core on exit for the purposes of running
  487. .Sy ::findleaks .
  488. .It Sy ZFS_COLOR
  489. Use ANSI color in
  490. .Nm zpool Cm status
  491. and
  492. .Nm zpool Cm iostat
  493. output.
  494. .It Sy ZPOOL_AUTO_POWER_ON_SLOT
  495. Automatically attempt to turn on the drives enclosure slot power to a drive when
  496. running the
  497. .Nm zpool Cm online
  498. or
  499. .Nm zpool Cm clear
  500. commands.
  501. This has the same effect as passing the
  502. .Fl -power
  503. option to those commands.
  504. .It Sy ZPOOL_POWER_ON_SLOT_TIMEOUT_MS
  505. The maximum time in milliseconds to wait for a slot power sysfs value
  506. to return the correct value after writing it.
  507. For example, after writing "on" to the sysfs enclosure slot power_control file,
  508. it can take some time for the enclosure to power down the slot and return
  509. "on" if you read back the 'power_control' value.
  510. Defaults to 30 seconds (30000ms) if not set.
  511. .It Sy ZPOOL_IMPORT_PATH
  512. The search path for devices or files to use with the pool.
  513. This is a colon-separated list of directories in which
  514. .Nm
  515. looks for device nodes and files.
  516. Similar to the
  517. .Fl d
  518. option in
  519. .Nm zpool import .
  520. .It Sy ZPOOL_IMPORT_UDEV_TIMEOUT_MS
  521. The maximum time in milliseconds that
  522. .Nm zpool import
  523. will wait for an expected device to be available.
  524. .It Sy ZPOOL_STATUS_NON_NATIVE_ASHIFT_IGNORE
  525. If set, suppress warning about non-native vdev ashift in
  526. .Nm zpool Cm status .
  527. The value is not used, only the presence or absence of the variable matters.
  528. .It Sy ZPOOL_VDEV_NAME_GUID
  529. Cause
  530. .Nm
  531. subcommands to output vdev guids by default.
  532. This behavior is identical to the
  533. .Nm zpool Cm status Fl g
  534. command line option.
  535. .It Sy ZPOOL_VDEV_NAME_FOLLOW_LINKS
  536. Cause
  537. .Nm
  538. subcommands to follow links for vdev names by default.
  539. This behavior is identical to the
  540. .Nm zpool Cm status Fl L
  541. command line option.
  542. .It Sy ZPOOL_VDEV_NAME_PATH
  543. Cause
  544. .Nm
  545. subcommands to output full vdev path names by default.
  546. This behavior is identical to the
  547. .Nm zpool Cm status Fl P
  548. command line option.
  549. .It Sy ZFS_VDEV_DEVID_OPT_OUT
  550. Older OpenZFS implementations had issues when attempting to display pool
  551. config vdev names if a
  552. .Sy devid
  553. NVP value is present in the pool's config.
  554. .Pp
  555. For example, a pool that originated on illumos platform would have a
  556. .Sy devid
  557. value in the config and
  558. .Nm zpool Cm status
  559. would fail when listing the config.
  560. This would also be true for future Linux-based pools.
  561. .Pp
  562. A pool can be stripped of any
  563. .Sy devid
  564. values on import or prevented from adding
  565. them on
  566. .Nm zpool Cm create
  567. or
  568. .Nm zpool Cm add
  569. by setting
  570. .Sy ZFS_VDEV_DEVID_OPT_OUT .
  571. .Pp
  572. .It Sy ZPOOL_SCRIPTS_AS_ROOT
  573. Allow a privileged user to run
  574. .Nm zpool Cm status Ns / Ns Cm iostat Fl c .
  575. Normally, only unprivileged users are allowed to run
  576. .Fl c .
  577. .It Sy ZPOOL_SCRIPTS_PATH
  578. The search path for scripts when running
  579. .Nm zpool Cm status Ns / Ns Cm iostat Fl c .
  580. This is a colon-separated list of directories and overrides the default
  581. .Pa ~/.zpool.d
  582. and
  583. .Pa /etc/zfs/zpool.d
  584. search paths.
  585. .It Sy ZPOOL_SCRIPTS_ENABLED
  586. Allow a user to run
  587. .Nm zpool Cm status Ns / Ns Cm iostat Fl c .
  588. If
  589. .Sy ZPOOL_SCRIPTS_ENABLED
  590. is not set, it is assumed that the user is allowed to run
  591. .Nm zpool Cm status Ns / Ns Cm iostat Fl c .
  592. .\" Shared with zfs.8
  593. .It Sy ZFS_MODULE_TIMEOUT
  594. Time, in seconds, to wait for
  595. .Pa /dev/zfs
  596. to appear.
  597. Defaults to
  598. .Sy 10 ,
  599. max
  600. .Sy 600 Pq 10 minutes .
  601. If
  602. .Pf < Sy 0 ,
  603. wait forever; if
  604. .Sy 0 ,
  605. don't wait.
  606. .El
  607. .
  608. .Sh INTERFACE STABILITY
  609. .Sy Evolving
  610. .
  611. .Sh SEE ALSO
  612. .Xr zfs 4 ,
  613. .Xr zpool-features 7 ,
  614. .Xr zpoolconcepts 7 ,
  615. .Xr zpoolprops 7 ,
  616. .Xr zed 8 ,
  617. .Xr zfs 8 ,
  618. .Xr zpool-add 8 ,
  619. .Xr zpool-attach 8 ,
  620. .Xr zpool-checkpoint 8 ,
  621. .Xr zpool-clear 8 ,
  622. .Xr zpool-create 8 ,
  623. .Xr zpool-ddtprune 8 ,
  624. .Xr zpool-destroy 8 ,
  625. .Xr zpool-detach 8 ,
  626. .Xr zpool-events 8 ,
  627. .Xr zpool-export 8 ,
  628. .Xr zpool-get 8 ,
  629. .Xr zpool-history 8 ,
  630. .Xr zpool-import 8 ,
  631. .Xr zpool-initialize 8 ,
  632. .Xr zpool-iostat 8 ,
  633. .Xr zpool-labelclear 8 ,
  634. .Xr zpool-list 8 ,
  635. .Xr zpool-offline 8 ,
  636. .Xr zpool-online 8 ,
  637. .Xr zpool-prefetch 8 ,
  638. .Xr zpool-reguid 8 ,
  639. .Xr zpool-remove 8 ,
  640. .Xr zpool-reopen 8 ,
  641. .Xr zpool-replace 8 ,
  642. .Xr zpool-resilver 8 ,
  643. .Xr zpool-scrub 8 ,
  644. .Xr zpool-set 8 ,
  645. .Xr zpool-split 8 ,
  646. .Xr zpool-status 8 ,
  647. .Xr zpool-sync 8 ,
  648. .Xr zpool-trim 8 ,
  649. .Xr zpool-upgrade 8 ,
  650. .Xr zpool-wait 8