diff --git a/iscsi-scst/kernel/iscsi.c b/iscsi-scst/kernel/iscsi.c index ab9993650..9e4e7ec36 100644 --- a/iscsi-scst/kernel/iscsi.c +++ b/iscsi-scst/kernel/iscsi.c @@ -1403,7 +1403,7 @@ static int cmnd_prepare_recv_pdu(struct iscsi_conn *conn, while (1) { unsigned int sg_len; char __user *addr; - + if (unlikely(buff_offs >= bufflen)) { TRACE_DBG("Residual overflow (cmd %p, buff_offs %d, " "bufflen %d)", cmd, buff_offs, bufflen); diff --git a/iscsi-scst/usr/iscsi_adm.c b/iscsi-scst/usr/iscsi_adm.c index 168a06503..73e59b095 100644 --- a/iscsi-scst/usr/iscsi_adm.c +++ b/iscsi-scst/usr/iscsi_adm.c @@ -558,7 +558,7 @@ static int user_handle_del(struct iscsi_adm_req *req, char *user, char *pass) fprintf(stderr, "Username must be specified\n"); return -EINVAL; } - + if (pass) fprintf(stderr, "Ignoring specified password\n"); diff --git a/qla2x00t/qla_sup.c b/qla2x00t/qla_sup.c index 1728ab3cc..d83b02028 100644 --- a/qla2x00t/qla_sup.c +++ b/qla2x00t/qla_sup.c @@ -1700,7 +1700,7 @@ qla2x00_write_optrom_data(struct scsi_qla_host *ha, uint8_t *buf, */ rest_addr = 0xffff; sec_mask = 0x10000; - break; + break; } /* * ST m29w010b part - 16kb sector size diff --git a/qla_isp/README.scst b/qla_isp/README.scst index 27d60b47e..6f3825309 100644 --- a/qla_isp/README.scst +++ b/qla_isp/README.scst @@ -12,7 +12,7 @@ advantage over it: support of 24xx and 25xx series of Qlogic adapters. From other side, qla2x00t is simpler, smaller and much better tested on 22xx and 23xx, hence perform more reliable and, thus, is recommended for these adapters. Since 24xx/25xx become fully supported on qla2x00t we -encourage users to switch to this driver. +encourage users to switch to this driver. INSTALLATION diff --git a/scst/README b/scst/README index 5e6c45a59..817fa3ff3 100644 --- a/scst/README +++ b/scst/README @@ -58,7 +58,7 @@ The sysfs build supports only kernels 2.6.26 and higher, because in 2.6.26 internal kernel's sysfs interface had a major change, which made it heavily incompatible with pre-2.6.26 version. -At first, make sure that the link "/lib/modules/`you_kernel_version`/build" +At first, make sure that the link "/lib/modules/`you_kernel_version`/build" points to the source code for your currently running kernel. Then you should consider to apply necessary kernel patches. SCST has the @@ -641,7 +641,7 @@ Standard SCST dev handlers have at least the following common entries: threads. Valid only if threads_num attribute >0. - type - SCSI type of this device - + See below for more information about other entries of this subdirectory of the standard SCST dev handlers. @@ -670,7 +670,7 @@ of the standard SCST dev handlers. has the following entries: - None, one or more subdirectories for each existing SGV cache. - + - global_stats - file containing global SGV caches statistics. Each SGV cache's subdirectory has the following item: @@ -870,7 +870,7 @@ looking inside this file. DEST_GROUP_NAME. - "clear" - deletes all initiators from this group. - + For "add" and "del" commands INITIATOR_NAME can be a simple DOS-type patterns, containing '*' and '?' symbols. '*' means match all any symbols, '?' means match only any single symbol. For instance, @@ -1137,7 +1137,7 @@ For example: echo "add_device disk1 filename=/disk1; blocksize=4096; nv_cache=1" >/sys/kernel/scst_tgt/handlers/vdisk_fileio/mgmt -will create a FILEIO virtual device disk1 with backend file /disk1 +will create a FILEIO virtual device disk1 with backend file /disk1 with block size 4K and NV_CACHE enabled. Each vdisk_fileio's device has the following attributes in @@ -1285,12 +1285,12 @@ Caching ------- By default for performance reasons VDISK FILEIO devices use write back -caching policy. +caching policy. Generally, write back caching is reasonably safe for use and danger of it is greatly overestimated, because: -1. Modern HDDs have at least 16MB of cache working in write back mode by +1. Modern HDDs have at least 16MB of cache working in write back mode by default, so for a 10 drives RAID it is 160MB of a write back cache. You can consider, how many people are happy with it and how many disabled write back cache of their HDDs? Almost all and almost nobody @@ -1304,9 +1304,9 @@ to have acceptable performance their users have to use write back caching, hence on a power loss all not yet committed to flash chips, but acknowledged as written, data will be lost. -2. Most, if not all, modern enterprise level applications are well -prepared to work with write back cached storage. They know well when to -flush the cache and how to flush it to make the lost on crash data +2. Most, if not all, modern enterprise level applications are well +prepared to work with write back cached storage. They know well when to +flush the cache and how to flush it to make the lost on crash data acceptable. For instance, journaled file systems flush cache on each meta data @@ -1322,7 +1322,7 @@ using "barrier=1" and "barrier=flush" mount options correspondingly. You can check if the barriers turn on or off by looking in /proc/mounts. Windows and, AFAIK, other UNIX'es don't need any special explicit options and do necessary barrier actions on write-back caching devices -by default. +by default. But even in case of journaled file systems your unsaved cached data will still be lost in case of power/hardware/software failures, so you may @@ -1339,7 +1339,7 @@ impossible), or need a good UPS to protect yourself from not committed data loss. Note, on some real-life workloads write through caching might perform -better, than write back one with the barrier protection turned on. +better, than write back one with the barrier protection turned on. To limit this data loss with write back caching you can use files in /proc/sys/vm to limit amount of unflushed data in the system cache. @@ -1384,7 +1384,7 @@ Pass-through mode In the pass-through mode (i.e. using the pass-through device handlers scst_disk, scst_tape, etc) SCSI commands, coming from remote initiators, are passed to local SCSI devices on target as is, without any -modifications. +modifications. In the SYSFS interface all real SCSI devices are listed in /sys/kernel/scst_tgt/devices in form host:channel:id:lun numbers, for @@ -1828,6 +1828,6 @@ Thanks to: * Daniel Debonzi for a big part of SCST sysfs tree implementation - + Vladislav Bolkhovitin , http://scst.sourceforge.net diff --git a/scst/README_in-tree b/scst/README_in-tree index 8742d29ac..fd1c64e97 100644 --- a/scst/README_in-tree +++ b/scst/README_in-tree @@ -871,12 +871,12 @@ Caching ------- By default for performance reasons VDISK FILEIO devices use write back -caching policy. +caching policy. Generally, write back caching is reasonably safe for use and danger of it is greatly overestimated, because: -1. Modern HDDs have at least 16MB of cache working in write back mode by +1. Modern HDDs have at least 16MB of cache working in write back mode by default, so for a 10 drives RAID it is 160MB of a write back cache. You can consider, how many people are happy with it and how many disabled write back cache of their HDDs? Almost all and almost nobody @@ -890,9 +890,9 @@ to have acceptable performance their users have to use write back caching, hence on a power loss all not yet committed to flash chips, but acknowledged as written, data will be lost. -2. Most, if not all, modern enterprise level applications are well -prepared to work with write back cached storage. They know well when to -flush the cache and how to flush it to make the lost on crash data +2. Most, if not all, modern enterprise level applications are well +prepared to work with write back cached storage. They know well when to +flush the cache and how to flush it to make the lost on crash data acceptable. For instance, journaled file systems flush cache on each meta data @@ -925,7 +925,7 @@ impossible), or need a good UPS to protect yourself from not committed data loss. Note, on some real-life workloads write through caching might perform -better, than write back one with the barrier protection turned on. +better, than write back one with the barrier protection turned on. To limit this data loss with write back caching you can use files in /proc/sys/vm to limit amount of unflushed data in the system cache. diff --git a/usr/fileio/common.c b/usr/fileio/common.c index 9f9fa727f..316cf8102 100644 --- a/usr/fileio/common.c +++ b/usr/fileio/common.c @@ -1131,7 +1131,7 @@ static int rigid_geo_pg(unsigned char *p, int pcontrol, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0x3a, 0x98/* 15K RPM */, 0, 0}; int32_t ncyl, n; - + memcpy(p, geo_m_pg, sizeof(geo_m_pg)); ncyl = dev->nblocks / (DEF_HEADS * DEF_SECTORS); if ((dev->nblocks % (DEF_HEADS * DEF_SECTORS)) != 0) @@ -1457,7 +1457,7 @@ static void exec_read_capacity(struct vdisk_cmd *vcmd) buffer[6] = (blocksize >> (BYTE * 1)) & 0xFF; buffer[7] = (blocksize >> (BYTE * 0)) & 0xFF; - length = min(length, (int)sizeof(buffer)); + length = min(length, (int)sizeof(buffer)); memcpy(address, buffer, length); @@ -1519,7 +1519,7 @@ static void exec_read_capacity16(struct vdisk_cmd *vcmd) break; } - length = min(length, (int)sizeof(buffer)); + length = min(length, (int)sizeof(buffer)); memcpy(address, buffer, length);