Linux Audio

Check our new training course

Linux kernel drivers training

May 6-19, 2025
Register
Loading...
v3.15
   1/*
   2 * vvvvvvvvvvvvvvvvvvvvvvv Original vvvvvvvvvvvvvvvvvvvvvvvvvvvvvvv
   3 *  Copyright (C) 1992  Eric Youngdale
   4 *  Simulate a host adapter with 2 disks attached.  Do a lot of checking
   5 *  to make sure that we are not getting blocks mixed up, and PANIC if
   6 *  anything out of the ordinary is seen.
   7 * ^^^^^^^^^^^^^^^^^^^^^^^ Original ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
   8 *
   9 *  This version is more generic, simulating a variable number of disk
  10 *  (or disk like devices) sharing a common amount of RAM. To be more
  11 *  realistic, the simulated devices have the transport attributes of
  12 *  SAS disks.
  13 *
 
 
 
 
  14 *
  15 *  For documentation see http://sg.danny.cz/sg/sdebug26.html
  16 *
  17 *   D. Gilbert (dpg) work for Magneto-Optical device test [20010421]
  18 *   dpg: work for devfs large number of disks [20010809]
  19 *        forked for lk 2.5 series [20011216, 20020101]
  20 *        use vmalloc() more inquiry+mode_sense [20020302]
  21 *        add timers for delayed responses [20020721]
  22 *   Patrick Mansfield <patmans@us.ibm.com> max_luns+scsi_level [20021031]
  23 *   Mike Anderson <andmike@us.ibm.com> sysfs work [20021118]
  24 *   dpg: change style of boot options to "scsi_debug.num_tgts=2" and
  25 *        module options to "modprobe scsi_debug num_tgts=2" [20021221]
  26 */
  27
 
 
 
  28#include <linux/module.h>
  29
  30#include <linux/kernel.h>
  31#include <linux/errno.h>
  32#include <linux/timer.h>
  33#include <linux/slab.h>
  34#include <linux/types.h>
  35#include <linux/string.h>
  36#include <linux/genhd.h>
  37#include <linux/fs.h>
  38#include <linux/init.h>
  39#include <linux/proc_fs.h>
  40#include <linux/vmalloc.h>
  41#include <linux/moduleparam.h>
  42#include <linux/scatterlist.h>
  43#include <linux/blkdev.h>
  44#include <linux/crc-t10dif.h>
 
 
 
 
 
 
  45
  46#include <net/checksum.h>
  47
  48#include <asm/unaligned.h>
  49
  50#include <scsi/scsi.h>
  51#include <scsi/scsi_cmnd.h>
  52#include <scsi/scsi_device.h>
  53#include <scsi/scsi_host.h>
  54#include <scsi/scsicam.h>
  55#include <scsi/scsi_eh.h>
 
  56#include <scsi/scsi_dbg.h>
  57
  58#include "sd.h"
  59#include "scsi_logging.h"
  60
  61#define SCSI_DEBUG_VERSION "1.82"
  62static const char * scsi_debug_version_date = "20100324";
 
 
 
  63
  64/* Additional Sense Code (ASC) */
  65#define NO_ADDITIONAL_SENSE 0x0
  66#define LOGICAL_UNIT_NOT_READY 0x4
  67#define LOGICAL_UNIT_COMMUNICATION_FAILURE 0x8
  68#define UNRECOVERED_READ_ERR 0x11
  69#define PARAMETER_LIST_LENGTH_ERR 0x1a
  70#define INVALID_OPCODE 0x20
  71#define ADDR_OUT_OF_RANGE 0x21
  72#define INVALID_COMMAND_OPCODE 0x20
  73#define INVALID_FIELD_IN_CDB 0x24
  74#define INVALID_FIELD_IN_PARAM_LIST 0x26
  75#define POWERON_RESET 0x29
 
 
 
 
 
 
 
 
 
  76#define SAVING_PARAMS_UNSUP 0x39
  77#define TRANSPORT_PROBLEM 0x4b
  78#define THRESHOLD_EXCEEDED 0x5d
  79#define LOW_POWER_COND_ON 0x5e
 
 
 
  80
  81/* Additional Sense Code Qualifier (ASCQ) */
  82#define ACK_NAK_TO 0x3
  83
  84#define SDEBUG_TAGGED_QUEUING 0 /* 0 | MSG_SIMPLE_TAG | MSG_ORDERED_TAG */
  85
  86/* Default values for driver parameters */
  87#define DEF_NUM_HOST   1
  88#define DEF_NUM_TGTS   1
  89#define DEF_MAX_LUNS   1
  90/* With these defaults, this driver will make 1 host with 1 target
  91 * (id 0) containing 1 logical unit (lun 0). That is 1 device.
  92 */
  93#define DEF_ATO 1
  94#define DEF_DELAY   1
  95#define DEF_DEV_SIZE_MB   8
  96#define DEF_DIF 0
  97#define DEF_DIX 0
  98#define DEF_D_SENSE   0
  99#define DEF_EVERY_NTH   0
 100#define DEF_FAKE_RW	0
 101#define DEF_GUARD 0
 
 102#define DEF_LBPU 0
 103#define DEF_LBPWS 0
 104#define DEF_LBPWS10 0
 105#define DEF_LBPRZ 1
 106#define DEF_LOWEST_ALIGNED 0
 
 107#define DEF_NO_LUN_0   0
 108#define DEF_NUM_PARTS   0
 109#define DEF_OPTS   0
 110#define DEF_OPT_BLKS 64
 111#define DEF_PHYSBLK_EXP 0
 112#define DEF_PTYPE   0
 113#define DEF_REMOVABLE false
 114#define DEF_SCSI_LEVEL   5    /* INQUIRY, byte2 [5->SPC-3] */
 115#define DEF_SECTOR_SIZE 512
 116#define DEF_UNMAP_ALIGNMENT 0
 117#define DEF_UNMAP_GRANULARITY 1
 118#define DEF_UNMAP_MAX_BLOCKS 0xFFFFFFFF
 119#define DEF_UNMAP_MAX_DESC 256
 120#define DEF_VIRTUAL_GB   0
 121#define DEF_VPD_USE_HOSTNO 1
 122#define DEF_WRITESAME_LENGTH 0xFFFF
 123
 124/* bit mask values for scsi_debug_opts */
 125#define SCSI_DEBUG_OPT_NOISE   1
 126#define SCSI_DEBUG_OPT_MEDIUM_ERR   2
 127#define SCSI_DEBUG_OPT_TIMEOUT   4
 128#define SCSI_DEBUG_OPT_RECOVERED_ERR   8
 129#define SCSI_DEBUG_OPT_TRANSPORT_ERR   16
 130#define SCSI_DEBUG_OPT_DIF_ERR   32
 131#define SCSI_DEBUG_OPT_DIX_ERR   64
 132#define SCSI_DEBUG_OPT_MAC_TIMEOUT  128
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 133/* When "every_nth" > 0 then modulo "every_nth" commands:
 134 *   - a no response is simulated if SCSI_DEBUG_OPT_TIMEOUT is set
 135 *   - a RECOVERED_ERROR is simulated on successful read and write
 136 *     commands if SCSI_DEBUG_OPT_RECOVERED_ERR is set.
 137 *   - a TRANSPORT_ERROR is simulated on successful read and write
 138 *     commands if SCSI_DEBUG_OPT_TRANSPORT_ERR is set.
 139 *
 140 * When "every_nth" < 0 then after "- every_nth" commands:
 141 *   - a no response is simulated if SCSI_DEBUG_OPT_TIMEOUT is set
 142 *   - a RECOVERED_ERROR is simulated on successful read and write
 143 *     commands if SCSI_DEBUG_OPT_RECOVERED_ERR is set.
 144 *   - a TRANSPORT_ERROR is simulated on successful read and write
 145 *     commands if SCSI_DEBUG_OPT_TRANSPORT_ERR is set.
 146 * This will continue until some other action occurs (e.g. the user
 147 * writing a new value (other than -1 or 1) to every_nth via sysfs).
 
 148 */
 149
 150/* when 1==SCSI_DEBUG_OPT_MEDIUM_ERR, a medium error is simulated at this
 
 
 
 
 
 
 
 
 
 
 
 
 
 151 * sector on read commands: */
 152#define OPT_MEDIUM_ERR_ADDR   0x1234 /* that's sector 4660 in decimal */
 153#define OPT_MEDIUM_ERR_NUM    10     /* number of consecutive medium errs */
 154
 155/* If REPORT LUNS has luns >= 256 it can choose "flat space" (value 1)
 156 * or "peripheral device" addressing (value 0) */
 157#define SAM2_LUN_ADDRESS_METHOD 0
 158#define SAM2_WLUN_REPORT_LUNS 0xc101
 159
 160/* Can queue up to this number of commands. Typically commands that
 161 * that have a non-zero delay are queued. */
 162#define SCSI_DEBUG_CANQUEUE  255
 163
 164static int scsi_debug_add_host = DEF_NUM_HOST;
 165static int scsi_debug_ato = DEF_ATO;
 166static int scsi_debug_delay = DEF_DELAY;
 167static int scsi_debug_dev_size_mb = DEF_DEV_SIZE_MB;
 168static int scsi_debug_dif = DEF_DIF;
 169static int scsi_debug_dix = DEF_DIX;
 170static int scsi_debug_dsense = DEF_D_SENSE;
 171static int scsi_debug_every_nth = DEF_EVERY_NTH;
 172static int scsi_debug_fake_rw = DEF_FAKE_RW;
 173static unsigned int scsi_debug_guard = DEF_GUARD;
 174static int scsi_debug_lowest_aligned = DEF_LOWEST_ALIGNED;
 175static int scsi_debug_max_luns = DEF_MAX_LUNS;
 176static int scsi_debug_max_queue = SCSI_DEBUG_CANQUEUE;
 177static int scsi_debug_no_lun_0 = DEF_NO_LUN_0;
 178static int scsi_debug_no_uld = 0;
 179static int scsi_debug_num_parts = DEF_NUM_PARTS;
 180static int scsi_debug_num_tgts = DEF_NUM_TGTS; /* targets per host */
 181static int scsi_debug_opt_blks = DEF_OPT_BLKS;
 182static int scsi_debug_opts = DEF_OPTS;
 183static int scsi_debug_physblk_exp = DEF_PHYSBLK_EXP;
 184static int scsi_debug_ptype = DEF_PTYPE; /* SCSI peripheral type (0==disk) */
 185static int scsi_debug_scsi_level = DEF_SCSI_LEVEL;
 186static int scsi_debug_sector_size = DEF_SECTOR_SIZE;
 187static int scsi_debug_virtual_gb = DEF_VIRTUAL_GB;
 188static int scsi_debug_vpd_use_hostno = DEF_VPD_USE_HOSTNO;
 189static unsigned int scsi_debug_lbpu = DEF_LBPU;
 190static unsigned int scsi_debug_lbpws = DEF_LBPWS;
 191static unsigned int scsi_debug_lbpws10 = DEF_LBPWS10;
 192static unsigned int scsi_debug_lbprz = DEF_LBPRZ;
 193static unsigned int scsi_debug_unmap_alignment = DEF_UNMAP_ALIGNMENT;
 194static unsigned int scsi_debug_unmap_granularity = DEF_UNMAP_GRANULARITY;
 195static unsigned int scsi_debug_unmap_max_blocks = DEF_UNMAP_MAX_BLOCKS;
 196static unsigned int scsi_debug_unmap_max_desc = DEF_UNMAP_MAX_DESC;
 197static unsigned int scsi_debug_write_same_length = DEF_WRITESAME_LENGTH;
 198static bool scsi_debug_removable = DEF_REMOVABLE;
 199static bool scsi_debug_clustering;
 200
 201static int scsi_debug_cmnd_count = 0;
 202
 203#define DEV_READONLY(TGT)      (0)
 204
 205static unsigned int sdebug_store_sectors;
 206static sector_t sdebug_capacity;	/* in sectors */
 207
 208/* old BIOS stuff, kernel may get rid of them but some mode sense pages
 209   may still need them */
 210static int sdebug_heads;		/* heads per disk */
 211static int sdebug_cylinders_per;	/* cylinders per surface */
 212static int sdebug_sectors_per;		/* sectors per cylinder */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 213
 214#define SDEBUG_MAX_PARTS 4
 215
 216#define SDEBUG_SENSE_LEN 32
 217
 218#define SCSI_DEBUG_MAX_CMD_LEN 32
 219
 220static unsigned int scsi_debug_lbp(void)
 221{
 222	return scsi_debug_lbpu | scsi_debug_lbpws | scsi_debug_lbpws10;
 223}
 224
 225struct sdebug_dev_info {
 226	struct list_head dev_list;
 227	unsigned char sense_buff[SDEBUG_SENSE_LEN];	/* weak nexus */
 228	unsigned int channel;
 229	unsigned int target;
 230	unsigned int lun;
 
 231	struct sdebug_host_info *sdbg_host;
 232	unsigned int wlun;
 233	char reset;
 234	char stopped;
 235	char used;
 236};
 237
 238struct sdebug_host_info {
 239	struct list_head host_list;
 240	struct Scsi_Host *shost;
 241	struct device dev;
 242	struct list_head dev_info_list;
 243};
 244
 245#define to_sdebug_host(d)	\
 246	container_of(d, struct sdebug_host_info, dev)
 247
 248static LIST_HEAD(sdebug_host_list);
 249static DEFINE_SPINLOCK(sdebug_host_list_lock);
 250
 251typedef void (* done_funct_t) (struct scsi_cmnd *);
 
 
 
 252
 253struct sdebug_queued_cmd {
 254	int in_use;
 255	struct timer_list cmnd_timer;
 256	done_funct_t done_funct;
 257	struct scsi_cmnd * a_cmnd;
 258	int scsi_result;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 259};
 260static struct sdebug_queued_cmd queued_arr[SCSI_DEBUG_CANQUEUE];
 261
 262static unsigned char * fake_storep;	/* ramdisk storage */
 263static struct sd_dif_tuple *dif_storep;	/* protection info */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 264static void *map_storep;		/* provisioning map */
 265
 266static unsigned long map_size;
 267static int num_aborts = 0;
 268static int num_dev_resets = 0;
 269static int num_bus_resets = 0;
 270static int num_host_resets = 0;
 
 271static int dix_writes;
 272static int dix_reads;
 273static int dif_errors;
 274
 275static DEFINE_SPINLOCK(queued_arr_lock);
 
 
 276static DEFINE_RWLOCK(atomic_rw);
 277
 278static char sdebug_proc_name[] = "scsi_debug";
 
 279
 280static struct bus_type pseudo_lld_bus;
 281
 282static struct device_driver sdebug_driverfs_driver = {
 283	.name 		= sdebug_proc_name,
 284	.bus		= &pseudo_lld_bus,
 285};
 286
 287static const int check_condition_result =
 288		(DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION;
 289
 290static const int illegal_condition_result =
 291	(DRIVER_SENSE << 24) | (DID_ABORT << 16) | SAM_STAT_CHECK_CONDITION;
 292
 293static unsigned char ctrl_m_pg[] = {0xa, 10, 2, 0, 0, 0, 0, 0,
 294				    0, 0, 0x2, 0x4b};
 295static unsigned char iec_m_pg[] = {0x1c, 0xa, 0x08, 0, 0, 0, 0, 0,
 296			           0, 0, 0x0, 0x0};
 
 
 
 
 
 
 
 
 
 297
 298static void *fake_store(unsigned long long lba)
 299{
 300	lba = do_div(lba, sdebug_store_sectors);
 301
 302	return fake_storep + lba * scsi_debug_sector_size;
 303}
 304
 305static struct sd_dif_tuple *dif_store(sector_t sector)
 306{
 307	sector = do_div(sector, sdebug_store_sectors);
 308
 309	return dif_storep + sector;
 310}
 311
 312static int sdebug_add_adapter(void);
 313static void sdebug_remove_adapter(void);
 314
 315static void sdebug_max_tgts_luns(void)
 316{
 317	struct sdebug_host_info *sdbg_host;
 318	struct Scsi_Host *hpnt;
 319
 320	spin_lock(&sdebug_host_list_lock);
 321	list_for_each_entry(sdbg_host, &sdebug_host_list, host_list) {
 322		hpnt = sdbg_host->shost;
 323		if ((hpnt->this_id >= 0) &&
 324		    (scsi_debug_num_tgts > hpnt->this_id))
 325			hpnt->max_id = scsi_debug_num_tgts + 1;
 326		else
 327			hpnt->max_id = scsi_debug_num_tgts;
 328		/* scsi_debug_max_luns; */
 329		hpnt->max_lun = SAM2_WLUN_REPORT_LUNS;
 330	}
 331	spin_unlock(&sdebug_host_list_lock);
 332}
 333
 334static void mk_sense_buffer(struct sdebug_dev_info *devip, int key,
 335			    int asc, int asq)
 
 
 
 
 336{
 337	unsigned char *sbuff;
 
 
 338
 339	sbuff = devip->sense_buff;
 340	memset(sbuff, 0, SDEBUG_SENSE_LEN);
 341
 342	scsi_build_sense_buffer(scsi_debug_dsense, sbuff, key, asc, asq);
 343
 344	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
 345		printk(KERN_INFO "scsi_debug:    [sense_key,asc,ascq]: "
 346		      "[0x%x,0x%x,0x%x]\n", key, asc, asq);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 347}
 348
 349static void get_data_transfer_info(unsigned char *cmd,
 350				   unsigned long long *lba, unsigned int *num,
 351				   u32 *ei_lba)
 352{
 353	*ei_lba = 0;
 354
 355	switch (*cmd) {
 356	case VARIABLE_LENGTH_CMD:
 357		*lba = (u64)cmd[19] | (u64)cmd[18] << 8 |
 358			(u64)cmd[17] << 16 | (u64)cmd[16] << 24 |
 359			(u64)cmd[15] << 32 | (u64)cmd[14] << 40 |
 360			(u64)cmd[13] << 48 | (u64)cmd[12] << 56;
 361
 362		*ei_lba = (u32)cmd[23] | (u32)cmd[22] << 8 |
 363			(u32)cmd[21] << 16 | (u32)cmd[20] << 24;
 364
 365		*num = (u32)cmd[31] | (u32)cmd[30] << 8 | (u32)cmd[29] << 16 |
 366			(u32)cmd[28] << 24;
 367		break;
 368
 369	case WRITE_SAME_16:
 370	case WRITE_16:
 371	case READ_16:
 372		*lba = (u64)cmd[9] | (u64)cmd[8] << 8 |
 373			(u64)cmd[7] << 16 | (u64)cmd[6] << 24 |
 374			(u64)cmd[5] << 32 | (u64)cmd[4] << 40 |
 375			(u64)cmd[3] << 48 | (u64)cmd[2] << 56;
 376
 377		*num = (u32)cmd[13] | (u32)cmd[12] << 8 | (u32)cmd[11] << 16 |
 378			(u32)cmd[10] << 24;
 379		break;
 380	case WRITE_12:
 381	case READ_12:
 382		*lba = (u32)cmd[5] | (u32)cmd[4] << 8 | (u32)cmd[3] << 16 |
 383			(u32)cmd[2] << 24;
 384
 385		*num = (u32)cmd[9] | (u32)cmd[8] << 8 | (u32)cmd[7] << 16 |
 386			(u32)cmd[6] << 24;
 387		break;
 388	case WRITE_SAME:
 389	case WRITE_10:
 390	case READ_10:
 391	case XDWRITEREAD_10:
 392		*lba = (u32)cmd[5] | (u32)cmd[4] << 8 |	(u32)cmd[3] << 16 |
 393			(u32)cmd[2] << 24;
 394
 395		*num = (u32)cmd[8] | (u32)cmd[7] << 8;
 396		break;
 397	case WRITE_6:
 398	case READ_6:
 399		*lba = (u32)cmd[3] | (u32)cmd[2] << 8 |
 400			(u32)(cmd[1] & 0x1f) << 16;
 401		*num = (0 == cmd[4]) ? 256 : cmd[4];
 402		break;
 403	default:
 404		break;
 405	}
 406}
 407
 408static int scsi_debug_ioctl(struct scsi_device *dev, int cmd, void __user *arg)
 409{
 410	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts) {
 411		printk(KERN_INFO "scsi_debug: ioctl: cmd=0x%x\n", cmd);
 
 
 
 
 
 
 
 
 
 412	}
 413	return -EINVAL;
 414	/* return -ENOTTY; // correct return but upsets fdisk */
 415}
 416
 417static int check_readiness(struct scsi_cmnd * SCpnt, int reset_only,
 418			   struct sdebug_dev_info * devip)
 419{
 420	if (devip->reset) {
 421		if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
 422			printk(KERN_INFO "scsi_debug: Reporting Unit "
 423			       "attention: power on reset\n");
 424		devip->reset = 0;
 425		mk_sense_buffer(devip, UNIT_ATTENTION, POWERON_RESET, 0);
 426		return check_condition_result;
 
 
 
 427	}
 428	if ((0 == reset_only) && devip->stopped) {
 429		if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
 430			printk(KERN_INFO "scsi_debug: Reporting Not "
 431			       "ready: initializing command required\n");
 432		mk_sense_buffer(devip, NOT_READY, LOGICAL_UNIT_NOT_READY,
 433				0x2);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 434		return check_condition_result;
 435	}
 436	return 0;
 437}
 438
 439/* Returns 0 if ok else (DID_ERROR << 16). Sets scp->resid . */
 440static int fill_from_dev_buffer(struct scsi_cmnd *scp, unsigned char *arr,
 441				int arr_len)
 442{
 443	int act_len;
 444	struct scsi_data_buffer *sdb = scsi_in(scp);
 445
 446	if (!sdb->length)
 447		return 0;
 448	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_FROM_DEVICE))
 449		return (DID_ERROR << 16);
 450
 451	act_len = sg_copy_from_buffer(sdb->table.sgl, sdb->table.nents,
 452				      arr, arr_len);
 453	sdb->resid = scsi_bufflen(scp) - act_len;
 454
 455	return 0;
 456}
 457
 458/* Returns number of bytes fetched into 'arr' or -1 if error. */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 459static int fetch_to_dev_buffer(struct scsi_cmnd *scp, unsigned char *arr,
 460			       int arr_len)
 461{
 462	if (!scsi_bufflen(scp))
 463		return 0;
 464	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_TO_DEVICE))
 465		return -1;
 466
 467	return scsi_sg_copy_to_buffer(scp, arr, arr_len);
 468}
 469
 470
 471static const char * inq_vendor_id = "Linux   ";
 472static const char * inq_product_id = "scsi_debug      ";
 473static const char * inq_product_rev = "0004";
 474
 475static int inquiry_evpd_83(unsigned char * arr, int port_group_id,
 476			   int target_dev_id, int dev_id_num,
 477			   const char * dev_id_str,
 478			   int dev_id_str_len)
 
 
 
 
 
 479{
 480	int num, port_a;
 481	char b[32];
 482
 483	port_a = target_dev_id + 1;
 484	/* T10 vendor identifier field format (faked) */
 485	arr[0] = 0x2;	/* ASCII */
 486	arr[1] = 0x1;
 487	arr[2] = 0x0;
 488	memcpy(&arr[4], inq_vendor_id, 8);
 489	memcpy(&arr[12], inq_product_id, 16);
 490	memcpy(&arr[28], dev_id_str, dev_id_str_len);
 491	num = 8 + 16 + dev_id_str_len;
 492	arr[3] = num;
 493	num += 4;
 494	if (dev_id_num >= 0) {
 495		/* NAA-5, Logical unit identifier (binary) */
 496		arr[num++] = 0x1;	/* binary (not necessarily sas) */
 497		arr[num++] = 0x3;	/* PIV=0, lu, naa */
 498		arr[num++] = 0x0;
 499		arr[num++] = 0x8;
 500		arr[num++] = 0x53;  /* naa-5 ieee company id=0x333333 (fake) */
 501		arr[num++] = 0x33;
 502		arr[num++] = 0x33;
 503		arr[num++] = 0x30;
 504		arr[num++] = (dev_id_num >> 24);
 505		arr[num++] = (dev_id_num >> 16) & 0xff;
 506		arr[num++] = (dev_id_num >> 8) & 0xff;
 507		arr[num++] = dev_id_num & 0xff;
 
 
 
 
 
 
 508		/* Target relative port number */
 509		arr[num++] = 0x61;	/* proto=sas, binary */
 510		arr[num++] = 0x94;	/* PIV=1, target port, rel port */
 511		arr[num++] = 0x0;	/* reserved */
 512		arr[num++] = 0x4;	/* length */
 513		arr[num++] = 0x0;	/* reserved */
 514		arr[num++] = 0x0;	/* reserved */
 515		arr[num++] = 0x0;
 516		arr[num++] = 0x1;	/* relative port A */
 517	}
 518	/* NAA-5, Target port identifier */
 519	arr[num++] = 0x61;	/* proto=sas, binary */
 520	arr[num++] = 0x93;	/* piv=1, target port, naa */
 521	arr[num++] = 0x0;
 522	arr[num++] = 0x8;
 523	arr[num++] = 0x52;	/* naa-5, company id=0x222222 (fake) */
 524	arr[num++] = 0x22;
 525	arr[num++] = 0x22;
 526	arr[num++] = 0x20;
 527	arr[num++] = (port_a >> 24);
 528	arr[num++] = (port_a >> 16) & 0xff;
 529	arr[num++] = (port_a >> 8) & 0xff;
 530	arr[num++] = port_a & 0xff;
 531	/* NAA-5, Target port group identifier */
 532	arr[num++] = 0x61;	/* proto=sas, binary */
 533	arr[num++] = 0x95;	/* piv=1, target port group id */
 534	arr[num++] = 0x0;
 535	arr[num++] = 0x4;
 536	arr[num++] = 0;
 537	arr[num++] = 0;
 538	arr[num++] = (port_group_id >> 8) & 0xff;
 539	arr[num++] = port_group_id & 0xff;
 540	/* NAA-5, Target device identifier */
 541	arr[num++] = 0x61;	/* proto=sas, binary */
 542	arr[num++] = 0xa3;	/* piv=1, target device, naa */
 543	arr[num++] = 0x0;
 544	arr[num++] = 0x8;
 545	arr[num++] = 0x52;	/* naa-5, company id=0x222222 (fake) */
 546	arr[num++] = 0x22;
 547	arr[num++] = 0x22;
 548	arr[num++] = 0x20;
 549	arr[num++] = (target_dev_id >> 24);
 550	arr[num++] = (target_dev_id >> 16) & 0xff;
 551	arr[num++] = (target_dev_id >> 8) & 0xff;
 552	arr[num++] = target_dev_id & 0xff;
 553	/* SCSI name string: Target device identifier */
 554	arr[num++] = 0x63;	/* proto=sas, UTF-8 */
 555	arr[num++] = 0xa8;	/* piv=1, target device, SCSI name string */
 556	arr[num++] = 0x0;
 557	arr[num++] = 24;
 558	memcpy(arr + num, "naa.52222220", 12);
 559	num += 12;
 560	snprintf(b, sizeof(b), "%08X", target_dev_id);
 561	memcpy(arr + num, b, 8);
 562	num += 8;
 563	memset(arr + num, 0, 4);
 564	num += 4;
 565	return num;
 566}
 567
 568
 569static unsigned char vpd84_data[] = {
 570/* from 4th byte */ 0x22,0x22,0x22,0x0,0xbb,0x0,
 571    0x22,0x22,0x22,0x0,0xbb,0x1,
 572    0x22,0x22,0x22,0x0,0xbb,0x2,
 573};
 574
 575static int inquiry_evpd_84(unsigned char * arr)
 
 576{
 577	memcpy(arr, vpd84_data, sizeof(vpd84_data));
 578	return sizeof(vpd84_data);
 579}
 580
 581static int inquiry_evpd_85(unsigned char * arr)
 
 582{
 583	int num = 0;
 584	const char * na1 = "https://www.kernel.org/config";
 585	const char * na2 = "http://www.kernel.org/log";
 586	int plen, olen;
 587
 588	arr[num++] = 0x1;	/* lu, storage config */
 589	arr[num++] = 0x0;	/* reserved */
 590	arr[num++] = 0x0;
 591	olen = strlen(na1);
 592	plen = olen + 1;
 593	if (plen % 4)
 594		plen = ((plen / 4) + 1) * 4;
 595	arr[num++] = plen;	/* length, null termianted, padded */
 596	memcpy(arr + num, na1, olen);
 597	memset(arr + num + olen, 0, plen - olen);
 598	num += plen;
 599
 600	arr[num++] = 0x4;	/* lu, logging */
 601	arr[num++] = 0x0;	/* reserved */
 602	arr[num++] = 0x0;
 603	olen = strlen(na2);
 604	plen = olen + 1;
 605	if (plen % 4)
 606		plen = ((plen / 4) + 1) * 4;
 607	arr[num++] = plen;	/* length, null terminated, padded */
 608	memcpy(arr + num, na2, olen);
 609	memset(arr + num + olen, 0, plen - olen);
 610	num += plen;
 611
 612	return num;
 613}
 614
 615/* SCSI ports VPD page */
 616static int inquiry_evpd_88(unsigned char * arr, int target_dev_id)
 617{
 618	int num = 0;
 619	int port_a, port_b;
 620
 621	port_a = target_dev_id + 1;
 622	port_b = port_a + 1;
 623	arr[num++] = 0x0;	/* reserved */
 624	arr[num++] = 0x0;	/* reserved */
 625	arr[num++] = 0x0;
 626	arr[num++] = 0x1;	/* relative port 1 (primary) */
 627	memset(arr + num, 0, 6);
 628	num += 6;
 629	arr[num++] = 0x0;
 630	arr[num++] = 12;	/* length tp descriptor */
 631	/* naa-5 target port identifier (A) */
 632	arr[num++] = 0x61;	/* proto=sas, binary */
 633	arr[num++] = 0x93;	/* PIV=1, target port, NAA */
 634	arr[num++] = 0x0;	/* reserved */
 635	arr[num++] = 0x8;	/* length */
 636	arr[num++] = 0x52;	/* NAA-5, company_id=0x222222 (fake) */
 637	arr[num++] = 0x22;
 638	arr[num++] = 0x22;
 639	arr[num++] = 0x20;
 640	arr[num++] = (port_a >> 24);
 641	arr[num++] = (port_a >> 16) & 0xff;
 642	arr[num++] = (port_a >> 8) & 0xff;
 643	arr[num++] = port_a & 0xff;
 644
 645	arr[num++] = 0x0;	/* reserved */
 646	arr[num++] = 0x0;	/* reserved */
 647	arr[num++] = 0x0;
 648	arr[num++] = 0x2;	/* relative port 2 (secondary) */
 649	memset(arr + num, 0, 6);
 650	num += 6;
 651	arr[num++] = 0x0;
 652	arr[num++] = 12;	/* length tp descriptor */
 653	/* naa-5 target port identifier (B) */
 654	arr[num++] = 0x61;	/* proto=sas, binary */
 655	arr[num++] = 0x93;	/* PIV=1, target port, NAA */
 656	arr[num++] = 0x0;	/* reserved */
 657	arr[num++] = 0x8;	/* length */
 658	arr[num++] = 0x52;	/* NAA-5, company_id=0x222222 (fake) */
 659	arr[num++] = 0x22;
 660	arr[num++] = 0x22;
 661	arr[num++] = 0x20;
 662	arr[num++] = (port_b >> 24);
 663	arr[num++] = (port_b >> 16) & 0xff;
 664	arr[num++] = (port_b >> 8) & 0xff;
 665	arr[num++] = port_b & 0xff;
 666
 667	return num;
 668}
 669
 670
 671static unsigned char vpd89_data[] = {
 672/* from 4th byte */ 0,0,0,0,
 673'l','i','n','u','x',' ',' ',' ',
 674'S','A','T',' ','s','c','s','i','_','d','e','b','u','g',' ',' ',
 675'1','2','3','4',
 6760x34,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,
 6770xec,0,0,0,
 6780x5a,0xc,0xff,0x3f,0x37,0xc8,0x10,0,0,0,0,0,0x3f,0,0,0,
 6790,0,0,0,0x58,0x58,0x58,0x58,0x58,0x58,0x58,0x58,0x20,0x20,0x20,0x20,
 6800x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0,0,0,0x40,0x4,0,0x2e,0x33,
 6810x38,0x31,0x20,0x20,0x20,0x20,0x54,0x53,0x38,0x33,0x30,0x30,0x33,0x31,
 6820x53,0x41,
 6830x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,
 6840x20,0x20,
 6850x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,
 6860x10,0x80,
 6870,0,0,0x2f,0,0,0,0x2,0,0x2,0x7,0,0xff,0xff,0x1,0,
 6880x3f,0,0xc1,0xff,0x3e,0,0x10,0x1,0xb0,0xf8,0x50,0x9,0,0,0x7,0,
 6890x3,0,0x78,0,0x78,0,0xf0,0,0x78,0,0,0,0,0,0,0,
 6900,0,0,0,0,0,0,0,0x2,0,0,0,0,0,0,0,
 6910x7e,0,0x1b,0,0x6b,0x34,0x1,0x7d,0x3,0x40,0x69,0x34,0x1,0x3c,0x3,0x40,
 6920x7f,0x40,0,0,0,0,0xfe,0xfe,0,0,0,0,0,0xfe,0,0,
 6930,0,0,0,0,0,0,0,0xb0,0xf8,0x50,0x9,0,0,0,0,
 6940,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 6950,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 6960,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 6970x1,0,0xb0,0xf8,0x50,0x9,0xb0,0xf8,0x50,0x9,0x20,0x20,0x2,0,0xb6,0x42,
 6980,0x80,0x8a,0,0x6,0x3c,0xa,0x3c,0xff,0xff,0xc6,0x7,0,0x1,0,0x8,
 6990xf0,0xf,0,0x10,0x2,0,0x30,0,0,0,0,0,0,0,0x6,0xfe,
 7000,0,0x2,0,0x50,0,0x8a,0,0x4f,0x95,0,0,0x21,0,0xb,0,
 7010,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7020,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7030,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7040,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7050,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7060,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7070,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7080,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7090,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7100,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7110,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 7120,0,0,0,0,0,0,0,0,0,0,0,0,0,0xa5,0x51,
 713};
 714
 715static int inquiry_evpd_89(unsigned char * arr)
 
 716{
 717	memcpy(arr, vpd89_data, sizeof(vpd89_data));
 718	return sizeof(vpd89_data);
 719}
 720
 721
 722/* Block limits VPD page (SBC-3) */
 723static unsigned char vpdb0_data[] = {
 724	/* from 4th byte */ 0,0,0,4, 0,0,0x4,0, 0,0,0,64,
 725	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 726	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 727	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
 728};
 729
 730static int inquiry_evpd_b0(unsigned char * arr)
 
 731{
 732	unsigned int gran;
 733
 734	memcpy(arr, vpdb0_data, sizeof(vpdb0_data));
 735
 736	/* Optimal transfer length granularity */
 737	gran = 1 << scsi_debug_physblk_exp;
 738	arr[2] = (gran >> 8) & 0xff;
 739	arr[3] = gran & 0xff;
 740
 741	/* Maximum Transfer Length */
 742	if (sdebug_store_sectors > 0x400) {
 743		arr[4] = (sdebug_store_sectors >> 24) & 0xff;
 744		arr[5] = (sdebug_store_sectors >> 16) & 0xff;
 745		arr[6] = (sdebug_store_sectors >> 8) & 0xff;
 746		arr[7] = sdebug_store_sectors & 0xff;
 747	}
 748
 749	/* Optimal Transfer Length */
 750	put_unaligned_be32(scsi_debug_opt_blks, &arr[8]);
 751
 752	if (scsi_debug_lbpu) {
 753		/* Maximum Unmap LBA Count */
 754		put_unaligned_be32(scsi_debug_unmap_max_blocks, &arr[16]);
 755
 756		/* Maximum Unmap Block Descriptor Count */
 757		put_unaligned_be32(scsi_debug_unmap_max_desc, &arr[20]);
 758	}
 759
 760	/* Unmap Granularity Alignment */
 761	if (scsi_debug_unmap_alignment) {
 762		put_unaligned_be32(scsi_debug_unmap_alignment, &arr[28]);
 763		arr[28] |= 0x80; /* UGAVALID */
 764	}
 765
 766	/* Optimal Unmap Granularity */
 767	put_unaligned_be32(scsi_debug_unmap_granularity, &arr[24]);
 768
 769	/* Maximum WRITE SAME Length */
 770	put_unaligned_be64(scsi_debug_write_same_length, &arr[32]);
 771
 772	return 0x3c; /* Mandatory page length for Logical Block Provisioning */
 773
 774	return sizeof(vpdb0_data);
 775}
 776
 777/* Block device characteristics VPD page (SBC-3) */
 778static int inquiry_evpd_b1(unsigned char *arr)
 779{
 780	memset(arr, 0, 0x3c);
 781	arr[0] = 0;
 782	arr[1] = 1;	/* non rotating medium (e.g. solid state) */
 783	arr[2] = 0;
 784	arr[3] = 5;	/* less than 1.8" */
 785
 786	return 0x3c;
 787}
 788
 789/* Logical block provisioning VPD page (SBC-3) */
 790static int inquiry_evpd_b2(unsigned char *arr)
 791{
 792	memset(arr, 0, 0x4);
 793	arr[0] = 0;			/* threshold exponent */
 794
 795	if (scsi_debug_lbpu)
 796		arr[1] = 1 << 7;
 797
 798	if (scsi_debug_lbpws)
 799		arr[1] |= 1 << 6;
 800
 801	if (scsi_debug_lbpws10)
 802		arr[1] |= 1 << 5;
 803
 804	if (scsi_debug_lbprz)
 805		arr[1] |= 1 << 2;
 806
 
 807	return 0x4;
 808}
 809
 810#define SDEBUG_LONG_INQ_SZ 96
 811#define SDEBUG_MAX_INQ_ARR_SZ 584
 812
 813static int resp_inquiry(struct scsi_cmnd * scp, int target,
 814			struct sdebug_dev_info * devip)
 815{
 816	unsigned char pq_pdt;
 817	unsigned char * arr;
 818	unsigned char *cmd = (unsigned char *)scp->cmnd;
 819	int alloc_len, n, ret;
 
 820
 821	alloc_len = (cmd[3] << 8) + cmd[4];
 822	arr = kzalloc(SDEBUG_MAX_INQ_ARR_SZ, GFP_ATOMIC);
 823	if (! arr)
 824		return DID_REQUEUE << 16;
 825	if (devip->wlun)
 826		pq_pdt = 0x1e;	/* present, wlun */
 827	else if (scsi_debug_no_lun_0 && (0 == devip->lun))
 828		pq_pdt = 0x7f;	/* not present, no device type */
 
 
 829	else
 830		pq_pdt = (scsi_debug_ptype & 0x1f);
 831	arr[0] = pq_pdt;
 832	if (0x2 & cmd[1]) {  /* CMDDT bit set */
 833		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB,
 834			       	0);
 835		kfree(arr);
 836		return check_condition_result;
 837	} else if (0x1 & cmd[1]) {  /* EVPD bit set */
 838		int lu_id_num, port_group_id, target_dev_id, len;
 839		char lu_id_str[6];
 840		int host_no = devip->sdbg_host->shost->host_no;
 841		
 842		port_group_id = (((host_no + 1) & 0x7f) << 8) +
 843		    (devip->channel & 0x7f);
 844		if (0 == scsi_debug_vpd_use_hostno)
 845			host_no = 0;
 846		lu_id_num = devip->wlun ? -1 : (((host_no + 1) * 2000) +
 847			    (devip->target * 1000) + devip->lun);
 848		target_dev_id = ((host_no + 1) * 2000) +
 849				 (devip->target * 1000) - 3;
 850		len = scnprintf(lu_id_str, 6, "%d", lu_id_num);
 851		if (0 == cmd[2]) { /* supported vital product data pages */
 852			arr[1] = cmd[2];	/*sanity */
 853			n = 4;
 854			arr[n++] = 0x0;   /* this page */
 855			arr[n++] = 0x80;  /* unit serial number */
 856			arr[n++] = 0x83;  /* device identification */
 857			arr[n++] = 0x84;  /* software interface ident. */
 858			arr[n++] = 0x85;  /* management network addresses */
 859			arr[n++] = 0x86;  /* extended inquiry */
 860			arr[n++] = 0x87;  /* mode page policy */
 861			arr[n++] = 0x88;  /* SCSI ports */
 862			arr[n++] = 0x89;  /* ATA information */
 863			arr[n++] = 0xb0;  /* Block limits (SBC) */
 864			arr[n++] = 0xb1;  /* Block characteristics (SBC) */
 865			if (scsi_debug_lbp()) /* Logical Block Prov. (SBC) */
 866				arr[n++] = 0xb2;
 
 867			arr[3] = n - 4;	  /* number of supported VPD pages */
 868		} else if (0x80 == cmd[2]) { /* unit serial number */
 869			arr[1] = cmd[2];	/*sanity */
 870			arr[3] = len;
 871			memcpy(&arr[4], lu_id_str, len);
 872		} else if (0x83 == cmd[2]) { /* device identification */
 873			arr[1] = cmd[2];	/*sanity */
 874			arr[3] = inquiry_evpd_83(&arr[4], port_group_id,
 875						 target_dev_id, lu_id_num,
 876						 lu_id_str, len);
 
 877		} else if (0x84 == cmd[2]) { /* Software interface ident. */
 878			arr[1] = cmd[2];	/*sanity */
 879			arr[3] = inquiry_evpd_84(&arr[4]);
 880		} else if (0x85 == cmd[2]) { /* Management network addresses */
 881			arr[1] = cmd[2];	/*sanity */
 882			arr[3] = inquiry_evpd_85(&arr[4]);
 883		} else if (0x86 == cmd[2]) { /* extended inquiry */
 884			arr[1] = cmd[2];	/*sanity */
 885			arr[3] = 0x3c;	/* number of following entries */
 886			if (scsi_debug_dif == SD_DIF_TYPE3_PROTECTION)
 887				arr[4] = 0x4;	/* SPT: GRD_CHK:1 */
 888			else if (scsi_debug_dif)
 889				arr[4] = 0x5;   /* SPT: GRD_CHK:1, REF_CHK:1 */
 890			else
 891				arr[4] = 0x0;   /* no protection stuff */
 892			arr[5] = 0x7;   /* head of q, ordered + simple q's */
 893		} else if (0x87 == cmd[2]) { /* mode page policy */
 894			arr[1] = cmd[2];	/*sanity */
 895			arr[3] = 0x8;	/* number of following entries */
 896			arr[4] = 0x2;	/* disconnect-reconnect mp */
 897			arr[6] = 0x80;	/* mlus, shared */
 898			arr[8] = 0x18;	 /* protocol specific lu */
 899			arr[10] = 0x82;	 /* mlus, per initiator port */
 900		} else if (0x88 == cmd[2]) { /* SCSI Ports */
 901			arr[1] = cmd[2];	/*sanity */
 902			arr[3] = inquiry_evpd_88(&arr[4], target_dev_id);
 903		} else if (0x89 == cmd[2]) { /* ATA information */
 904			arr[1] = cmd[2];        /*sanity */
 905			n = inquiry_evpd_89(&arr[4]);
 906			arr[2] = (n >> 8);
 907			arr[3] = (n & 0xff);
 908		} else if (0xb0 == cmd[2]) { /* Block limits (SBC) */
 909			arr[1] = cmd[2];        /*sanity */
 910			arr[3] = inquiry_evpd_b0(&arr[4]);
 911		} else if (0xb1 == cmd[2]) { /* Block characteristics (SBC) */
 912			arr[1] = cmd[2];        /*sanity */
 913			arr[3] = inquiry_evpd_b1(&arr[4]);
 914		} else if (0xb2 == cmd[2]) { /* Logical Block Prov. (SBC) */
 915			arr[1] = cmd[2];        /*sanity */
 916			arr[3] = inquiry_evpd_b2(&arr[4]);
 917		} else {
 918			/* Illegal request, invalid field in cdb */
 919			mk_sense_buffer(devip, ILLEGAL_REQUEST,
 920					INVALID_FIELD_IN_CDB, 0);
 921			kfree(arr);
 922			return check_condition_result;
 923		}
 924		len = min(((arr[2] << 8) + arr[3]) + 4, alloc_len);
 925		ret = fill_from_dev_buffer(scp, arr,
 926			    min(len, SDEBUG_MAX_INQ_ARR_SZ));
 927		kfree(arr);
 928		return ret;
 929	}
 930	/* drops through here for a standard inquiry */
 931	arr[1] = scsi_debug_removable ? 0x80 : 0;	/* Removable disk */
 932	arr[2] = scsi_debug_scsi_level;
 933	arr[3] = 2;    /* response_data_format==2 */
 934	arr[4] = SDEBUG_LONG_INQ_SZ - 5;
 935	arr[5] = scsi_debug_dif ? 1 : 0; /* PROTECT bit */
 936	if (0 == scsi_debug_vpd_use_hostno)
 937		arr[5] = 0x10; /* claim: implicit TGPS */
 938	arr[6] = 0x10; /* claim: MultiP */
 939	/* arr[6] |= 0x40; ... claim: EncServ (enclosure services) */
 940	arr[7] = 0xa; /* claim: LINKED + CMDQUE */
 941	memcpy(&arr[8], inq_vendor_id, 8);
 942	memcpy(&arr[16], inq_product_id, 16);
 943	memcpy(&arr[32], inq_product_rev, 4);
 944	/* version descriptors (2 bytes each) follow */
 945	arr[58] = 0x0; arr[59] = 0x77; /* SAM-3 ANSI */
 946	arr[60] = 0x3; arr[61] = 0x14;  /* SPC-3 ANSI */
 947	n = 62;
 948	if (scsi_debug_ptype == 0) {
 949		arr[n++] = 0x3; arr[n++] = 0x3d; /* SBC-2 ANSI */
 950	} else if (scsi_debug_ptype == 1) {
 951		arr[n++] = 0x3; arr[n++] = 0x60; /* SSC-2 no version */
 
 
 952	}
 953	arr[n++] = 0xc; arr[n++] = 0xf;  /* SAS-1.1 rev 10 */
 954	ret = fill_from_dev_buffer(scp, arr,
 955			    min(alloc_len, SDEBUG_LONG_INQ_SZ));
 956	kfree(arr);
 957	return ret;
 958}
 959
 
 
 
 960static int resp_requests(struct scsi_cmnd * scp,
 961			 struct sdebug_dev_info * devip)
 962{
 963	unsigned char * sbuff;
 964	unsigned char *cmd = (unsigned char *)scp->cmnd;
 965	unsigned char arr[SDEBUG_SENSE_LEN];
 966	int want_dsense;
 967	int len = 18;
 968
 969	memset(arr, 0, sizeof(arr));
 970	if (devip->reset == 1)
 971		mk_sense_buffer(devip, 0, NO_ADDITIONAL_SENSE, 0);
 972	want_dsense = !!(cmd[1] & 1) || scsi_debug_dsense;
 973	sbuff = devip->sense_buff;
 974	if ((iec_m_pg[2] & 0x4) && (6 == (iec_m_pg[3] & 0xf))) {
 975		if (want_dsense) {
 976			arr[0] = 0x72;
 977			arr[1] = 0x0;		/* NO_SENSE in sense_key */
 978			arr[2] = THRESHOLD_EXCEEDED;
 979			arr[3] = 0xff;		/* TEST set and MRIE==6 */
 
 980		} else {
 981			arr[0] = 0x70;
 982			arr[2] = 0x0;		/* NO_SENSE in sense_key */
 983			arr[7] = 0xa;   	/* 18 byte sense buffer */
 984			arr[12] = THRESHOLD_EXCEEDED;
 985			arr[13] = 0xff;		/* TEST set and MRIE==6 */
 986		}
 987	} else {
 988		memcpy(arr, sbuff, SDEBUG_SENSE_LEN);
 989		if ((cmd[1] & 1) && (! scsi_debug_dsense)) {
 990			/* DESC bit set and sense_buff in fixed format */
 991			memset(arr, 0, sizeof(arr));
 
 
 
 
 
 
 
 
 
 
 
 992			arr[0] = 0x72;
 993			arr[1] = sbuff[2];     /* sense key */
 994			arr[2] = sbuff[12];    /* asc */
 995			arr[3] = sbuff[13];    /* ascq */
 996			len = 8;
 
 
 
 
 
 
 
 997		}
 
 998	}
 999	mk_sense_buffer(devip, 0, NO_ADDITIONAL_SENSE, 0);
1000	return fill_from_dev_buffer(scp, arr, len);
1001}
1002
1003static int resp_start_stop(struct scsi_cmnd * scp,
1004			   struct sdebug_dev_info * devip)
1005{
1006	unsigned char *cmd = (unsigned char *)scp->cmnd;
1007	int power_cond, errsts, start;
1008
1009	if ((errsts = check_readiness(scp, 1, devip)))
1010		return errsts;
1011	power_cond = (cmd[4] & 0xf0) >> 4;
1012	if (power_cond) {
1013		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB,
1014			       	0);
1015		return check_condition_result;
1016	}
1017	start = cmd[4] & 1;
1018	if (start == devip->stopped)
1019		devip->stopped = !start;
1020	return 0;
1021}
1022
1023static sector_t get_sdebug_capacity(void)
1024{
1025	if (scsi_debug_virtual_gb > 0)
1026		return (sector_t)scsi_debug_virtual_gb *
1027			(1073741824 / scsi_debug_sector_size);
 
 
1028	else
1029		return sdebug_store_sectors;
1030}
1031
1032#define SDEBUG_READCAP_ARR_SZ 8
1033static int resp_readcap(struct scsi_cmnd * scp,
1034			struct sdebug_dev_info * devip)
1035{
1036	unsigned char arr[SDEBUG_READCAP_ARR_SZ];
1037	unsigned int capac;
1038	int errsts;
1039
1040	if ((errsts = check_readiness(scp, 1, devip)))
1041		return errsts;
1042	/* following just in case virtual_gb changed */
1043	sdebug_capacity = get_sdebug_capacity();
1044	memset(arr, 0, SDEBUG_READCAP_ARR_SZ);
1045	if (sdebug_capacity < 0xffffffff) {
1046		capac = (unsigned int)sdebug_capacity - 1;
1047		arr[0] = (capac >> 24);
1048		arr[1] = (capac >> 16) & 0xff;
1049		arr[2] = (capac >> 8) & 0xff;
1050		arr[3] = capac & 0xff;
1051	} else {
1052		arr[0] = 0xff;
1053		arr[1] = 0xff;
1054		arr[2] = 0xff;
1055		arr[3] = 0xff;
1056	}
1057	arr[6] = (scsi_debug_sector_size >> 8) & 0xff;
1058	arr[7] = scsi_debug_sector_size & 0xff;
1059	return fill_from_dev_buffer(scp, arr, SDEBUG_READCAP_ARR_SZ);
1060}
1061
1062#define SDEBUG_READCAP16_ARR_SZ 32
1063static int resp_readcap16(struct scsi_cmnd * scp,
1064			  struct sdebug_dev_info * devip)
1065{
1066	unsigned char *cmd = (unsigned char *)scp->cmnd;
1067	unsigned char arr[SDEBUG_READCAP16_ARR_SZ];
1068	unsigned long long capac;
1069	int errsts, k, alloc_len;
1070
1071	if ((errsts = check_readiness(scp, 1, devip)))
1072		return errsts;
1073	alloc_len = ((cmd[10] << 24) + (cmd[11] << 16) + (cmd[12] << 8)
1074		     + cmd[13]);
1075	/* following just in case virtual_gb changed */
1076	sdebug_capacity = get_sdebug_capacity();
1077	memset(arr, 0, SDEBUG_READCAP16_ARR_SZ);
1078	capac = sdebug_capacity - 1;
1079	for (k = 0; k < 8; ++k, capac >>= 8)
1080		arr[7 - k] = capac & 0xff;
1081	arr[8] = (scsi_debug_sector_size >> 24) & 0xff;
1082	arr[9] = (scsi_debug_sector_size >> 16) & 0xff;
1083	arr[10] = (scsi_debug_sector_size >> 8) & 0xff;
1084	arr[11] = scsi_debug_sector_size & 0xff;
1085	arr[13] = scsi_debug_physblk_exp & 0xf;
1086	arr[14] = (scsi_debug_lowest_aligned >> 8) & 0x3f;
1087
1088	if (scsi_debug_lbp()) {
1089		arr[14] |= 0x80; /* LBPME */
1090		if (scsi_debug_lbprz)
1091			arr[14] |= 0x40; /* LBPRZ */
 
 
 
 
1092	}
1093
1094	arr[15] = scsi_debug_lowest_aligned & 0xff;
1095
1096	if (scsi_debug_dif) {
1097		arr[12] = (scsi_debug_dif - 1) << 1; /* P_TYPE */
1098		arr[12] |= 1; /* PROT_EN */
1099	}
1100
1101	return fill_from_dev_buffer(scp, arr,
1102				    min(alloc_len, SDEBUG_READCAP16_ARR_SZ));
1103}
1104
1105#define SDEBUG_MAX_TGTPGS_ARR_SZ 1412
1106
1107static int resp_report_tgtpgs(struct scsi_cmnd * scp,
1108			      struct sdebug_dev_info * devip)
1109{
1110	unsigned char *cmd = (unsigned char *)scp->cmnd;
1111	unsigned char * arr;
1112	int host_no = devip->sdbg_host->shost->host_no;
1113	int n, ret, alen, rlen;
1114	int port_group_a, port_group_b, port_a, port_b;
1115
1116	alen = ((cmd[6] << 24) + (cmd[7] << 16) + (cmd[8] << 8)
1117		+ cmd[9]);
1118
1119	arr = kzalloc(SDEBUG_MAX_TGTPGS_ARR_SZ, GFP_ATOMIC);
1120	if (! arr)
1121		return DID_REQUEUE << 16;
1122	/*
1123	 * EVPD page 0x88 states we have two ports, one
1124	 * real and a fake port with no device connected.
1125	 * So we create two port groups with one port each
1126	 * and set the group with port B to unavailable.
1127	 */
1128	port_a = 0x1; /* relative port A */
1129	port_b = 0x2; /* relative port B */
1130	port_group_a = (((host_no + 1) & 0x7f) << 8) +
1131	    (devip->channel & 0x7f);
1132	port_group_b = (((host_no + 1) & 0x7f) << 8) +
1133	    (devip->channel & 0x7f) + 0x80;
1134
1135	/*
1136	 * The asymmetric access state is cycled according to the host_id.
1137	 */
1138	n = 4;
1139	if (0 == scsi_debug_vpd_use_hostno) {
1140	    arr[n++] = host_no % 3; /* Asymm access state */
1141	    arr[n++] = 0x0F; /* claim: all states are supported */
1142	} else {
1143	    arr[n++] = 0x0; /* Active/Optimized path */
1144	    arr[n++] = 0x01; /* claim: only support active/optimized paths */
1145	}
1146	arr[n++] = (port_group_a >> 8) & 0xff;
1147	arr[n++] = port_group_a & 0xff;
1148	arr[n++] = 0;    /* Reserved */
1149	arr[n++] = 0;    /* Status code */
1150	arr[n++] = 0;    /* Vendor unique */
1151	arr[n++] = 0x1;  /* One port per group */
1152	arr[n++] = 0;    /* Reserved */
1153	arr[n++] = 0;    /* Reserved */
1154	arr[n++] = (port_a >> 8) & 0xff;
1155	arr[n++] = port_a & 0xff;
1156	arr[n++] = 3;    /* Port unavailable */
1157	arr[n++] = 0x08; /* claim: only unavailalbe paths are supported */
1158	arr[n++] = (port_group_b >> 8) & 0xff;
1159	arr[n++] = port_group_b & 0xff;
1160	arr[n++] = 0;    /* Reserved */
1161	arr[n++] = 0;    /* Status code */
1162	arr[n++] = 0;    /* Vendor unique */
1163	arr[n++] = 0x1;  /* One port per group */
1164	arr[n++] = 0;    /* Reserved */
1165	arr[n++] = 0;    /* Reserved */
1166	arr[n++] = (port_b >> 8) & 0xff;
1167	arr[n++] = port_b & 0xff;
1168
1169	rlen = n - 4;
1170	arr[0] = (rlen >> 24) & 0xff;
1171	arr[1] = (rlen >> 16) & 0xff;
1172	arr[2] = (rlen >> 8) & 0xff;
1173	arr[3] = rlen & 0xff;
1174
1175	/*
1176	 * Return the smallest value of either
1177	 * - The allocated length
1178	 * - The constructed command length
1179	 * - The maximum array size
1180	 */
1181	rlen = min(alen,n);
1182	ret = fill_from_dev_buffer(scp, arr,
1183				   min(rlen, SDEBUG_MAX_TGTPGS_ARR_SZ));
1184	kfree(arr);
1185	return ret;
1186}
1187
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1188/* <<Following mode page info copied from ST318451LW>> */
1189
1190static int resp_err_recov_pg(unsigned char * p, int pcontrol, int target)
1191{	/* Read-Write Error Recovery page for mode_sense */
1192	unsigned char err_recov_pg[] = {0x1, 0xa, 0xc0, 11, 240, 0, 0, 0,
1193					5, 0, 0xff, 0xff};
1194
1195	memcpy(p, err_recov_pg, sizeof(err_recov_pg));
1196	if (1 == pcontrol)
1197		memset(p + 2, 0, sizeof(err_recov_pg) - 2);
1198	return sizeof(err_recov_pg);
1199}
1200
1201static int resp_disconnect_pg(unsigned char * p, int pcontrol, int target)
1202{ 	/* Disconnect-Reconnect page for mode_sense */
1203	unsigned char disconnect_pg[] = {0x2, 0xe, 128, 128, 0, 10, 0, 0,
1204					 0, 0, 0, 0, 0, 0, 0, 0};
1205
1206	memcpy(p, disconnect_pg, sizeof(disconnect_pg));
1207	if (1 == pcontrol)
1208		memset(p + 2, 0, sizeof(disconnect_pg) - 2);
1209	return sizeof(disconnect_pg);
1210}
1211
1212static int resp_format_pg(unsigned char * p, int pcontrol, int target)
1213{       /* Format device page for mode_sense */
1214	unsigned char format_pg[] = {0x3, 0x16, 0, 0, 0, 0, 0, 0,
1215				     0, 0, 0, 0, 0, 0, 0, 0,
1216				     0, 0, 0, 0, 0x40, 0, 0, 0};
1217
1218	memcpy(p, format_pg, sizeof(format_pg));
1219	p[10] = (sdebug_sectors_per >> 8) & 0xff;
1220	p[11] = sdebug_sectors_per & 0xff;
1221	p[12] = (scsi_debug_sector_size >> 8) & 0xff;
1222	p[13] = scsi_debug_sector_size & 0xff;
1223	if (scsi_debug_removable)
1224		p[20] |= 0x20; /* should agree with INQUIRY */
1225	if (1 == pcontrol)
1226		memset(p + 2, 0, sizeof(format_pg) - 2);
1227	return sizeof(format_pg);
1228}
1229
 
 
 
 
1230static int resp_caching_pg(unsigned char * p, int pcontrol, int target)
1231{ 	/* Caching page for mode_sense */
1232	unsigned char caching_pg[] = {0x8, 18, 0x14, 0, 0xff, 0xff, 0, 0,
 
 
1233		0xff, 0xff, 0xff, 0xff, 0x80, 0x14, 0, 0,     0, 0, 0, 0};
1234
 
 
1235	memcpy(p, caching_pg, sizeof(caching_pg));
1236	if (1 == pcontrol)
1237		memset(p + 2, 0, sizeof(caching_pg) - 2);
 
 
1238	return sizeof(caching_pg);
1239}
1240
 
 
 
1241static int resp_ctrl_m_pg(unsigned char * p, int pcontrol, int target)
1242{ 	/* Control mode page for mode_sense */
1243	unsigned char ch_ctrl_m_pg[] = {/* 0xa, 10, */ 0x6, 0, 0, 0, 0, 0,
1244				        0, 0, 0, 0};
1245	unsigned char d_ctrl_m_pg[] = {0xa, 10, 2, 0, 0, 0, 0, 0,
1246				     0, 0, 0x2, 0x4b};
1247
1248	if (scsi_debug_dsense)
1249		ctrl_m_pg[2] |= 0x4;
1250	else
1251		ctrl_m_pg[2] &= ~0x4;
1252
1253	if (scsi_debug_ato)
1254		ctrl_m_pg[5] |= 0x80; /* ATO=1 */
1255
1256	memcpy(p, ctrl_m_pg, sizeof(ctrl_m_pg));
1257	if (1 == pcontrol)
1258		memcpy(p + 2, ch_ctrl_m_pg, sizeof(ch_ctrl_m_pg));
1259	else if (2 == pcontrol)
1260		memcpy(p, d_ctrl_m_pg, sizeof(d_ctrl_m_pg));
1261	return sizeof(ctrl_m_pg);
1262}
1263
1264
1265static int resp_iec_m_pg(unsigned char * p, int pcontrol, int target)
1266{	/* Informational Exceptions control mode page for mode_sense */
1267	unsigned char ch_iec_m_pg[] = {/* 0x1c, 0xa, */ 0x4, 0xf, 0, 0, 0, 0,
1268				       0, 0, 0x0, 0x0};
1269	unsigned char d_iec_m_pg[] = {0x1c, 0xa, 0x08, 0, 0, 0, 0, 0,
1270				      0, 0, 0x0, 0x0};
1271
1272	memcpy(p, iec_m_pg, sizeof(iec_m_pg));
1273	if (1 == pcontrol)
1274		memcpy(p + 2, ch_iec_m_pg, sizeof(ch_iec_m_pg));
1275	else if (2 == pcontrol)
1276		memcpy(p, d_iec_m_pg, sizeof(d_iec_m_pg));
1277	return sizeof(iec_m_pg);
1278}
1279
1280static int resp_sas_sf_m_pg(unsigned char * p, int pcontrol, int target)
1281{	/* SAS SSP mode page - short format for mode_sense */
1282	unsigned char sas_sf_m_pg[] = {0x19, 0x6,
1283		0x6, 0x0, 0x7, 0xd0, 0x0, 0x0};
1284
1285	memcpy(p, sas_sf_m_pg, sizeof(sas_sf_m_pg));
1286	if (1 == pcontrol)
1287		memset(p + 2, 0, sizeof(sas_sf_m_pg) - 2);
1288	return sizeof(sas_sf_m_pg);
1289}
1290
1291
1292static int resp_sas_pcd_m_spg(unsigned char * p, int pcontrol, int target,
1293			      int target_dev_id)
1294{	/* SAS phy control and discover mode page for mode_sense */
1295	unsigned char sas_pcd_m_pg[] = {0x59, 0x1, 0, 0x64, 0, 0x6, 0, 2,
1296		    0, 0, 0, 0, 0x10, 0x9, 0x8, 0x0,
1297		    0x52, 0x22, 0x22, 0x20, 0x0, 0x0, 0x0, 0x0,
1298		    0x51, 0x11, 0x11, 0x10, 0x0, 0x0, 0x0, 0x1,
1299		    0x2, 0, 0, 0, 0, 0, 0, 0,
1300		    0x88, 0x99, 0, 0, 0, 0, 0, 0,
1301		    0, 0, 0, 0, 0, 0, 0, 0,
1302		    0, 1, 0, 0, 0x10, 0x9, 0x8, 0x0,
1303		    0x52, 0x22, 0x22, 0x20, 0x0, 0x0, 0x0, 0x0,
1304		    0x51, 0x11, 0x11, 0x10, 0x0, 0x0, 0x0, 0x1,
1305		    0x3, 0, 0, 0, 0, 0, 0, 0,
1306		    0x88, 0x99, 0, 0, 0, 0, 0, 0,
1307		    0, 0, 0, 0, 0, 0, 0, 0,
1308		};
1309	int port_a, port_b;
1310
 
 
 
 
1311	port_a = target_dev_id + 1;
1312	port_b = port_a + 1;
1313	memcpy(p, sas_pcd_m_pg, sizeof(sas_pcd_m_pg));
1314	p[20] = (port_a >> 24);
1315	p[21] = (port_a >> 16) & 0xff;
1316	p[22] = (port_a >> 8) & 0xff;
1317	p[23] = port_a & 0xff;
1318	p[48 + 20] = (port_b >> 24);
1319	p[48 + 21] = (port_b >> 16) & 0xff;
1320	p[48 + 22] = (port_b >> 8) & 0xff;
1321	p[48 + 23] = port_b & 0xff;
1322	if (1 == pcontrol)
1323		memset(p + 4, 0, sizeof(sas_pcd_m_pg) - 4);
1324	return sizeof(sas_pcd_m_pg);
1325}
1326
1327static int resp_sas_sha_m_spg(unsigned char * p, int pcontrol)
1328{	/* SAS SSP shared protocol specific port mode subpage */
1329	unsigned char sas_sha_m_pg[] = {0x59, 0x2, 0, 0xc, 0, 0x6, 0x10, 0,
1330		    0, 0, 0, 0, 0, 0, 0, 0,
1331		};
1332
1333	memcpy(p, sas_sha_m_pg, sizeof(sas_sha_m_pg));
1334	if (1 == pcontrol)
1335		memset(p + 4, 0, sizeof(sas_sha_m_pg) - 4);
1336	return sizeof(sas_sha_m_pg);
1337}
1338
1339#define SDEBUG_MAX_MSENSE_SZ 256
1340
1341static int resp_mode_sense(struct scsi_cmnd * scp, int target,
1342			   struct sdebug_dev_info * devip)
1343{
1344	unsigned char dbd, llbaa;
1345	int pcontrol, pcode, subpcode, bd_len;
1346	unsigned char dev_spec;
1347	int k, alloc_len, msense_6, offset, len, errsts, target_dev_id;
 
1348	unsigned char * ap;
1349	unsigned char arr[SDEBUG_MAX_MSENSE_SZ];
1350	unsigned char *cmd = (unsigned char *)scp->cmnd;
 
1351
1352	if ((errsts = check_readiness(scp, 1, devip)))
1353		return errsts;
1354	dbd = !!(cmd[1] & 0x8);
1355	pcontrol = (cmd[2] & 0xc0) >> 6;
1356	pcode = cmd[2] & 0x3f;
1357	subpcode = cmd[3];
1358	msense_6 = (MODE_SENSE == cmd[0]);
1359	llbaa = msense_6 ? 0 : !!(cmd[1] & 0x10);
1360	if ((0 == scsi_debug_ptype) && (0 == dbd))
 
1361		bd_len = llbaa ? 16 : 8;
1362	else
1363		bd_len = 0;
1364	alloc_len = msense_6 ? cmd[4] : ((cmd[7] << 8) | cmd[8]);
1365	memset(arr, 0, SDEBUG_MAX_MSENSE_SZ);
1366	if (0x3 == pcontrol) {  /* Saving values not supported */
1367		mk_sense_buffer(devip, ILLEGAL_REQUEST, SAVING_PARAMS_UNSUP,
1368			       	0);
1369		return check_condition_result;
1370	}
1371	target_dev_id = ((devip->sdbg_host->shost->host_no + 1) * 2000) +
1372			(devip->target * 1000) - 3;
1373	/* set DPOFUA bit for disks */
1374	if (0 == scsi_debug_ptype)
1375		dev_spec = (DEV_READONLY(target) ? 0x80 : 0x0) | 0x10;
1376	else
1377		dev_spec = 0x0;
1378	if (msense_6) {
1379		arr[2] = dev_spec;
1380		arr[3] = bd_len;
1381		offset = 4;
1382	} else {
1383		arr[3] = dev_spec;
1384		if (16 == bd_len)
1385			arr[4] = 0x1;	/* set LONGLBA bit */
1386		arr[7] = bd_len;	/* assume 255 or less */
1387		offset = 8;
1388	}
1389	ap = arr + offset;
1390	if ((bd_len > 0) && (!sdebug_capacity))
1391		sdebug_capacity = get_sdebug_capacity();
1392
1393	if (8 == bd_len) {
1394		if (sdebug_capacity > 0xfffffffe) {
1395			ap[0] = 0xff;
1396			ap[1] = 0xff;
1397			ap[2] = 0xff;
1398			ap[3] = 0xff;
1399		} else {
1400			ap[0] = (sdebug_capacity >> 24) & 0xff;
1401			ap[1] = (sdebug_capacity >> 16) & 0xff;
1402			ap[2] = (sdebug_capacity >> 8) & 0xff;
1403			ap[3] = sdebug_capacity & 0xff;
1404		}
1405		ap[6] = (scsi_debug_sector_size >> 8) & 0xff;
1406		ap[7] = scsi_debug_sector_size & 0xff;
1407		offset += bd_len;
1408		ap = arr + offset;
1409	} else if (16 == bd_len) {
1410		unsigned long long capac = sdebug_capacity;
1411
1412        	for (k = 0; k < 8; ++k, capac >>= 8)
1413                	ap[7 - k] = capac & 0xff;
1414		ap[12] = (scsi_debug_sector_size >> 24) & 0xff;
1415		ap[13] = (scsi_debug_sector_size >> 16) & 0xff;
1416		ap[14] = (scsi_debug_sector_size >> 8) & 0xff;
1417		ap[15] = scsi_debug_sector_size & 0xff;
1418		offset += bd_len;
1419		ap = arr + offset;
1420	}
1421
1422	if ((subpcode > 0x0) && (subpcode < 0xff) && (0x19 != pcode)) {
1423		/* TODO: Control Extension page */
1424		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB,
1425			       	0);
1426		return check_condition_result;
1427	}
 
 
1428	switch (pcode) {
1429	case 0x1:	/* Read-Write error recovery page, direct access */
1430		len = resp_err_recov_pg(ap, pcontrol, target);
1431		offset += len;
1432		break;
1433	case 0x2:	/* Disconnect-Reconnect page, all devices */
1434		len = resp_disconnect_pg(ap, pcontrol, target);
1435		offset += len;
1436		break;
1437        case 0x3:       /* Format device page, direct access */
1438                len = resp_format_pg(ap, pcontrol, target);
1439                offset += len;
 
 
 
1440                break;
1441	case 0x8:	/* Caching page, direct access */
1442		len = resp_caching_pg(ap, pcontrol, target);
1443		offset += len;
 
 
 
1444		break;
1445	case 0xa:	/* Control Mode page, all devices */
1446		len = resp_ctrl_m_pg(ap, pcontrol, target);
1447		offset += len;
1448		break;
1449	case 0x19:	/* if spc==1 then sas phy, control+discover */
1450		if ((subpcode > 0x2) && (subpcode < 0xff)) {
1451		        mk_sense_buffer(devip, ILLEGAL_REQUEST,
1452					INVALID_FIELD_IN_CDB, 0);
1453			return check_condition_result;
1454	        }
1455		len = 0;
1456		if ((0x0 == subpcode) || (0xff == subpcode))
1457			len += resp_sas_sf_m_pg(ap + len, pcontrol, target);
1458		if ((0x1 == subpcode) || (0xff == subpcode))
1459			len += resp_sas_pcd_m_spg(ap + len, pcontrol, target,
1460						  target_dev_id);
1461		if ((0x2 == subpcode) || (0xff == subpcode))
1462			len += resp_sas_sha_m_spg(ap + len, pcontrol);
1463		offset += len;
1464		break;
1465	case 0x1c:	/* Informational Exceptions Mode page, all devices */
1466		len = resp_iec_m_pg(ap, pcontrol, target);
1467		offset += len;
1468		break;
1469	case 0x3f:	/* Read all Mode pages */
1470		if ((0 == subpcode) || (0xff == subpcode)) {
1471			len = resp_err_recov_pg(ap, pcontrol, target);
1472			len += resp_disconnect_pg(ap + len, pcontrol, target);
1473			len += resp_format_pg(ap + len, pcontrol, target);
1474			len += resp_caching_pg(ap + len, pcontrol, target);
 
 
 
 
1475			len += resp_ctrl_m_pg(ap + len, pcontrol, target);
1476			len += resp_sas_sf_m_pg(ap + len, pcontrol, target);
1477			if (0xff == subpcode) {
1478				len += resp_sas_pcd_m_spg(ap + len, pcontrol,
1479						  target, target_dev_id);
1480				len += resp_sas_sha_m_spg(ap + len, pcontrol);
1481			}
1482			len += resp_iec_m_pg(ap + len, pcontrol, target);
 
1483		} else {
1484			mk_sense_buffer(devip, ILLEGAL_REQUEST,
1485					INVALID_FIELD_IN_CDB, 0);
1486			return check_condition_result;
1487                }
1488		offset += len;
1489		break;
1490	default:
1491		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB,
1492			       	0);
 
 
 
1493		return check_condition_result;
1494	}
1495	if (msense_6)
1496		arr[0] = offset - 1;
1497	else {
1498		arr[0] = ((offset - 2) >> 8) & 0xff;
1499		arr[1] = (offset - 2) & 0xff;
1500	}
1501	return fill_from_dev_buffer(scp, arr, min(alloc_len, offset));
1502}
1503
1504#define SDEBUG_MAX_MSELECT_SZ 512
1505
1506static int resp_mode_select(struct scsi_cmnd * scp, int mselect6,
1507			    struct sdebug_dev_info * devip)
1508{
1509	int pf, sp, ps, md_len, bd_len, off, spf, pg_len;
1510	int param_len, res, errsts, mpage;
1511	unsigned char arr[SDEBUG_MAX_MSELECT_SZ];
1512	unsigned char *cmd = (unsigned char *)scp->cmnd;
 
1513
1514	if ((errsts = check_readiness(scp, 1, devip)))
1515		return errsts;
1516	memset(arr, 0, sizeof(arr));
1517	pf = cmd[1] & 0x10;
1518	sp = cmd[1] & 0x1;
1519	param_len = mselect6 ? cmd[4] : ((cmd[7] << 8) + cmd[8]);
1520	if ((0 == pf) || sp || (param_len > SDEBUG_MAX_MSELECT_SZ)) {
1521		mk_sense_buffer(devip, ILLEGAL_REQUEST,
1522				INVALID_FIELD_IN_CDB, 0);
1523		return check_condition_result;
1524	}
1525        res = fetch_to_dev_buffer(scp, arr, param_len);
1526        if (-1 == res)
1527                return (DID_ERROR << 16);
1528        else if ((res < param_len) &&
1529                 (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
1530                printk(KERN_INFO "scsi_debug: mode_select: cdb indicated=%d, "
1531                       " IO sent=%d bytes\n", param_len, res);
1532	md_len = mselect6 ? (arr[0] + 1) : ((arr[0] << 8) + arr[1] + 2);
1533	bd_len = mselect6 ? arr[3] : ((arr[6] << 8) + arr[7]);
1534	if (md_len > 2) {
1535		mk_sense_buffer(devip, ILLEGAL_REQUEST,
1536				INVALID_FIELD_IN_PARAM_LIST, 0);
1537		return check_condition_result;
1538	}
1539	off = bd_len + (mselect6 ? 4 : 8);
1540	mpage = arr[off] & 0x3f;
1541	ps = !!(arr[off] & 0x80);
1542	if (ps) {
1543		mk_sense_buffer(devip, ILLEGAL_REQUEST,
1544				INVALID_FIELD_IN_PARAM_LIST, 0);
1545		return check_condition_result;
1546	}
1547	spf = !!(arr[off] & 0x40);
1548	pg_len = spf ? ((arr[off + 2] << 8) + arr[off + 3] + 4) :
1549		       (arr[off + 1] + 2);
1550	if ((pg_len + off) > param_len) {
1551		mk_sense_buffer(devip, ILLEGAL_REQUEST,
1552				PARAMETER_LIST_LENGTH_ERR, 0);
1553		return check_condition_result;
1554	}
1555	switch (mpage) {
 
 
 
 
 
 
 
1556	case 0xa:      /* Control Mode page */
1557		if (ctrl_m_pg[1] == arr[off + 1]) {
1558			memcpy(ctrl_m_pg + 2, arr + off + 2,
1559			       sizeof(ctrl_m_pg) - 2);
1560			scsi_debug_dsense = !!(ctrl_m_pg[2] & 0x4);
1561			return 0;
1562		}
1563		break;
1564	case 0x1c:      /* Informational Exceptions Mode page */
1565		if (iec_m_pg[1] == arr[off + 1]) {
1566			memcpy(iec_m_pg + 2, arr + off + 2,
1567			       sizeof(iec_m_pg) - 2);
1568			return 0;
1569		}
1570		break;
1571	default:
1572		break;
1573	}
1574	mk_sense_buffer(devip, ILLEGAL_REQUEST,
1575			INVALID_FIELD_IN_PARAM_LIST, 0);
1576	return check_condition_result;
 
 
 
1577}
1578
1579static int resp_temp_l_pg(unsigned char * arr)
1580{
1581	unsigned char temp_l_pg[] = {0x0, 0x0, 0x3, 0x2, 0x0, 38,
1582				     0x0, 0x1, 0x3, 0x2, 0x0, 65,
1583		};
1584
1585        memcpy(arr, temp_l_pg, sizeof(temp_l_pg));
1586        return sizeof(temp_l_pg);
1587}
1588
1589static int resp_ie_l_pg(unsigned char * arr)
1590{
1591	unsigned char ie_l_pg[] = {0x0, 0x0, 0x3, 0x3, 0x0, 0x0, 38,
1592		};
1593
1594        memcpy(arr, ie_l_pg, sizeof(ie_l_pg));
1595	if (iec_m_pg[2] & 0x4) {	/* TEST bit set */
1596		arr[4] = THRESHOLD_EXCEEDED;
1597		arr[5] = 0xff;
1598	}
1599        return sizeof(ie_l_pg);
1600}
1601
1602#define SDEBUG_MAX_LSENSE_SZ 512
1603
1604static int resp_log_sense(struct scsi_cmnd * scp,
1605                          struct sdebug_dev_info * devip)
1606{
1607	int ppc, sp, pcontrol, pcode, subpcode, alloc_len, errsts, len, n;
1608	unsigned char arr[SDEBUG_MAX_LSENSE_SZ];
1609	unsigned char *cmd = (unsigned char *)scp->cmnd;
1610
1611	if ((errsts = check_readiness(scp, 1, devip)))
1612		return errsts;
1613	memset(arr, 0, sizeof(arr));
1614	ppc = cmd[1] & 0x2;
1615	sp = cmd[1] & 0x1;
1616	if (ppc || sp) {
1617		mk_sense_buffer(devip, ILLEGAL_REQUEST,
1618				INVALID_FIELD_IN_CDB, 0);
1619		return check_condition_result;
1620	}
1621	pcontrol = (cmd[2] & 0xc0) >> 6;
1622	pcode = cmd[2] & 0x3f;
1623	subpcode = cmd[3] & 0xff;
1624	alloc_len = (cmd[7] << 8) + cmd[8];
1625	arr[0] = pcode;
1626	if (0 == subpcode) {
1627		switch (pcode) {
1628		case 0x0:	/* Supported log pages log page */
1629			n = 4;
1630			arr[n++] = 0x0;		/* this page */
1631			arr[n++] = 0xd;		/* Temperature */
1632			arr[n++] = 0x2f;	/* Informational exceptions */
1633			arr[3] = n - 4;
1634			break;
1635		case 0xd:	/* Temperature log page */
1636			arr[3] = resp_temp_l_pg(arr + 4);
1637			break;
1638		case 0x2f:	/* Informational exceptions log page */
1639			arr[3] = resp_ie_l_pg(arr + 4);
1640			break;
1641		default:
1642			mk_sense_buffer(devip, ILLEGAL_REQUEST,
1643					INVALID_FIELD_IN_CDB, 0);
1644			return check_condition_result;
1645		}
1646	} else if (0xff == subpcode) {
1647		arr[0] |= 0x40;
1648		arr[1] = subpcode;
1649		switch (pcode) {
1650		case 0x0:	/* Supported log pages and subpages log page */
1651			n = 4;
1652			arr[n++] = 0x0;
1653			arr[n++] = 0x0;		/* 0,0 page */
1654			arr[n++] = 0x0;
1655			arr[n++] = 0xff;	/* this page */
1656			arr[n++] = 0xd;
1657			arr[n++] = 0x0;		/* Temperature */
1658			arr[n++] = 0x2f;
1659			arr[n++] = 0x0;	/* Informational exceptions */
1660			arr[3] = n - 4;
1661			break;
1662		case 0xd:	/* Temperature subpages */
1663			n = 4;
1664			arr[n++] = 0xd;
1665			arr[n++] = 0x0;		/* Temperature */
1666			arr[3] = n - 4;
1667			break;
1668		case 0x2f:	/* Informational exceptions subpages */
1669			n = 4;
1670			arr[n++] = 0x2f;
1671			arr[n++] = 0x0;		/* Informational exceptions */
1672			arr[3] = n - 4;
1673			break;
1674		default:
1675			mk_sense_buffer(devip, ILLEGAL_REQUEST,
1676					INVALID_FIELD_IN_CDB, 0);
1677			return check_condition_result;
1678		}
1679	} else {
1680		mk_sense_buffer(devip, ILLEGAL_REQUEST,
1681				INVALID_FIELD_IN_CDB, 0);
1682		return check_condition_result;
1683	}
1684	len = min(((arr[2] << 8) + arr[3]) + 4, alloc_len);
1685	return fill_from_dev_buffer(scp, arr,
1686		    min(len, SDEBUG_MAX_INQ_ARR_SZ));
1687}
1688
1689static int check_device_access_params(struct sdebug_dev_info *devi,
1690				      unsigned long long lba, unsigned int num)
1691{
1692	if (lba + num > sdebug_capacity) {
1693		mk_sense_buffer(devi, ILLEGAL_REQUEST, ADDR_OUT_OF_RANGE, 0);
1694		return check_condition_result;
1695	}
1696	/* transfer length excessive (tie in to block limits VPD page) */
1697	if (num > sdebug_store_sectors) {
1698		mk_sense_buffer(devi, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
 
1699		return check_condition_result;
1700	}
1701	return 0;
1702}
1703
1704/* Returns number of bytes copied or -1 if error. */
1705static int do_device_access(struct scsi_cmnd *scmd,
1706			    struct sdebug_dev_info *devi,
1707			    unsigned long long lba, unsigned int num, int write)
1708{
1709	int ret;
1710	unsigned long long block, rest = 0;
1711	struct scsi_data_buffer *sdb;
1712	enum dma_data_direction dir;
1713	size_t (*func)(struct scatterlist *, unsigned int, void *, size_t,
1714		       off_t);
1715
1716	if (write) {
1717		sdb = scsi_out(scmd);
1718		dir = DMA_TO_DEVICE;
1719		func = sg_pcopy_to_buffer;
1720	} else {
1721		sdb = scsi_in(scmd);
1722		dir = DMA_FROM_DEVICE;
1723		func = sg_pcopy_from_buffer;
1724	}
1725
1726	if (!sdb->length)
1727		return 0;
1728	if (!(scsi_bidi_cmnd(scmd) || scmd->sc_data_direction == dir))
1729		return -1;
1730
1731	block = do_div(lba, sdebug_store_sectors);
1732	if (block + num > sdebug_store_sectors)
1733		rest = block + num - sdebug_store_sectors;
1734
1735	ret = func(sdb->table.sgl, sdb->table.nents,
1736		   fake_storep + (block * scsi_debug_sector_size),
1737		   (num - rest) * scsi_debug_sector_size, 0);
1738	if (ret != (num - rest) * scsi_debug_sector_size)
1739		return ret;
1740
1741	if (rest) {
1742		ret += func(sdb->table.sgl, sdb->table.nents,
1743			    fake_storep, rest * scsi_debug_sector_size,
1744			    (num - rest) * scsi_debug_sector_size);
1745	}
1746
1747	return ret;
1748}
1749
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1750static __be16 dif_compute_csum(const void *buf, int len)
1751{
1752	__be16 csum;
1753
1754	if (scsi_debug_guard)
1755		csum = (__force __be16)ip_compute_csum(buf, len);
1756	else
1757		csum = cpu_to_be16(crc_t10dif(buf, len));
1758
1759	return csum;
1760}
1761
1762static int dif_verify(struct sd_dif_tuple *sdt, const void *data,
1763		      sector_t sector, u32 ei_lba)
1764{
1765	__be16 csum = dif_compute_csum(data, scsi_debug_sector_size);
1766
1767	if (sdt->guard_tag != csum) {
1768		pr_err("%s: GUARD check failed on sector %lu rcvd 0x%04x, data 0x%04x\n",
1769			__func__,
1770			(unsigned long)sector,
1771			be16_to_cpu(sdt->guard_tag),
1772			be16_to_cpu(csum));
1773		return 0x01;
1774	}
1775	if (scsi_debug_dif == SD_DIF_TYPE1_PROTECTION &&
1776	    be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) {
1777		pr_err("%s: REF check failed on sector %lu\n",
1778			__func__, (unsigned long)sector);
1779		return 0x03;
1780	}
1781	if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
1782	    be32_to_cpu(sdt->ref_tag) != ei_lba) {
1783		pr_err("%s: REF check failed on sector %lu\n",
1784			__func__, (unsigned long)sector);
1785		return 0x03;
1786	}
1787	return 0;
1788}
1789
1790static void dif_copy_prot(struct scsi_cmnd *SCpnt, sector_t sector,
1791			  unsigned int sectors, bool read)
1792{
1793	size_t resid;
1794	void *paddr;
1795	const void *dif_store_end = dif_storep + sdebug_store_sectors;
1796	struct sg_mapping_iter miter;
1797
1798	/* Bytes of protection data to copy into sgl */
1799	resid = sectors * sizeof(*dif_storep);
1800
1801	sg_miter_start(&miter, scsi_prot_sglist(SCpnt),
1802			scsi_prot_sg_count(SCpnt), SG_MITER_ATOMIC |
1803			(read ? SG_MITER_TO_SG : SG_MITER_FROM_SG));
1804
1805	while (sg_miter_next(&miter) && resid > 0) {
1806		size_t len = min(miter.length, resid);
1807		void *start = dif_store(sector);
1808		size_t rest = 0;
1809
1810		if (dif_store_end < start + len)
1811			rest = start + len - dif_store_end;
1812
1813		paddr = miter.addr;
1814
1815		if (read)
1816			memcpy(paddr, start, len - rest);
1817		else
1818			memcpy(start, paddr, len - rest);
1819
1820		if (rest) {
1821			if (read)
1822				memcpy(paddr + len - rest, dif_storep, rest);
1823			else
1824				memcpy(dif_storep, paddr + len - rest, rest);
1825		}
1826
1827		sector += len / sizeof(*dif_storep);
1828		resid -= len;
1829	}
1830	sg_miter_stop(&miter);
1831}
1832
1833static int prot_verify_read(struct scsi_cmnd *SCpnt, sector_t start_sec,
1834			    unsigned int sectors, u32 ei_lba)
1835{
1836	unsigned int i;
1837	struct sd_dif_tuple *sdt;
1838	sector_t sector;
1839
1840	for (i = 0; i < sectors; i++, ei_lba++) {
1841		int ret;
1842
1843		sector = start_sec + i;
1844		sdt = dif_store(sector);
1845
1846		if (sdt->app_tag == cpu_to_be16(0xffff))
1847			continue;
1848
1849		ret = dif_verify(sdt, fake_store(sector), sector, ei_lba);
1850		if (ret) {
1851			dif_errors++;
1852			return ret;
1853		}
1854	}
1855
1856	dif_copy_prot(SCpnt, start_sec, sectors, true);
1857	dix_reads++;
1858
1859	return 0;
1860}
1861
1862static int resp_read(struct scsi_cmnd *SCpnt, unsigned long long lba,
1863		     unsigned int num, struct sdebug_dev_info *devip,
1864		     u32 ei_lba)
1865{
 
 
 
 
 
1866	unsigned long iflags;
1867	int ret;
 
1868
1869	ret = check_device_access_params(devip, lba, num);
1870	if (ret)
1871		return ret;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1872
1873	if ((SCSI_DEBUG_OPT_MEDIUM_ERR & scsi_debug_opts) &&
1874	    (lba <= (OPT_MEDIUM_ERR_ADDR + OPT_MEDIUM_ERR_NUM - 1)) &&
1875	    ((lba + num) > OPT_MEDIUM_ERR_ADDR)) {
1876		/* claim unrecoverable read error */
1877		mk_sense_buffer(devip, MEDIUM_ERROR, UNRECOVERED_READ_ERR, 0);
1878		/* set info field and valid bit for fixed descriptor */
1879		if (0x70 == (devip->sense_buff[0] & 0x7f)) {
1880			devip->sense_buff[0] |= 0x80;	/* Valid bit */
1881			ret = (lba < OPT_MEDIUM_ERR_ADDR)
1882			      ? OPT_MEDIUM_ERR_ADDR : (int)lba;
1883			devip->sense_buff[3] = (ret >> 24) & 0xff;
1884			devip->sense_buff[4] = (ret >> 16) & 0xff;
1885			devip->sense_buff[5] = (ret >> 8) & 0xff;
1886			devip->sense_buff[6] = ret & 0xff;
1887		}
1888	        scsi_set_resid(SCpnt, scsi_bufflen(SCpnt));
1889		return check_condition_result;
1890	}
1891
1892	read_lock_irqsave(&atomic_rw, iflags);
1893
1894	/* DIX + T10 DIF */
1895	if (scsi_debug_dix && scsi_prot_sg_count(SCpnt)) {
1896		int prot_ret = prot_verify_read(SCpnt, lba, num, ei_lba);
1897
1898		if (prot_ret) {
1899			read_unlock_irqrestore(&atomic_rw, iflags);
1900			mk_sense_buffer(devip, ABORTED_COMMAND, 0x10, prot_ret);
1901			return illegal_condition_result;
1902		}
1903	}
1904
1905	ret = do_device_access(SCpnt, devip, lba, num, 0);
1906	read_unlock_irqrestore(&atomic_rw, iflags);
1907	if (ret == -1)
1908		return DID_ERROR << 16;
1909
1910	scsi_in(SCpnt)->resid = scsi_bufflen(SCpnt) - ret;
1911
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1912	return 0;
1913}
1914
1915void dump_sector(unsigned char *buf, int len)
1916{
1917	int i, j;
1918
1919	printk(KERN_ERR ">>> Sector Dump <<<\n");
1920
 
1921	for (i = 0 ; i < len ; i += 16) {
1922		printk(KERN_ERR "%04d: ", i);
1923
1924		for (j = 0 ; j < 16 ; j++) {
1925			unsigned char c = buf[i+j];
 
1926			if (c >= 0x20 && c < 0x7e)
1927				printk(" %c ", buf[i+j]);
 
1928			else
1929				printk("%02x ", buf[i+j]);
 
1930		}
1931
1932		printk("\n");
1933	}
1934}
1935
1936static int prot_verify_write(struct scsi_cmnd *SCpnt, sector_t start_sec,
1937			     unsigned int sectors, u32 ei_lba)
1938{
1939	int ret;
1940	struct sd_dif_tuple *sdt;
1941	void *daddr;
1942	sector_t sector = start_sec;
1943	int ppage_offset;
1944	int dpage_offset;
1945	struct sg_mapping_iter diter;
1946	struct sg_mapping_iter piter;
1947
1948	BUG_ON(scsi_sg_count(SCpnt) == 0);
1949	BUG_ON(scsi_prot_sg_count(SCpnt) == 0);
1950
1951	sg_miter_start(&piter, scsi_prot_sglist(SCpnt),
1952			scsi_prot_sg_count(SCpnt),
1953			SG_MITER_ATOMIC | SG_MITER_FROM_SG);
1954	sg_miter_start(&diter, scsi_sglist(SCpnt), scsi_sg_count(SCpnt),
1955			SG_MITER_ATOMIC | SG_MITER_FROM_SG);
1956
1957	/* For each protection page */
1958	while (sg_miter_next(&piter)) {
1959		dpage_offset = 0;
1960		if (WARN_ON(!sg_miter_next(&diter))) {
1961			ret = 0x01;
1962			goto out;
1963		}
1964
1965		for (ppage_offset = 0; ppage_offset < piter.length;
1966		     ppage_offset += sizeof(struct sd_dif_tuple)) {
1967			/* If we're at the end of the current
1968			 * data page advance to the next one
1969			 */
1970			if (dpage_offset >= diter.length) {
1971				if (WARN_ON(!sg_miter_next(&diter))) {
1972					ret = 0x01;
1973					goto out;
1974				}
1975				dpage_offset = 0;
1976			}
1977
1978			sdt = piter.addr + ppage_offset;
1979			daddr = diter.addr + dpage_offset;
1980
1981			ret = dif_verify(sdt, daddr, sector, ei_lba);
1982			if (ret) {
1983				dump_sector(daddr, scsi_debug_sector_size);
1984				goto out;
1985			}
1986
1987			sector++;
1988			ei_lba++;
1989			dpage_offset += scsi_debug_sector_size;
1990		}
1991		diter.consumed = dpage_offset;
1992		sg_miter_stop(&diter);
1993	}
1994	sg_miter_stop(&piter);
1995
1996	dif_copy_prot(SCpnt, start_sec, sectors, false);
1997	dix_writes++;
1998
1999	return 0;
2000
2001out:
2002	dif_errors++;
2003	sg_miter_stop(&diter);
2004	sg_miter_stop(&piter);
2005	return ret;
2006}
2007
2008static unsigned long lba_to_map_index(sector_t lba)
2009{
2010	if (scsi_debug_unmap_alignment) {
2011		lba += scsi_debug_unmap_granularity -
2012			scsi_debug_unmap_alignment;
2013	}
2014	do_div(lba, scsi_debug_unmap_granularity);
2015
2016	return lba;
2017}
2018
2019static sector_t map_index_to_lba(unsigned long index)
2020{
2021	sector_t lba = index * scsi_debug_unmap_granularity;
2022
2023	if (scsi_debug_unmap_alignment) {
2024		lba -= scsi_debug_unmap_granularity -
2025			scsi_debug_unmap_alignment;
2026	}
2027
 
 
2028	return lba;
2029}
2030
2031static unsigned int map_state(sector_t lba, unsigned int *num)
2032{
2033	sector_t end;
2034	unsigned int mapped;
2035	unsigned long index;
2036	unsigned long next;
2037
2038	index = lba_to_map_index(lba);
2039	mapped = test_bit(index, map_storep);
2040
2041	if (mapped)
2042		next = find_next_zero_bit(map_storep, map_size, index);
2043	else
2044		next = find_next_bit(map_storep, map_size, index);
2045
2046	end = min_t(sector_t, sdebug_store_sectors,  map_index_to_lba(next));
2047	*num = end - lba;
2048
2049	return mapped;
2050}
2051
2052static void map_region(sector_t lba, unsigned int len)
2053{
2054	sector_t end = lba + len;
2055
2056	while (lba < end) {
2057		unsigned long index = lba_to_map_index(lba);
2058
2059		if (index < map_size)
2060			set_bit(index, map_storep);
2061
2062		lba = map_index_to_lba(index + 1);
2063	}
2064}
2065
2066static void unmap_region(sector_t lba, unsigned int len)
2067{
2068	sector_t end = lba + len;
2069
2070	while (lba < end) {
2071		unsigned long index = lba_to_map_index(lba);
2072
2073		if (lba == map_index_to_lba(index) &&
2074		    lba + scsi_debug_unmap_granularity <= end &&
2075		    index < map_size) {
2076			clear_bit(index, map_storep);
2077			if (scsi_debug_lbprz) {
2078				memset(fake_storep +
2079				       lba * scsi_debug_sector_size, 0,
2080				       scsi_debug_sector_size *
2081				       scsi_debug_unmap_granularity);
 
2082			}
2083			if (dif_storep) {
2084				memset(dif_storep + lba, 0xff,
2085				       sizeof(*dif_storep) *
2086				       scsi_debug_unmap_granularity);
2087			}
2088		}
2089		lba = map_index_to_lba(index + 1);
2090	}
2091}
2092
2093static int resp_write(struct scsi_cmnd *SCpnt, unsigned long long lba,
2094		      unsigned int num, struct sdebug_dev_info *devip,
2095		      u32 ei_lba)
2096{
 
 
 
 
2097	unsigned long iflags;
2098	int ret;
 
2099
2100	ret = check_device_access_params(devip, lba, num);
2101	if (ret)
2102		return ret;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2103
2104	write_lock_irqsave(&atomic_rw, iflags);
2105
2106	/* DIX + T10 DIF */
2107	if (scsi_debug_dix && scsi_prot_sg_count(SCpnt)) {
2108		int prot_ret = prot_verify_write(SCpnt, lba, num, ei_lba);
2109
2110		if (prot_ret) {
2111			write_unlock_irqrestore(&atomic_rw, iflags);
2112			mk_sense_buffer(devip, ILLEGAL_REQUEST, 0x10, prot_ret);
2113			return illegal_condition_result;
2114		}
2115	}
2116
2117	ret = do_device_access(SCpnt, devip, lba, num, 1);
2118	if (scsi_debug_lbp())
2119		map_region(lba, num);
2120	write_unlock_irqrestore(&atomic_rw, iflags);
2121	if (-1 == ret)
2122		return (DID_ERROR << 16);
2123	else if ((ret < (num * scsi_debug_sector_size)) &&
2124		 (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
2125		printk(KERN_INFO "scsi_debug: write: cdb indicated=%u, "
2126		       " IO sent=%d bytes\n", num * scsi_debug_sector_size, ret);
2127
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2128	return 0;
2129}
2130
2131static int resp_write_same(struct scsi_cmnd *scmd, unsigned long long lba,
2132		      unsigned int num, struct sdebug_dev_info *devip,
2133			   u32 ei_lba, unsigned int unmap)
2134{
2135	unsigned long iflags;
2136	unsigned long long i;
2137	int ret;
 
2138
2139	ret = check_device_access_params(devip, lba, num);
2140	if (ret)
2141		return ret;
2142
2143	if (num > scsi_debug_write_same_length) {
2144		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB,
2145				0);
2146		return check_condition_result;
2147	}
2148
2149	write_lock_irqsave(&atomic_rw, iflags);
2150
2151	if (unmap && scsi_debug_lbp()) {
2152		unmap_region(lba, num);
2153		goto out;
2154	}
2155
2156	/* Else fetch one logical block */
2157	ret = fetch_to_dev_buffer(scmd,
2158				  fake_storep + (lba * scsi_debug_sector_size),
2159				  scsi_debug_sector_size);
 
 
 
 
2160
2161	if (-1 == ret) {
2162		write_unlock_irqrestore(&atomic_rw, iflags);
2163		return (DID_ERROR << 16);
2164	} else if ((ret < (num * scsi_debug_sector_size)) &&
2165		 (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
2166		printk(KERN_INFO "scsi_debug: write same: cdb indicated=%u, "
2167		       " IO sent=%d bytes\n", num * scsi_debug_sector_size, ret);
 
2168
2169	/* Copy first sector to remaining blocks */
2170	for (i = 1 ; i < num ; i++)
2171		memcpy(fake_storep + ((lba + i) * scsi_debug_sector_size),
2172		       fake_storep + (lba * scsi_debug_sector_size),
2173		       scsi_debug_sector_size);
2174
2175	if (scsi_debug_lbp())
2176		map_region(lba, num);
2177out:
2178	write_unlock_irqrestore(&atomic_rw, iflags);
2179
2180	return 0;
2181}
2182
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2183struct unmap_block_desc {
2184	__be64	lba;
2185	__be32	blocks;
2186	__be32	__reserved;
2187};
2188
2189static int resp_unmap(struct scsi_cmnd * scmd, struct sdebug_dev_info * devip)
2190{
2191	unsigned char *buf;
2192	struct unmap_block_desc *desc;
2193	unsigned int i, payload_len, descriptors;
2194	int ret;
2195	unsigned long iflags;
2196
2197	ret = check_readiness(scmd, 1, devip);
2198	if (ret)
2199		return ret;
2200
2201	payload_len = get_unaligned_be16(&scmd->cmnd[7]);
2202	BUG_ON(scsi_bufflen(scmd) != payload_len);
 
 
2203
2204	descriptors = (payload_len - 8) / 16;
 
 
 
 
2205
2206	buf = kmalloc(scsi_bufflen(scmd), GFP_ATOMIC);
2207	if (!buf)
 
 
2208		return check_condition_result;
 
2209
2210	scsi_sg_copy_to_buffer(scmd, buf, scsi_bufflen(scmd));
2211
2212	BUG_ON(get_unaligned_be16(&buf[0]) != payload_len - 2);
2213	BUG_ON(get_unaligned_be16(&buf[2]) != descriptors * 16);
2214
2215	desc = (void *)&buf[8];
2216
2217	write_lock_irqsave(&atomic_rw, iflags);
2218
2219	for (i = 0 ; i < descriptors ; i++) {
2220		unsigned long long lba = get_unaligned_be64(&desc[i].lba);
2221		unsigned int num = get_unaligned_be32(&desc[i].blocks);
2222
2223		ret = check_device_access_params(devip, lba, num);
2224		if (ret)
2225			goto out;
2226
2227		unmap_region(lba, num);
2228	}
2229
2230	ret = 0;
2231
2232out:
2233	write_unlock_irqrestore(&atomic_rw, iflags);
2234	kfree(buf);
2235
2236	return ret;
2237}
2238
2239#define SDEBUG_GET_LBA_STATUS_LEN 32
2240
2241static int resp_get_lba_status(struct scsi_cmnd * scmd,
2242			       struct sdebug_dev_info * devip)
2243{
2244	unsigned long long lba;
2245	unsigned int alloc_len, mapped, num;
2246	unsigned char arr[SDEBUG_GET_LBA_STATUS_LEN];
 
2247	int ret;
2248
2249	ret = check_readiness(scmd, 1, devip);
2250	if (ret)
2251		return ret;
2252
2253	lba = get_unaligned_be64(&scmd->cmnd[2]);
2254	alloc_len = get_unaligned_be32(&scmd->cmnd[10]);
2255
2256	if (alloc_len < 24)
2257		return 0;
2258
2259	ret = check_device_access_params(devip, lba, 1);
2260	if (ret)
2261		return ret;
2262
2263	mapped = map_state(lba, &num);
 
 
 
 
 
 
 
 
 
 
2264
2265	memset(arr, 0, SDEBUG_GET_LBA_STATUS_LEN);
2266	put_unaligned_be32(20, &arr[0]);	/* Parameter Data Length */
2267	put_unaligned_be64(lba, &arr[8]);	/* LBA */
2268	put_unaligned_be32(num, &arr[16]);	/* Number of blocks */
2269	arr[20] = !mapped;			/* mapped = 0, unmapped = 1 */
2270
2271	return fill_from_dev_buffer(scmd, arr, SDEBUG_GET_LBA_STATUS_LEN);
2272}
2273
2274#define SDEBUG_RLUN_ARR_SZ 256
2275
2276static int resp_report_luns(struct scsi_cmnd * scp,
2277			    struct sdebug_dev_info * devip)
 
 
 
 
 
 
 
 
2278{
 
2279	unsigned int alloc_len;
2280	int lun_cnt, i, upper, num, n, wlun, lun;
2281	unsigned char *cmd = (unsigned char *)scp->cmnd;
2282	int select_report = (int)cmd[2];
2283	struct scsi_lun *one_lun;
2284	unsigned char arr[SDEBUG_RLUN_ARR_SZ];
2285	unsigned char * max_addr;
2286
2287	alloc_len = cmd[9] + (cmd[8] << 8) + (cmd[7] << 16) + (cmd[6] << 24);
2288	if ((alloc_len < 4) || (select_report > 2)) {
2289		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB,
2290			       	0);
2291		return check_condition_result;
2292	}
2293	/* can produce response with up to 16k luns (lun 0 to lun 16383) */
2294	memset(arr, 0, SDEBUG_RLUN_ARR_SZ);
2295	lun_cnt = scsi_debug_max_luns;
2296	if (1 == select_report)
 
 
 
 
 
 
 
 
 
 
 
 
2297		lun_cnt = 0;
2298	else if (scsi_debug_no_lun_0 && (lun_cnt > 0))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2299		--lun_cnt;
2300	wlun = (select_report > 0) ? 1 : 0;
2301	num = lun_cnt + wlun;
2302	arr[2] = ((sizeof(struct scsi_lun) * num) >> 8) & 0xff;
2303	arr[3] = (sizeof(struct scsi_lun) * num) & 0xff;
2304	n = min((int)((SDEBUG_RLUN_ARR_SZ - 8) /
2305			    sizeof(struct scsi_lun)), num);
2306	if (n < num) {
2307		wlun = 0;
2308		lun_cnt = n;
2309	}
2310	one_lun = (struct scsi_lun *) &arr[8];
2311	max_addr = arr + SDEBUG_RLUN_ARR_SZ;
2312	for (i = 0, lun = (scsi_debug_no_lun_0 ? 1 : 0);
2313             ((i < lun_cnt) && ((unsigned char *)(one_lun + i) < max_addr));
2314	     i++, lun++) {
2315		upper = (lun >> 8) & 0x3f;
2316		if (upper)
2317			one_lun[i].scsi_lun[0] =
2318			    (upper | (SAM2_LUN_ADDRESS_METHOD << 6));
2319		one_lun[i].scsi_lun[1] = lun & 0xff;
2320	}
2321	if (wlun) {
2322		one_lun[i].scsi_lun[0] = (SAM2_WLUN_REPORT_LUNS >> 8) & 0xff;
2323		one_lun[i].scsi_lun[1] = SAM2_WLUN_REPORT_LUNS & 0xff;
2324		i++;
 
 
 
 
2325	}
2326	alloc_len = (unsigned char *)(one_lun + i) - arr;
2327	return fill_from_dev_buffer(scp, arr,
2328				    min((int)alloc_len, SDEBUG_RLUN_ARR_SZ));
 
 
 
 
2329}
2330
2331static int resp_xdwriteread(struct scsi_cmnd *scp, unsigned long long lba,
2332			    unsigned int num, struct sdebug_dev_info *devip)
2333{
2334	int j;
2335	unsigned char *kaddr, *buf;
2336	unsigned int offset;
2337	struct scsi_data_buffer *sdb = scsi_in(scp);
2338	struct sg_mapping_iter miter;
2339
2340	/* better not to use temporary buffer. */
2341	buf = kmalloc(scsi_bufflen(scp), GFP_ATOMIC);
2342	if (!buf) {
2343		mk_sense_buffer(devip, NOT_READY,
2344				LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
2345		return check_condition_result;
2346	}
2347
2348	scsi_sg_copy_to_buffer(scp, buf, scsi_bufflen(scp));
2349
2350	offset = 0;
2351	sg_miter_start(&miter, sdb->table.sgl, sdb->table.nents,
2352			SG_MITER_ATOMIC | SG_MITER_TO_SG);
2353
2354	while (sg_miter_next(&miter)) {
2355		kaddr = miter.addr;
2356		for (j = 0; j < miter.length; j++)
2357			*(kaddr + j) ^= *(buf + offset + j);
2358
2359		offset += miter.length;
2360	}
2361	sg_miter_stop(&miter);
2362	kfree(buf);
2363
2364	return 0;
2365}
2366
2367/* When timer goes off this function is called. */
2368static void timer_intr_handler(unsigned long indx)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2369{
2370	struct sdebug_queued_cmd * sqcp;
 
2371	unsigned long iflags;
 
 
 
 
2372
2373	if (indx >= scsi_debug_max_queue) {
2374		printk(KERN_ERR "scsi_debug:timer_intr_handler: indx too "
2375		       "large\n");
 
 
 
 
 
 
2376		return;
2377	}
2378	spin_lock_irqsave(&queued_arr_lock, iflags);
2379	sqcp = &queued_arr[(int)indx];
2380	if (! sqcp->in_use) {
2381		printk(KERN_ERR "scsi_debug:timer_intr_handler: Unexpected "
2382		       "interrupt\n");
2383		spin_unlock_irqrestore(&queued_arr_lock, iflags);
 
2384		return;
2385	}
2386	sqcp->in_use = 0;
2387	if (sqcp->done_funct) {
2388		sqcp->a_cmnd->result = sqcp->scsi_result;
2389		sqcp->done_funct(sqcp->a_cmnd); /* callback to mid level */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2390	}
2391	sqcp->done_funct = NULL;
2392	spin_unlock_irqrestore(&queued_arr_lock, iflags);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2393}
2394
 
 
2395
2396static struct sdebug_dev_info *
2397sdebug_device_create(struct sdebug_host_info *sdbg_host, gfp_t flags)
2398{
2399	struct sdebug_dev_info *devip;
2400
2401	devip = kzalloc(sizeof(*devip), flags);
2402	if (devip) {
 
 
 
 
 
 
 
 
 
 
 
2403		devip->sdbg_host = sdbg_host;
2404		list_add_tail(&devip->dev_list, &sdbg_host->dev_info_list);
2405	}
2406	return devip;
2407}
2408
2409static struct sdebug_dev_info * devInfoReg(struct scsi_device * sdev)
2410{
2411	struct sdebug_host_info * sdbg_host;
2412	struct sdebug_dev_info * open_devip = NULL;
2413	struct sdebug_dev_info * devip =
2414			(struct sdebug_dev_info *)sdev->hostdata;
2415
2416	if (devip)
2417		return devip;
2418	sdbg_host = *(struct sdebug_host_info **)shost_priv(sdev->host);
2419	if (!sdbg_host) {
2420                printk(KERN_ERR "Host info NULL\n");
2421		return NULL;
2422        }
2423	list_for_each_entry(devip, &sdbg_host->dev_info_list, dev_list) {
2424		if ((devip->used) && (devip->channel == sdev->channel) &&
2425                    (devip->target == sdev->id) &&
2426                    (devip->lun == sdev->lun))
2427                        return devip;
2428		else {
2429			if ((!devip->used) && (!open_devip))
2430				open_devip = devip;
2431		}
2432	}
2433	if (!open_devip) { /* try and make a new one */
2434		open_devip = sdebug_device_create(sdbg_host, GFP_ATOMIC);
2435		if (!open_devip) {
2436			printk(KERN_ERR "%s: out of memory at line %d\n",
2437				__func__, __LINE__);
2438			return NULL;
2439		}
2440	}
2441
2442	open_devip->channel = sdev->channel;
2443	open_devip->target = sdev->id;
2444	open_devip->lun = sdev->lun;
2445	open_devip->sdbg_host = sdbg_host;
2446	open_devip->reset = 1;
2447	open_devip->used = 1;
2448	memset(open_devip->sense_buff, 0, SDEBUG_SENSE_LEN);
2449	if (scsi_debug_dsense)
2450		open_devip->sense_buff[0] = 0x72;
2451	else {
2452		open_devip->sense_buff[0] = 0x70;
2453		open_devip->sense_buff[7] = 0xa;
2454	}
2455	if (sdev->lun == SAM2_WLUN_REPORT_LUNS)
2456		open_devip->wlun = SAM2_WLUN_REPORT_LUNS & 0xff;
2457
2458	return open_devip;
2459}
2460
2461static int scsi_debug_slave_alloc(struct scsi_device *sdp)
2462{
2463	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2464		printk(KERN_INFO "scsi_debug: slave_alloc <%u %u %u %u>\n",
2465		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
2466	queue_flag_set_unlocked(QUEUE_FLAG_BIDI, sdp->request_queue);
2467	return 0;
2468}
2469
2470static int scsi_debug_slave_configure(struct scsi_device *sdp)
2471{
2472	struct sdebug_dev_info *devip;
 
2473
2474	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2475		printk(KERN_INFO "scsi_debug: slave_configure <%u %u %u %u>\n",
2476		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
2477	if (sdp->host->max_cmd_len != SCSI_DEBUG_MAX_CMD_LEN)
2478		sdp->host->max_cmd_len = SCSI_DEBUG_MAX_CMD_LEN;
2479	devip = devInfoReg(sdp);
2480	if (NULL == devip)
2481		return 1;	/* no resources, will be marked offline */
 
 
2482	sdp->hostdata = devip;
2483	if (sdp->host->cmd_per_lun)
2484		scsi_adjust_queue_depth(sdp, SDEBUG_TAGGED_QUEUING,
2485					sdp->host->cmd_per_lun);
2486	blk_queue_max_segment_size(sdp->request_queue, 256 * 1024);
2487	if (scsi_debug_no_uld)
2488		sdp->no_uld_attach = 1;
2489	return 0;
2490}
2491
2492static void scsi_debug_slave_destroy(struct scsi_device *sdp)
2493{
2494	struct sdebug_dev_info *devip =
2495		(struct sdebug_dev_info *)sdp->hostdata;
2496
2497	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2498		printk(KERN_INFO "scsi_debug: slave_destroy <%u %u %u %u>\n",
2499		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
2500	if (devip) {
2501		/* make this slot available for re-use */
2502		devip->used = 0;
2503		sdp->hostdata = NULL;
2504	}
2505}
2506
2507/* Returns 1 if found 'cmnd' and deleted its timer. else returns 0 */
2508static int stop_queued_cmnd(struct scsi_cmnd *cmnd)
 
 
 
 
 
 
 
 
 
 
 
2509{
2510	unsigned long iflags;
2511	int k;
 
2512	struct sdebug_queued_cmd *sqcp;
 
 
2513
2514	spin_lock_irqsave(&queued_arr_lock, iflags);
2515	for (k = 0; k < scsi_debug_max_queue; ++k) {
2516		sqcp = &queued_arr[k];
2517		if (sqcp->in_use && (cmnd == sqcp->a_cmnd)) {
2518			del_timer_sync(&sqcp->cmnd_timer);
2519			sqcp->in_use = 0;
2520			sqcp->a_cmnd = NULL;
2521			break;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2522		}
 
2523	}
2524	spin_unlock_irqrestore(&queued_arr_lock, iflags);
2525	return (k < scsi_debug_max_queue) ? 1 : 0;
2526}
2527
2528/* Deletes (stops) timers of all queued commands */
2529static void stop_all_queued(void)
2530{
2531	unsigned long iflags;
2532	int k;
 
2533	struct sdebug_queued_cmd *sqcp;
 
 
2534
2535	spin_lock_irqsave(&queued_arr_lock, iflags);
2536	for (k = 0; k < scsi_debug_max_queue; ++k) {
2537		sqcp = &queued_arr[k];
2538		if (sqcp->in_use && sqcp->a_cmnd) {
2539			del_timer_sync(&sqcp->cmnd_timer);
2540			sqcp->in_use = 0;
2541			sqcp->a_cmnd = NULL;
 
 
 
 
 
 
 
 
 
 
 
2542		}
 
2543	}
2544	spin_unlock_irqrestore(&queued_arr_lock, iflags);
2545}
2546
2547static int scsi_debug_abort(struct scsi_cmnd * SCpnt)
 
2548{
2549	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2550		printk(KERN_INFO "scsi_debug: abort\n");
2551	++num_aborts;
2552	stop_queued_cmnd(SCpnt);
2553	return SUCCESS;
 
 
 
 
 
 
2554}
2555
2556static int scsi_debug_biosparam(struct scsi_device *sdev,
2557		struct block_device * bdev, sector_t capacity, int *info)
2558{
2559	int res;
2560	unsigned char *buf;
2561
2562	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2563		printk(KERN_INFO "scsi_debug: biosparam\n");
2564	buf = scsi_bios_ptable(bdev);
2565	if (buf) {
2566		res = scsi_partsize(buf, capacity,
2567				    &info[2], &info[0], &info[1]);
2568		kfree(buf);
2569		if (! res)
2570			return res;
2571	}
2572	info[0] = sdebug_heads;
2573	info[1] = sdebug_sectors_per;
2574	info[2] = sdebug_cylinders_per;
2575	return 0;
2576}
2577
2578static int scsi_debug_device_reset(struct scsi_cmnd * SCpnt)
2579{
2580	struct sdebug_dev_info * devip;
2581
2582	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2583		printk(KERN_INFO "scsi_debug: device_reset\n");
2584	++num_dev_resets;
2585	if (SCpnt) {
2586		devip = devInfoReg(SCpnt->device);
 
 
 
 
 
2587		if (devip)
2588			devip->reset = 1;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2589	}
 
 
 
 
2590	return SUCCESS;
2591}
2592
2593static int scsi_debug_bus_reset(struct scsi_cmnd * SCpnt)
2594{
2595	struct sdebug_host_info *sdbg_host;
2596        struct sdebug_dev_info * dev_info;
2597        struct scsi_device * sdp;
2598        struct Scsi_Host * hp;
 
2599
2600	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2601		printk(KERN_INFO "scsi_debug: bus_reset\n");
2602	++num_bus_resets;
2603	if (SCpnt && ((sdp = SCpnt->device)) && ((hp = sdp->host))) {
 
 
 
 
 
 
2604		sdbg_host = *(struct sdebug_host_info **)shost_priv(hp);
2605		if (sdbg_host) {
2606			list_for_each_entry(dev_info,
2607                                            &sdbg_host->dev_info_list,
2608                                            dev_list)
2609				dev_info->reset = 1;
 
 
2610		}
2611	}
 
 
 
 
2612	return SUCCESS;
2613}
2614
2615static int scsi_debug_host_reset(struct scsi_cmnd * SCpnt)
2616{
2617	struct sdebug_host_info * sdbg_host;
2618        struct sdebug_dev_info * dev_info;
 
2619
2620	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
2621		printk(KERN_INFO "scsi_debug: host_reset\n");
2622	++num_host_resets;
 
 
2623        spin_lock(&sdebug_host_list_lock);
2624        list_for_each_entry(sdbg_host, &sdebug_host_list, host_list) {
2625                list_for_each_entry(dev_info, &sdbg_host->dev_info_list,
2626                                    dev_list)
2627                        dev_info->reset = 1;
 
 
2628        }
2629        spin_unlock(&sdebug_host_list_lock);
2630	stop_all_queued();
 
 
 
2631	return SUCCESS;
2632}
2633
2634/* Initializes timers in queued array */
2635static void __init init_all_queued(void)
2636{
2637	unsigned long iflags;
2638	int k;
2639	struct sdebug_queued_cmd * sqcp;
2640
2641	spin_lock_irqsave(&queued_arr_lock, iflags);
2642	for (k = 0; k < scsi_debug_max_queue; ++k) {
2643		sqcp = &queued_arr[k];
2644		init_timer(&sqcp->cmnd_timer);
2645		sqcp->in_use = 0;
2646		sqcp->a_cmnd = NULL;
2647	}
2648	spin_unlock_irqrestore(&queued_arr_lock, iflags);
2649}
2650
2651static void __init sdebug_build_parts(unsigned char *ramp,
2652				      unsigned long store_size)
2653{
2654	struct partition * pp;
2655	int starts[SDEBUG_MAX_PARTS + 2];
2656	int sectors_per_part, num_sectors, k;
2657	int heads_by_sects, start_sec, end_sec;
2658
2659	/* assume partition table already zeroed */
2660	if ((scsi_debug_num_parts < 1) || (store_size < 1048576))
2661		return;
2662	if (scsi_debug_num_parts > SDEBUG_MAX_PARTS) {
2663		scsi_debug_num_parts = SDEBUG_MAX_PARTS;
2664		printk(KERN_WARNING "scsi_debug:build_parts: reducing "
2665				    "partitions to %d\n", SDEBUG_MAX_PARTS);
2666	}
2667	num_sectors = (int)sdebug_store_sectors;
2668	sectors_per_part = (num_sectors - sdebug_sectors_per)
2669			   / scsi_debug_num_parts;
2670	heads_by_sects = sdebug_heads * sdebug_sectors_per;
2671        starts[0] = sdebug_sectors_per;
2672	for (k = 1; k < scsi_debug_num_parts; ++k)
2673		starts[k] = ((k * sectors_per_part) / heads_by_sects)
2674			    * heads_by_sects;
2675	starts[scsi_debug_num_parts] = num_sectors;
2676	starts[scsi_debug_num_parts + 1] = 0;
2677
2678	ramp[510] = 0x55;	/* magic partition markings */
2679	ramp[511] = 0xAA;
2680	pp = (struct partition *)(ramp + 0x1be);
2681	for (k = 0; starts[k + 1]; ++k, ++pp) {
2682		start_sec = starts[k];
2683		end_sec = starts[k + 1] - 1;
2684		pp->boot_ind = 0;
2685
2686		pp->cyl = start_sec / heads_by_sects;
2687		pp->head = (start_sec - (pp->cyl * heads_by_sects))
2688			   / sdebug_sectors_per;
2689		pp->sector = (start_sec % sdebug_sectors_per) + 1;
2690
2691		pp->end_cyl = end_sec / heads_by_sects;
2692		pp->end_head = (end_sec - (pp->end_cyl * heads_by_sects))
2693			       / sdebug_sectors_per;
2694		pp->end_sector = (end_sec % sdebug_sectors_per) + 1;
2695
2696		pp->start_sect = cpu_to_le32(start_sec);
2697		pp->nr_sects = cpu_to_le32(end_sec - start_sec + 1);
2698		pp->sys_ind = 0x83;	/* plain Linux partition */
2699	}
2700}
2701
2702static int schedule_resp(struct scsi_cmnd * cmnd,
2703			 struct sdebug_dev_info * devip,
2704			 done_funct_t done, int scsi_result, int delta_jiff)
2705{
2706	if ((SCSI_DEBUG_OPT_NOISE & scsi_debug_opts) && cmnd) {
2707		if (scsi_result) {
2708			struct scsi_device * sdp = cmnd->device;
2709
2710			printk(KERN_INFO "scsi_debug:    <%u %u %u %u> "
2711			       "non-zero result=0x%x\n", sdp->host->host_no,
2712			       sdp->channel, sdp->id, sdp->lun, scsi_result);
2713		}
2714	}
2715	if (cmnd && devip) {
2716		/* simulate autosense by this driver */
2717		if (SAM_STAT_CHECK_CONDITION == (scsi_result & 0xff))
2718			memcpy(cmnd->sense_buffer, devip->sense_buff,
2719			       (SCSI_SENSE_BUFFERSIZE > SDEBUG_SENSE_LEN) ?
2720			       SDEBUG_SENSE_LEN : SCSI_SENSE_BUFFERSIZE);
2721	}
2722	if (delta_jiff <= 0) {
2723		if (cmnd)
2724			cmnd->result = scsi_result;
2725		if (done)
2726			done(cmnd);
2727		return 0;
2728	} else {
2729		unsigned long iflags;
2730		int k;
2731		struct sdebug_queued_cmd * sqcp = NULL;
2732
2733		spin_lock_irqsave(&queued_arr_lock, iflags);
2734		for (k = 0; k < scsi_debug_max_queue; ++k) {
2735			sqcp = &queued_arr[k];
2736			if (! sqcp->in_use)
2737				break;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2738		}
2739		if (k >= scsi_debug_max_queue) {
2740			spin_unlock_irqrestore(&queued_arr_lock, iflags);
2741			printk(KERN_WARNING "scsi_debug: can_queue exceeded\n");
2742			return 1;	/* report busy to mid level */
2743		}
2744		sqcp->in_use = 1;
2745		sqcp->a_cmnd = cmnd;
2746		sqcp->scsi_result = scsi_result;
2747		sqcp->done_funct = done;
2748		sqcp->cmnd_timer.function = timer_intr_handler;
2749		sqcp->cmnd_timer.data = k;
2750		sqcp->cmnd_timer.expires = jiffies + delta_jiff;
2751		add_timer(&sqcp->cmnd_timer);
2752		spin_unlock_irqrestore(&queued_arr_lock, iflags);
2753		if (cmnd)
2754			cmnd->result = 0;
2755		return 0;
2756	}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2757}
 
2758/* Note: The following macros create attribute files in the
2759   /sys/module/scsi_debug/parameters directory. Unfortunately this
2760   driver is unaware of a change and cannot trigger auxiliary actions
2761   as it can when the corresponding attribute in the
2762   /sys/bus/pseudo/drivers/scsi_debug directory is changed.
2763 */
2764module_param_named(add_host, scsi_debug_add_host, int, S_IRUGO | S_IWUSR);
2765module_param_named(ato, scsi_debug_ato, int, S_IRUGO);
2766module_param_named(clustering, scsi_debug_clustering, bool, S_IRUGO | S_IWUSR);
2767module_param_named(delay, scsi_debug_delay, int, S_IRUGO | S_IWUSR);
2768module_param_named(dev_size_mb, scsi_debug_dev_size_mb, int, S_IRUGO);
2769module_param_named(dif, scsi_debug_dif, int, S_IRUGO);
2770module_param_named(dix, scsi_debug_dix, int, S_IRUGO);
2771module_param_named(dsense, scsi_debug_dsense, int, S_IRUGO | S_IWUSR);
2772module_param_named(every_nth, scsi_debug_every_nth, int, S_IRUGO | S_IWUSR);
2773module_param_named(fake_rw, scsi_debug_fake_rw, int, S_IRUGO | S_IWUSR);
2774module_param_named(guard, scsi_debug_guard, uint, S_IRUGO);
2775module_param_named(lbpu, scsi_debug_lbpu, int, S_IRUGO);
2776module_param_named(lbpws, scsi_debug_lbpws, int, S_IRUGO);
2777module_param_named(lbpws10, scsi_debug_lbpws10, int, S_IRUGO);
2778module_param_named(lbprz, scsi_debug_lbprz, int, S_IRUGO);
2779module_param_named(lowest_aligned, scsi_debug_lowest_aligned, int, S_IRUGO);
2780module_param_named(max_luns, scsi_debug_max_luns, int, S_IRUGO | S_IWUSR);
2781module_param_named(max_queue, scsi_debug_max_queue, int, S_IRUGO | S_IWUSR);
2782module_param_named(no_lun_0, scsi_debug_no_lun_0, int, S_IRUGO | S_IWUSR);
2783module_param_named(no_uld, scsi_debug_no_uld, int, S_IRUGO);
2784module_param_named(num_parts, scsi_debug_num_parts, int, S_IRUGO);
2785module_param_named(num_tgts, scsi_debug_num_tgts, int, S_IRUGO | S_IWUSR);
2786module_param_named(opt_blks, scsi_debug_opt_blks, int, S_IRUGO);
2787module_param_named(opts, scsi_debug_opts, int, S_IRUGO | S_IWUSR);
2788module_param_named(physblk_exp, scsi_debug_physblk_exp, int, S_IRUGO);
2789module_param_named(ptype, scsi_debug_ptype, int, S_IRUGO | S_IWUSR);
2790module_param_named(removable, scsi_debug_removable, bool, S_IRUGO | S_IWUSR);
2791module_param_named(scsi_level, scsi_debug_scsi_level, int, S_IRUGO);
2792module_param_named(sector_size, scsi_debug_sector_size, int, S_IRUGO);
2793module_param_named(unmap_alignment, scsi_debug_unmap_alignment, int, S_IRUGO);
2794module_param_named(unmap_granularity, scsi_debug_unmap_granularity, int, S_IRUGO);
2795module_param_named(unmap_max_blocks, scsi_debug_unmap_max_blocks, int, S_IRUGO);
2796module_param_named(unmap_max_desc, scsi_debug_unmap_max_desc, int, S_IRUGO);
2797module_param_named(virtual_gb, scsi_debug_virtual_gb, int, S_IRUGO | S_IWUSR);
2798module_param_named(vpd_use_hostno, scsi_debug_vpd_use_hostno, int,
 
 
 
 
 
 
2799		   S_IRUGO | S_IWUSR);
2800module_param_named(write_same_length, scsi_debug_write_same_length, int,
2801		   S_IRUGO | S_IWUSR);
2802
2803MODULE_AUTHOR("Eric Youngdale + Douglas Gilbert");
2804MODULE_DESCRIPTION("SCSI debug adapter driver");
2805MODULE_LICENSE("GPL");
2806MODULE_VERSION(SCSI_DEBUG_VERSION);
2807
2808MODULE_PARM_DESC(add_host, "0..127 hosts allowed(def=1)");
2809MODULE_PARM_DESC(ato, "application tag ownership: 0=disk 1=host (def=1)");
2810MODULE_PARM_DESC(clustering, "when set enables larger transfers (def=0)");
2811MODULE_PARM_DESC(delay, "# of jiffies to delay response(def=1)");
2812MODULE_PARM_DESC(dev_size_mb, "size in MB of ram shared by devs(def=8)");
2813MODULE_PARM_DESC(dif, "data integrity field type: 0-3 (def=0)");
2814MODULE_PARM_DESC(dix, "data integrity extensions mask (def=0)");
2815MODULE_PARM_DESC(dsense, "use descriptor sense format(def=0 -> fixed)");
2816MODULE_PARM_DESC(every_nth, "timeout every nth command(def=0)");
2817MODULE_PARM_DESC(fake_rw, "fake reads/writes instead of copying (def=0)");
2818MODULE_PARM_DESC(guard, "protection checksum: 0=crc, 1=ip (def=0)");
 
2819MODULE_PARM_DESC(lbpu, "enable LBP, support UNMAP command (def=0)");
2820MODULE_PARM_DESC(lbpws, "enable LBP, support WRITE SAME(16) with UNMAP bit (def=0)");
2821MODULE_PARM_DESC(lbpws10, "enable LBP, support WRITE SAME(10) with UNMAP bit (def=0)");
2822MODULE_PARM_DESC(lbprz, "unmapped blocks return 0 on read (def=1)");
 
2823MODULE_PARM_DESC(lowest_aligned, "lowest aligned lba (def=0)");
2824MODULE_PARM_DESC(max_luns, "number of LUNs per target to simulate(def=1)");
2825MODULE_PARM_DESC(max_queue, "max number of queued commands (1 to 255(def))");
 
2826MODULE_PARM_DESC(no_lun_0, "no LU number 0 (def=0 -> have lun 0)");
2827MODULE_PARM_DESC(no_uld, "stop ULD (e.g. sd driver) attaching (def=0))");
2828MODULE_PARM_DESC(num_parts, "number of partitions(def=0)");
2829MODULE_PARM_DESC(num_tgts, "number of targets per host to simulate(def=1)");
2830MODULE_PARM_DESC(opt_blks, "optimal transfer length in block (def=64)");
2831MODULE_PARM_DESC(opts, "1->noise, 2->medium_err, 4->timeout, 8->recovered_err... (def=0)");
2832MODULE_PARM_DESC(physblk_exp, "physical block exponent (def=0)");
2833MODULE_PARM_DESC(ptype, "SCSI peripheral type(def=0[disk])");
2834MODULE_PARM_DESC(removable, "claim to have removable media (def=0)");
2835MODULE_PARM_DESC(scsi_level, "SCSI level to simulate(def=5[SPC-3])");
2836MODULE_PARM_DESC(sector_size, "logical block size in bytes (def=512)");
 
 
 
2837MODULE_PARM_DESC(unmap_alignment, "lowest aligned thin provisioning lba (def=0)");
2838MODULE_PARM_DESC(unmap_granularity, "thin provisioning granularity in blocks (def=1)");
2839MODULE_PARM_DESC(unmap_max_blocks, "max # of blocks can be unmapped in one cmd (def=0xffffffff)");
2840MODULE_PARM_DESC(unmap_max_desc, "max # of ranges that can be unmapped in one cmd (def=256)");
2841MODULE_PARM_DESC(virtual_gb, "virtual gigabyte size (def=0 -> use dev_size_mb)");
 
 
2842MODULE_PARM_DESC(vpd_use_hostno, "0 -> dev ids ignore hostno (def=1 -> unique dev ids)");
2843MODULE_PARM_DESC(write_same_length, "Maximum blocks per WRITE SAME cmd (def=0xffff)");
2844
2845static char sdebug_info[256];
 
2846
2847static const char * scsi_debug_info(struct Scsi_Host * shp)
2848{
2849	sprintf(sdebug_info, "scsi_debug, version %s [%s], "
2850		"dev_size_mb=%d, opts=0x%x", SCSI_DEBUG_VERSION,
2851		scsi_debug_version_date, scsi_debug_dev_size_mb,
2852		scsi_debug_opts);
 
 
 
 
 
 
2853	return sdebug_info;
2854}
2855
2856/* scsi_debug_proc_info
2857 * Used if the driver currently has no own support for /proc/scsi
2858 */
2859static int scsi_debug_write_info(struct Scsi_Host *host, char *buffer, int length)
2860{
2861	char arr[16];
2862	int opts;
2863	int minLen = length > 15 ? 15 : length;
2864
2865	if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO))
2866		return -EACCES;
2867	memcpy(arr, buffer, minLen);
2868	arr[minLen] = '\0';
2869	if (1 != sscanf(arr, "%d", &opts))
2870		return -EINVAL;
2871	scsi_debug_opts = opts;
2872	if (scsi_debug_every_nth != 0)
2873		scsi_debug_cmnd_count = 0;
 
 
2874	return length;
2875}
2876
 
 
 
2877static int scsi_debug_show_info(struct seq_file *m, struct Scsi_Host *host)
2878{
2879	seq_printf(m, "scsi_debug adapter driver, version "
2880	    "%s [%s]\n"
2881	    "num_tgts=%d, shared (ram) size=%d MB, opts=0x%x, "
2882	    "every_nth=%d(curr:%d)\n"
2883	    "delay=%d, max_luns=%d, scsi_level=%d\n"
2884	    "sector_size=%d bytes, cylinders=%d, heads=%d, sectors=%d\n"
2885	    "number of aborts=%d, device_reset=%d, bus_resets=%d, "
2886	    "host_resets=%d\ndix_reads=%d dix_writes=%d dif_errors=%d\n",
2887	    SCSI_DEBUG_VERSION, scsi_debug_version_date, scsi_debug_num_tgts,
2888	    scsi_debug_dev_size_mb, scsi_debug_opts, scsi_debug_every_nth,
2889	    scsi_debug_cmnd_count, scsi_debug_delay,
2890	    scsi_debug_max_luns, scsi_debug_scsi_level,
2891	    scsi_debug_sector_size, sdebug_cylinders_per, sdebug_heads,
2892	    sdebug_sectors_per, num_aborts, num_dev_resets, num_bus_resets,
2893	    num_host_resets, dix_reads, dix_writes, dif_errors);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2894	return 0;
2895}
2896
2897static ssize_t delay_show(struct device_driver *ddp, char *buf)
2898{
2899        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_delay);
2900}
2901
 
 
2902static ssize_t delay_store(struct device_driver *ddp, const char *buf,
2903			   size_t count)
2904{
2905        int delay;
2906	char work[20];
2907
2908        if (1 == sscanf(buf, "%10s", work)) {
2909		if ((1 == sscanf(work, "%d", &delay)) && (delay >= 0)) {
2910			scsi_debug_delay = delay;
2911			return count;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2912		}
 
2913	}
2914	return -EINVAL;
2915}
2916static DRIVER_ATTR_RW(delay);
2917
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2918static ssize_t opts_show(struct device_driver *ddp, char *buf)
2919{
2920        return scnprintf(buf, PAGE_SIZE, "0x%x\n", scsi_debug_opts);
2921}
2922
2923static ssize_t opts_store(struct device_driver *ddp, const char *buf,
2924			  size_t count)
2925{
2926        int opts;
2927	char work[20];
2928
2929        if (1 == sscanf(buf, "%10s", work)) {
2930		if (0 == strnicmp(work,"0x", 2)) {
2931			if (1 == sscanf(&work[2], "%x", &opts))
2932				goto opts_done;
2933		} else {
2934			if (1 == sscanf(work, "%d", &opts))
2935				goto opts_done;
2936		}
2937	}
2938	return -EINVAL;
2939opts_done:
2940	scsi_debug_opts = opts;
2941	scsi_debug_cmnd_count = 0;
 
 
2942	return count;
2943}
2944static DRIVER_ATTR_RW(opts);
2945
2946static ssize_t ptype_show(struct device_driver *ddp, char *buf)
2947{
2948        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_ptype);
2949}
2950static ssize_t ptype_store(struct device_driver *ddp, const char *buf,
2951			   size_t count)
2952{
2953        int n;
2954
2955	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
2956		scsi_debug_ptype = n;
2957		return count;
2958	}
2959	return -EINVAL;
2960}
2961static DRIVER_ATTR_RW(ptype);
2962
2963static ssize_t dsense_show(struct device_driver *ddp, char *buf)
2964{
2965        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dsense);
2966}
2967static ssize_t dsense_store(struct device_driver *ddp, const char *buf,
2968			    size_t count)
2969{
2970        int n;
2971
2972	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
2973		scsi_debug_dsense = n;
2974		return count;
2975	}
2976	return -EINVAL;
2977}
2978static DRIVER_ATTR_RW(dsense);
2979
2980static ssize_t fake_rw_show(struct device_driver *ddp, char *buf)
2981{
2982        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_fake_rw);
2983}
2984static ssize_t fake_rw_store(struct device_driver *ddp, const char *buf,
2985			     size_t count)
2986{
2987        int n;
2988
2989	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
2990		scsi_debug_fake_rw = n;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2991		return count;
2992	}
2993	return -EINVAL;
2994}
2995static DRIVER_ATTR_RW(fake_rw);
2996
2997static ssize_t no_lun_0_show(struct device_driver *ddp, char *buf)
2998{
2999        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_no_lun_0);
3000}
3001static ssize_t no_lun_0_store(struct device_driver *ddp, const char *buf,
3002			      size_t count)
3003{
3004        int n;
3005
3006	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
3007		scsi_debug_no_lun_0 = n;
3008		return count;
3009	}
3010	return -EINVAL;
3011}
3012static DRIVER_ATTR_RW(no_lun_0);
3013
3014static ssize_t num_tgts_show(struct device_driver *ddp, char *buf)
3015{
3016        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_num_tgts);
3017}
3018static ssize_t num_tgts_store(struct device_driver *ddp, const char *buf,
3019			      size_t count)
3020{
3021        int n;
3022
3023	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
3024		scsi_debug_num_tgts = n;
3025		sdebug_max_tgts_luns();
3026		return count;
3027	}
3028	return -EINVAL;
3029}
3030static DRIVER_ATTR_RW(num_tgts);
3031
3032static ssize_t dev_size_mb_show(struct device_driver *ddp, char *buf)
3033{
3034        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dev_size_mb);
3035}
3036static DRIVER_ATTR_RO(dev_size_mb);
3037
3038static ssize_t num_parts_show(struct device_driver *ddp, char *buf)
3039{
3040        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_num_parts);
3041}
3042static DRIVER_ATTR_RO(num_parts);
3043
3044static ssize_t every_nth_show(struct device_driver *ddp, char *buf)
3045{
3046        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_every_nth);
3047}
3048static ssize_t every_nth_store(struct device_driver *ddp, const char *buf,
3049			       size_t count)
3050{
3051        int nth;
3052
3053	if ((count > 0) && (1 == sscanf(buf, "%d", &nth))) {
3054		scsi_debug_every_nth = nth;
3055		scsi_debug_cmnd_count = 0;
 
 
 
 
3056		return count;
3057	}
3058	return -EINVAL;
3059}
3060static DRIVER_ATTR_RW(every_nth);
3061
3062static ssize_t max_luns_show(struct device_driver *ddp, char *buf)
3063{
3064        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_max_luns);
3065}
3066static ssize_t max_luns_store(struct device_driver *ddp, const char *buf,
3067			      size_t count)
3068{
3069        int n;
 
3070
3071	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
3072		scsi_debug_max_luns = n;
 
 
 
 
 
3073		sdebug_max_tgts_luns();
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3074		return count;
3075	}
3076	return -EINVAL;
3077}
3078static DRIVER_ATTR_RW(max_luns);
3079
3080static ssize_t max_queue_show(struct device_driver *ddp, char *buf)
3081{
3082        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_max_queue);
3083}
 
 
3084static ssize_t max_queue_store(struct device_driver *ddp, const char *buf,
3085			       size_t count)
3086{
3087        int n;
 
3088
3089	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n > 0) &&
3090	    (n <= SCSI_DEBUG_CANQUEUE)) {
3091		scsi_debug_max_queue = n;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3092		return count;
3093	}
3094	return -EINVAL;
3095}
3096static DRIVER_ATTR_RW(max_queue);
3097
3098static ssize_t no_uld_show(struct device_driver *ddp, char *buf)
3099{
3100        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_no_uld);
3101}
3102static DRIVER_ATTR_RO(no_uld);
3103
3104static ssize_t scsi_level_show(struct device_driver *ddp, char *buf)
3105{
3106        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_scsi_level);
3107}
3108static DRIVER_ATTR_RO(scsi_level);
3109
3110static ssize_t virtual_gb_show(struct device_driver *ddp, char *buf)
3111{
3112        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_virtual_gb);
3113}
3114static ssize_t virtual_gb_store(struct device_driver *ddp, const char *buf,
3115				size_t count)
3116{
3117        int n;
 
3118
3119	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
3120		scsi_debug_virtual_gb = n;
3121
3122		sdebug_capacity = get_sdebug_capacity();
3123
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3124		return count;
3125	}
3126	return -EINVAL;
3127}
3128static DRIVER_ATTR_RW(virtual_gb);
3129
3130static ssize_t add_host_show(struct device_driver *ddp, char *buf)
3131{
3132        return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_add_host);
3133}
3134
 
 
 
3135static ssize_t add_host_store(struct device_driver *ddp, const char *buf,
3136			      size_t count)
3137{
3138	int delta_hosts;
3139
3140	if (sscanf(buf, "%d", &delta_hosts) != 1)
3141		return -EINVAL;
3142	if (delta_hosts > 0) {
3143		do {
3144			sdebug_add_adapter();
3145		} while (--delta_hosts);
3146	} else if (delta_hosts < 0) {
3147		do {
3148			sdebug_remove_adapter();
3149		} while (++delta_hosts);
3150	}
3151	return count;
3152}
3153static DRIVER_ATTR_RW(add_host);
3154
3155static ssize_t vpd_use_hostno_show(struct device_driver *ddp, char *buf)
3156{
3157	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_vpd_use_hostno);
3158}
3159static ssize_t vpd_use_hostno_store(struct device_driver *ddp, const char *buf,
3160				    size_t count)
3161{
3162	int n;
3163
3164	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
3165		scsi_debug_vpd_use_hostno = n;
3166		return count;
3167	}
3168	return -EINVAL;
3169}
3170static DRIVER_ATTR_RW(vpd_use_hostno);
3171
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3172static ssize_t sector_size_show(struct device_driver *ddp, char *buf)
3173{
3174	return scnprintf(buf, PAGE_SIZE, "%u\n", scsi_debug_sector_size);
3175}
3176static DRIVER_ATTR_RO(sector_size);
3177
 
 
 
 
 
 
3178static ssize_t dix_show(struct device_driver *ddp, char *buf)
3179{
3180	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dix);
3181}
3182static DRIVER_ATTR_RO(dix);
3183
3184static ssize_t dif_show(struct device_driver *ddp, char *buf)
3185{
3186	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dif);
3187}
3188static DRIVER_ATTR_RO(dif);
3189
3190static ssize_t guard_show(struct device_driver *ddp, char *buf)
3191{
3192	return scnprintf(buf, PAGE_SIZE, "%u\n", scsi_debug_guard);
3193}
3194static DRIVER_ATTR_RO(guard);
3195
3196static ssize_t ato_show(struct device_driver *ddp, char *buf)
3197{
3198	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_ato);
3199}
3200static DRIVER_ATTR_RO(ato);
3201
3202static ssize_t map_show(struct device_driver *ddp, char *buf)
3203{
3204	ssize_t count;
3205
3206	if (!scsi_debug_lbp())
3207		return scnprintf(buf, PAGE_SIZE, "0-%u\n",
3208				 sdebug_store_sectors);
3209
3210	count = bitmap_scnlistprintf(buf, PAGE_SIZE, map_storep, map_size);
3211
3212	buf[count++] = '\n';
3213	buf[count++] = 0;
3214
3215	return count;
3216}
3217static DRIVER_ATTR_RO(map);
3218
3219static ssize_t removable_show(struct device_driver *ddp, char *buf)
3220{
3221	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_removable ? 1 : 0);
3222}
3223static ssize_t removable_store(struct device_driver *ddp, const char *buf,
3224			       size_t count)
3225{
3226	int n;
3227
3228	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
3229		scsi_debug_removable = (n > 0);
3230		return count;
3231	}
3232	return -EINVAL;
3233}
3234static DRIVER_ATTR_RW(removable);
3235
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3236/* Note: The following array creates attribute files in the
3237   /sys/bus/pseudo/drivers/scsi_debug directory. The advantage of these
3238   files (over those found in the /sys/module/scsi_debug/parameters
3239   directory) is that auxiliary actions can be triggered when an attribute
3240   is changed. For example see: sdebug_add_host_store() above.
3241 */
3242
3243static struct attribute *sdebug_drv_attrs[] = {
3244	&driver_attr_delay.attr,
3245	&driver_attr_opts.attr,
3246	&driver_attr_ptype.attr,
3247	&driver_attr_dsense.attr,
3248	&driver_attr_fake_rw.attr,
3249	&driver_attr_no_lun_0.attr,
3250	&driver_attr_num_tgts.attr,
3251	&driver_attr_dev_size_mb.attr,
3252	&driver_attr_num_parts.attr,
3253	&driver_attr_every_nth.attr,
3254	&driver_attr_max_luns.attr,
3255	&driver_attr_max_queue.attr,
3256	&driver_attr_no_uld.attr,
3257	&driver_attr_scsi_level.attr,
3258	&driver_attr_virtual_gb.attr,
3259	&driver_attr_add_host.attr,
3260	&driver_attr_vpd_use_hostno.attr,
3261	&driver_attr_sector_size.attr,
 
 
3262	&driver_attr_dix.attr,
3263	&driver_attr_dif.attr,
3264	&driver_attr_guard.attr,
3265	&driver_attr_ato.attr,
3266	&driver_attr_map.attr,
3267	&driver_attr_removable.attr,
 
 
 
 
3268	NULL,
3269};
3270ATTRIBUTE_GROUPS(sdebug_drv);
3271
3272static struct device *pseudo_primary;
3273
3274static int __init scsi_debug_init(void)
3275{
3276	unsigned long sz;
3277	int host_to_add;
3278	int k;
3279	int ret;
3280
3281	switch (scsi_debug_sector_size) {
 
 
 
 
 
 
 
 
3282	case  512:
3283	case 1024:
3284	case 2048:
3285	case 4096:
3286		break;
3287	default:
3288		printk(KERN_ERR "scsi_debug_init: invalid sector_size %d\n",
3289		       scsi_debug_sector_size);
3290		return -EINVAL;
3291	}
3292
3293	switch (scsi_debug_dif) {
3294
3295	case SD_DIF_TYPE0_PROTECTION:
3296	case SD_DIF_TYPE1_PROTECTION:
3297	case SD_DIF_TYPE2_PROTECTION:
3298	case SD_DIF_TYPE3_PROTECTION:
 
3299		break;
3300
3301	default:
3302		printk(KERN_ERR "scsi_debug_init: dif must be 0, 1, 2 or 3\n");
 
 
 
 
 
3303		return -EINVAL;
3304	}
3305
3306	if (scsi_debug_guard > 1) {
3307		printk(KERN_ERR "scsi_debug_init: guard must be 0 or 1\n");
3308		return -EINVAL;
3309	}
3310
3311	if (scsi_debug_ato > 1) {
3312		printk(KERN_ERR "scsi_debug_init: ato must be 0 or 1\n");
3313		return -EINVAL;
3314	}
 
 
 
 
3315
3316	if (scsi_debug_physblk_exp > 15) {
3317		printk(KERN_ERR "scsi_debug_init: invalid physblk_exp %u\n",
3318		       scsi_debug_physblk_exp);
3319		return -EINVAL;
3320	}
3321
3322	if (scsi_debug_lowest_aligned > 0x3fff) {
3323		printk(KERN_ERR "scsi_debug_init: lowest_aligned too big: %u\n",
3324		       scsi_debug_lowest_aligned);
3325		return -EINVAL;
3326	}
 
 
 
 
 
 
3327
3328	if (scsi_debug_dev_size_mb < 1)
3329		scsi_debug_dev_size_mb = 1;  /* force minimum 1 MB ramdisk */
3330	sz = (unsigned long)scsi_debug_dev_size_mb * 1048576;
3331	sdebug_store_sectors = sz / scsi_debug_sector_size;
3332	sdebug_capacity = get_sdebug_capacity();
3333
3334	/* play around with geometry, don't waste too much on track 0 */
3335	sdebug_heads = 8;
3336	sdebug_sectors_per = 32;
3337	if (scsi_debug_dev_size_mb >= 16)
3338		sdebug_heads = 32;
3339	else if (scsi_debug_dev_size_mb >= 256)
3340		sdebug_heads = 64;
 
 
3341	sdebug_cylinders_per = (unsigned long)sdebug_capacity /
3342			       (sdebug_sectors_per * sdebug_heads);
3343	if (sdebug_cylinders_per >= 1024) {
3344		/* other LLDs do this; implies >= 1GB ram disk ... */
3345		sdebug_heads = 255;
3346		sdebug_sectors_per = 63;
3347		sdebug_cylinders_per = (unsigned long)sdebug_capacity /
3348			       (sdebug_sectors_per * sdebug_heads);
3349	}
3350
3351	fake_storep = vmalloc(sz);
3352	if (NULL == fake_storep) {
3353		printk(KERN_ERR "scsi_debug_init: out of memory, 1\n");
3354		return -ENOMEM;
 
 
 
 
 
 
3355	}
3356	memset(fake_storep, 0, sz);
3357	if (scsi_debug_num_parts > 0)
3358		sdebug_build_parts(fake_storep, sz);
3359
3360	if (scsi_debug_dix) {
3361		int dif_size;
3362
3363		dif_size = sdebug_store_sectors * sizeof(struct sd_dif_tuple);
3364		dif_storep = vmalloc(dif_size);
3365
3366		printk(KERN_ERR "scsi_debug_init: dif_storep %u bytes @ %p\n",
3367		       dif_size, dif_storep);
3368
3369		if (dif_storep == NULL) {
3370			printk(KERN_ERR "scsi_debug_init: out of mem. (DIX)\n");
3371			ret = -ENOMEM;
3372			goto free_vm;
3373		}
3374
3375		memset(dif_storep, 0xff, dif_size);
3376	}
3377
3378	/* Logical Block Provisioning */
3379	if (scsi_debug_lbp()) {
3380		scsi_debug_unmap_max_blocks =
3381			clamp(scsi_debug_unmap_max_blocks, 0U, 0xffffffffU);
3382
3383		scsi_debug_unmap_max_desc =
3384			clamp(scsi_debug_unmap_max_desc, 0U, 256U);
3385
3386		scsi_debug_unmap_granularity =
3387			clamp(scsi_debug_unmap_granularity, 1U, 0xffffffffU);
3388
3389		if (scsi_debug_unmap_alignment &&
3390		    scsi_debug_unmap_granularity <=
3391		    scsi_debug_unmap_alignment) {
3392			printk(KERN_ERR
3393			       "%s: ERR: unmap_granularity <= unmap_alignment\n",
3394			       __func__);
3395			return -EINVAL;
3396		}
3397
3398		map_size = lba_to_map_index(sdebug_store_sectors - 1) + 1;
3399		map_storep = vmalloc(BITS_TO_LONGS(map_size) * sizeof(long));
3400
3401		printk(KERN_INFO "scsi_debug_init: %lu provisioning blocks\n",
3402		       map_size);
3403
3404		if (map_storep == NULL) {
3405			printk(KERN_ERR "scsi_debug_init: out of mem. (MAP)\n");
3406			ret = -ENOMEM;
3407			goto free_vm;
3408		}
3409
3410		bitmap_zero(map_storep, map_size);
3411
3412		/* Map first 1KB for partition table */
3413		if (scsi_debug_num_parts)
3414			map_region(0, 2);
3415	}
3416
3417	pseudo_primary = root_device_register("pseudo_0");
3418	if (IS_ERR(pseudo_primary)) {
3419		printk(KERN_WARNING "scsi_debug: root_device_register() error\n");
3420		ret = PTR_ERR(pseudo_primary);
3421		goto free_vm;
3422	}
3423	ret = bus_register(&pseudo_lld_bus);
3424	if (ret < 0) {
3425		printk(KERN_WARNING "scsi_debug: bus_register error: %d\n",
3426			ret);
3427		goto dev_unreg;
3428	}
3429	ret = driver_register(&sdebug_driverfs_driver);
3430	if (ret < 0) {
3431		printk(KERN_WARNING "scsi_debug: driver_register error: %d\n",
3432			ret);
3433		goto bus_unreg;
3434	}
3435
3436	init_all_queued();
3437
3438	host_to_add = scsi_debug_add_host;
3439        scsi_debug_add_host = 0;
3440
3441        for (k = 0; k < host_to_add; k++) {
3442                if (sdebug_add_adapter()) {
3443                        printk(KERN_ERR "scsi_debug_init: "
3444                               "sdebug_add_adapter failed k=%d\n", k);
3445                        break;
3446                }
3447        }
3448
3449	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts) {
3450		printk(KERN_INFO "scsi_debug_init: built %d host(s)\n",
3451		       scsi_debug_add_host);
3452	}
3453	return 0;
3454
3455bus_unreg:
3456	bus_unregister(&pseudo_lld_bus);
3457dev_unreg:
3458	root_device_unregister(pseudo_primary);
3459free_vm:
3460	if (map_storep)
3461		vfree(map_storep);
3462	if (dif_storep)
3463		vfree(dif_storep);
3464	vfree(fake_storep);
3465
 
3466	return ret;
3467}
3468
3469static void __exit scsi_debug_exit(void)
3470{
3471	int k = scsi_debug_add_host;
3472
3473	stop_all_queued();
 
3474	for (; k; k--)
3475		sdebug_remove_adapter();
3476	driver_unregister(&sdebug_driverfs_driver);
3477	bus_unregister(&pseudo_lld_bus);
3478	root_device_unregister(pseudo_primary);
3479
3480	if (dif_storep)
3481		vfree(dif_storep);
3482
3483	vfree(fake_storep);
 
3484}
3485
3486device_initcall(scsi_debug_init);
3487module_exit(scsi_debug_exit);
3488
3489static void sdebug_release_adapter(struct device * dev)
3490{
3491        struct sdebug_host_info *sdbg_host;
3492
3493	sdbg_host = to_sdebug_host(dev);
3494        kfree(sdbg_host);
3495}
3496
3497static int sdebug_add_adapter(void)
3498{
3499	int k, devs_per_host;
3500        int error = 0;
3501        struct sdebug_host_info *sdbg_host;
3502	struct sdebug_dev_info *sdbg_devinfo, *tmp;
3503
3504        sdbg_host = kzalloc(sizeof(*sdbg_host),GFP_KERNEL);
3505        if (NULL == sdbg_host) {
3506                printk(KERN_ERR "%s: out of memory at line %d\n",
3507                       __func__, __LINE__);
3508                return -ENOMEM;
3509        }
3510
3511        INIT_LIST_HEAD(&sdbg_host->dev_info_list);
3512
3513	devs_per_host = scsi_debug_num_tgts * scsi_debug_max_luns;
3514        for (k = 0; k < devs_per_host; k++) {
3515		sdbg_devinfo = sdebug_device_create(sdbg_host, GFP_KERNEL);
3516		if (!sdbg_devinfo) {
3517                        printk(KERN_ERR "%s: out of memory at line %d\n",
3518                               __func__, __LINE__);
3519                        error = -ENOMEM;
3520			goto clean;
3521                }
3522        }
3523
3524        spin_lock(&sdebug_host_list_lock);
3525        list_add_tail(&sdbg_host->host_list, &sdebug_host_list);
3526        spin_unlock(&sdebug_host_list_lock);
3527
3528        sdbg_host->dev.bus = &pseudo_lld_bus;
3529        sdbg_host->dev.parent = pseudo_primary;
3530        sdbg_host->dev.release = &sdebug_release_adapter;
3531        dev_set_name(&sdbg_host->dev, "adapter%d", scsi_debug_add_host);
3532
3533        error = device_register(&sdbg_host->dev);
3534
3535        if (error)
3536		goto clean;
3537
3538	++scsi_debug_add_host;
3539        return error;
3540
3541clean:
3542	list_for_each_entry_safe(sdbg_devinfo, tmp, &sdbg_host->dev_info_list,
3543				 dev_list) {
3544		list_del(&sdbg_devinfo->dev_list);
3545		kfree(sdbg_devinfo);
3546	}
3547
3548	kfree(sdbg_host);
3549        return error;
3550}
3551
3552static void sdebug_remove_adapter(void)
3553{
3554        struct sdebug_host_info * sdbg_host = NULL;
3555
3556        spin_lock(&sdebug_host_list_lock);
3557        if (!list_empty(&sdebug_host_list)) {
3558                sdbg_host = list_entry(sdebug_host_list.prev,
3559                                       struct sdebug_host_info, host_list);
3560		list_del(&sdbg_host->host_list);
3561	}
3562        spin_unlock(&sdebug_host_list_lock);
3563
3564	if (!sdbg_host)
3565		return;
3566
3567        device_unregister(&sdbg_host->dev);
3568        --scsi_debug_add_host;
3569}
3570
3571static
3572int scsi_debug_queuecommand_lck(struct scsi_cmnd *SCpnt, done_funct_t done)
3573{
3574	unsigned char *cmd = (unsigned char *) SCpnt->cmnd;
3575	int len, k;
3576	unsigned int num;
3577	unsigned long long lba;
3578	u32 ei_lba;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3579	int errsts = 0;
3580	int target = SCpnt->device->id;
3581	struct sdebug_dev_info *devip = NULL;
3582	int inj_recovered = 0;
3583	int inj_transport = 0;
3584	int inj_dif = 0;
3585	int inj_dix = 0;
3586	int delay_override = 0;
3587	int unmap = 0;
3588
3589	scsi_set_resid(SCpnt, 0);
3590	if ((SCSI_DEBUG_OPT_NOISE & scsi_debug_opts) && cmd) {
3591		printk(KERN_INFO "scsi_debug: cmd ");
3592		for (k = 0, len = SCpnt->cmd_len; k < len; ++k)
3593			printk("%02x ", (int)cmd[k]);
3594		printk("\n");
3595	}
3596
3597	if (target == SCpnt->device->host->hostt->this_id) {
3598		printk(KERN_INFO "scsi_debug: initiator's id used as "
3599		       "target!\n");
3600		return schedule_resp(SCpnt, NULL, done,
3601				     DID_NO_CONNECT << 16, 0);
3602	}
3603
3604	if ((SCpnt->device->lun >= scsi_debug_max_luns) &&
3605	    (SCpnt->device->lun != SAM2_WLUN_REPORT_LUNS))
3606		return schedule_resp(SCpnt, NULL, done,
3607				     DID_NO_CONNECT << 16, 0);
3608	devip = devInfoReg(SCpnt->device);
3609	if (NULL == devip)
3610		return schedule_resp(SCpnt, NULL, done,
3611				     DID_NO_CONNECT << 16, 0);
3612
3613	if ((scsi_debug_every_nth != 0) &&
3614	    (++scsi_debug_cmnd_count >= abs(scsi_debug_every_nth))) {
3615		scsi_debug_cmnd_count = 0;
3616		if (scsi_debug_every_nth < -1)
3617			scsi_debug_every_nth = -1;
3618		if (SCSI_DEBUG_OPT_TIMEOUT & scsi_debug_opts)
3619			return 0; /* ignore command causing timeout */
3620		else if (SCSI_DEBUG_OPT_MAC_TIMEOUT & scsi_debug_opts &&
3621			 scsi_medium_access_command(SCpnt))
3622			return 0; /* time out reads and writes */
3623		else if (SCSI_DEBUG_OPT_RECOVERED_ERR & scsi_debug_opts)
3624			inj_recovered = 1; /* to reads and writes below */
3625		else if (SCSI_DEBUG_OPT_TRANSPORT_ERR & scsi_debug_opts)
3626			inj_transport = 1; /* to reads and writes below */
3627		else if (SCSI_DEBUG_OPT_DIF_ERR & scsi_debug_opts)
3628			inj_dif = 1; /* to reads and writes below */
3629		else if (SCSI_DEBUG_OPT_DIX_ERR & scsi_debug_opts)
3630			inj_dix = 1; /* to reads and writes below */
3631	}
3632
3633	if (devip->wlun) {
3634		switch (*cmd) {
3635		case INQUIRY:
3636		case REQUEST_SENSE:
3637		case TEST_UNIT_READY:
3638		case REPORT_LUNS:
3639			break;  /* only allowable wlun commands */
3640		default:
3641			if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
3642				printk(KERN_INFO "scsi_debug: Opcode: 0x%x "
3643				       "not supported for wlun\n", *cmd);
3644			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3645					INVALID_OPCODE, 0);
3646			errsts = check_condition_result;
3647			return schedule_resp(SCpnt, devip, done, errsts,
3648					     0);
3649		}
 
 
 
 
 
 
 
3650	}
3651
3652	switch (*cmd) {
3653	case INQUIRY:     /* mandatory, ignore unit attention */
3654		delay_override = 1;
3655		errsts = resp_inquiry(SCpnt, target, devip);
3656		break;
3657	case REQUEST_SENSE:	/* mandatory, ignore unit attention */
3658		delay_override = 1;
3659		errsts = resp_requests(SCpnt, devip);
3660		break;
3661	case REZERO_UNIT:	/* actually this is REWIND for SSC */
3662	case START_STOP:
3663		errsts = resp_start_stop(SCpnt, devip);
3664		break;
3665	case ALLOW_MEDIUM_REMOVAL:
3666		errsts = check_readiness(SCpnt, 1, devip);
3667		if (errsts)
3668			break;
3669		if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
3670			printk(KERN_INFO "scsi_debug: Medium removal %s\n",
3671			       cmd[4] ? "inhibited" : "enabled");
3672		break;
3673	case SEND_DIAGNOSTIC:     /* mandatory */
3674		errsts = check_readiness(SCpnt, 1, devip);
3675		break;
3676	case TEST_UNIT_READY:     /* mandatory */
3677		delay_override = 1;
3678		errsts = check_readiness(SCpnt, 0, devip);
3679		break;
3680	case RESERVE:
3681		errsts = check_readiness(SCpnt, 1, devip);
3682		break;
3683	case RESERVE_10:
3684		errsts = check_readiness(SCpnt, 1, devip);
3685		break;
3686	case RELEASE:
3687		errsts = check_readiness(SCpnt, 1, devip);
3688		break;
3689	case RELEASE_10:
3690		errsts = check_readiness(SCpnt, 1, devip);
3691		break;
3692	case READ_CAPACITY:
3693		errsts = resp_readcap(SCpnt, devip);
3694		break;
3695	case SERVICE_ACTION_IN:
3696		if (cmd[1] == SAI_READ_CAPACITY_16)
3697			errsts = resp_readcap16(SCpnt, devip);
3698		else if (cmd[1] == SAI_GET_LBA_STATUS) {
3699
3700			if (scsi_debug_lbp() == 0) {
3701				mk_sense_buffer(devip, ILLEGAL_REQUEST,
3702						INVALID_COMMAND_OPCODE, 0);
3703				errsts = check_condition_result;
3704			} else
3705				errsts = resp_get_lba_status(SCpnt, devip);
3706		} else {
3707			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3708					INVALID_OPCODE, 0);
3709			errsts = check_condition_result;
3710		}
3711		break;
3712	case MAINTENANCE_IN:
3713		if (MI_REPORT_TARGET_PGS != cmd[1]) {
3714			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3715					INVALID_OPCODE, 0);
3716			errsts = check_condition_result;
3717			break;
3718		}
3719		errsts = resp_report_tgtpgs(SCpnt, devip);
3720		break;
3721	case READ_16:
3722	case READ_12:
3723	case READ_10:
3724		/* READ{10,12,16} and DIF Type 2 are natural enemies */
3725		if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
3726		    cmd[1] & 0xe0) {
3727			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3728					INVALID_COMMAND_OPCODE, 0);
3729			errsts = check_condition_result;
3730			break;
3731		}
3732
3733		if ((scsi_debug_dif == SD_DIF_TYPE1_PROTECTION ||
3734		     scsi_debug_dif == SD_DIF_TYPE3_PROTECTION) &&
3735		    (cmd[1] & 0xe0) == 0)
3736			printk(KERN_ERR "Unprotected RD/WR to DIF device\n");
3737
3738		/* fall through */
3739	case READ_6:
3740read:
3741		errsts = check_readiness(SCpnt, 0, devip);
3742		if (errsts)
3743			break;
3744		if (scsi_debug_fake_rw)
3745			break;
3746		get_data_transfer_info(cmd, &lba, &num, &ei_lba);
3747		errsts = resp_read(SCpnt, lba, num, devip, ei_lba);
3748		if (inj_recovered && (0 == errsts)) {
3749			mk_sense_buffer(devip, RECOVERED_ERROR,
3750					THRESHOLD_EXCEEDED, 0);
3751			errsts = check_condition_result;
3752		} else if (inj_transport && (0 == errsts)) {
3753			mk_sense_buffer(devip, ABORTED_COMMAND,
3754					TRANSPORT_PROBLEM, ACK_NAK_TO);
3755			errsts = check_condition_result;
3756		} else if (inj_dif && (0 == errsts)) {
3757			mk_sense_buffer(devip, ABORTED_COMMAND, 0x10, 1);
3758			errsts = illegal_condition_result;
3759		} else if (inj_dix && (0 == errsts)) {
3760			mk_sense_buffer(devip, ILLEGAL_REQUEST, 0x10, 1);
3761			errsts = illegal_condition_result;
3762		}
3763		break;
3764	case REPORT_LUNS:	/* mandatory, ignore unit attention */
3765		delay_override = 1;
3766		errsts = resp_report_luns(SCpnt, devip);
3767		break;
3768	case VERIFY:		/* 10 byte SBC-2 command */
3769		errsts = check_readiness(SCpnt, 0, devip);
3770		break;
3771	case WRITE_16:
3772	case WRITE_12:
3773	case WRITE_10:
3774		/* WRITE{10,12,16} and DIF Type 2 are natural enemies */
3775		if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
3776		    cmd[1] & 0xe0) {
3777			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3778					INVALID_COMMAND_OPCODE, 0);
3779			errsts = check_condition_result;
3780			break;
3781		}
3782
3783		if ((scsi_debug_dif == SD_DIF_TYPE1_PROTECTION ||
3784		     scsi_debug_dif == SD_DIF_TYPE3_PROTECTION) &&
3785		    (cmd[1] & 0xe0) == 0)
3786			printk(KERN_ERR "Unprotected RD/WR to DIF device\n");
3787
3788		/* fall through */
3789	case WRITE_6:
3790write:
3791		errsts = check_readiness(SCpnt, 0, devip);
3792		if (errsts)
3793			break;
3794		if (scsi_debug_fake_rw)
3795			break;
3796		get_data_transfer_info(cmd, &lba, &num, &ei_lba);
3797		errsts = resp_write(SCpnt, lba, num, devip, ei_lba);
3798		if (inj_recovered && (0 == errsts)) {
3799			mk_sense_buffer(devip, RECOVERED_ERROR,
3800					THRESHOLD_EXCEEDED, 0);
3801			errsts = check_condition_result;
3802		} else if (inj_dif && (0 == errsts)) {
3803			mk_sense_buffer(devip, ABORTED_COMMAND, 0x10, 1);
3804			errsts = illegal_condition_result;
3805		} else if (inj_dix && (0 == errsts)) {
3806			mk_sense_buffer(devip, ILLEGAL_REQUEST, 0x10, 1);
3807			errsts = illegal_condition_result;
3808		}
3809		break;
3810	case WRITE_SAME_16:
3811	case WRITE_SAME:
3812		if (cmd[1] & 0x8) {
3813			if ((*cmd == WRITE_SAME_16 && scsi_debug_lbpws == 0) ||
3814			    (*cmd == WRITE_SAME && scsi_debug_lbpws10 == 0)) {
3815				mk_sense_buffer(devip, ILLEGAL_REQUEST,
3816						INVALID_FIELD_IN_CDB, 0);
3817				errsts = check_condition_result;
3818			} else
3819				unmap = 1;
3820		}
3821		if (errsts)
3822			break;
3823		errsts = check_readiness(SCpnt, 0, devip);
3824		if (errsts)
3825			break;
3826		get_data_transfer_info(cmd, &lba, &num, &ei_lba);
3827		errsts = resp_write_same(SCpnt, lba, num, devip, ei_lba, unmap);
3828		break;
3829	case UNMAP:
3830		errsts = check_readiness(SCpnt, 0, devip);
3831		if (errsts)
3832			break;
3833
3834		if (scsi_debug_unmap_max_desc == 0 || scsi_debug_lbpu == 0) {
3835			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3836					INVALID_COMMAND_OPCODE, 0);
3837			errsts = check_condition_result;
3838		} else
3839			errsts = resp_unmap(SCpnt, devip);
3840		break;
3841	case MODE_SENSE:
3842	case MODE_SENSE_10:
3843		errsts = resp_mode_sense(SCpnt, target, devip);
3844		break;
3845	case MODE_SELECT:
3846		errsts = resp_mode_select(SCpnt, 1, devip);
3847		break;
3848	case MODE_SELECT_10:
3849		errsts = resp_mode_select(SCpnt, 0, devip);
3850		break;
3851	case LOG_SENSE:
3852		errsts = resp_log_sense(SCpnt, devip);
3853		break;
3854	case SYNCHRONIZE_CACHE:
3855		delay_override = 1;
3856		errsts = check_readiness(SCpnt, 0, devip);
3857		break;
3858	case WRITE_BUFFER:
3859		errsts = check_readiness(SCpnt, 1, devip);
3860		break;
3861	case XDWRITEREAD_10:
3862		if (!scsi_bidi_cmnd(SCpnt)) {
3863			mk_sense_buffer(devip, ILLEGAL_REQUEST,
3864					INVALID_FIELD_IN_CDB, 0);
3865			errsts = check_condition_result;
3866			break;
3867		}
3868
3869		errsts = check_readiness(SCpnt, 0, devip);
3870		if (errsts)
3871			break;
3872		if (scsi_debug_fake_rw)
3873			break;
3874		get_data_transfer_info(cmd, &lba, &num, &ei_lba);
3875		errsts = resp_read(SCpnt, lba, num, devip, ei_lba);
3876		if (errsts)
3877			break;
3878		errsts = resp_write(SCpnt, lba, num, devip, ei_lba);
3879		if (errsts)
3880			break;
3881		errsts = resp_xdwriteread(SCpnt, lba, num, devip);
3882		break;
3883	case VARIABLE_LENGTH_CMD:
3884		if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION) {
3885
3886			if ((cmd[10] & 0xe0) == 0)
3887				printk(KERN_ERR
3888				       "Unprotected RD/WR to DIF device\n");
3889
3890			if (cmd[9] == READ_32) {
3891				BUG_ON(SCpnt->cmd_len < 32);
3892				goto read;
3893			}
3894
3895			if (cmd[9] == WRITE_32) {
3896				BUG_ON(SCpnt->cmd_len < 32);
3897				goto write;
3898			}
3899		}
3900
3901		mk_sense_buffer(devip, ILLEGAL_REQUEST,
3902				INVALID_FIELD_IN_CDB, 0);
3903		errsts = check_condition_result;
3904		break;
3905
3906	default:
3907		if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
3908			printk(KERN_INFO "scsi_debug: Opcode: 0x%x not "
3909			       "supported\n", *cmd);
3910		errsts = check_readiness(SCpnt, 1, devip);
3911		if (errsts)
3912			break;	/* Unit attention takes precedence */
3913		mk_sense_buffer(devip, ILLEGAL_REQUEST, INVALID_OPCODE, 0);
 
 
 
 
 
 
3914		errsts = check_condition_result;
3915		break;
3916	}
3917	return schedule_resp(SCpnt, devip, done, errsts,
3918			     (delay_override ? 0 : scsi_debug_delay));
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3919}
3920
3921static DEF_SCSI_QCMD(scsi_debug_queuecommand)
3922
3923static struct scsi_host_template sdebug_driver_template = {
3924	.show_info =		scsi_debug_show_info,
3925	.write_info =		scsi_debug_write_info,
3926	.proc_name =		sdebug_proc_name,
3927	.name =			"SCSI DEBUG",
3928	.info =			scsi_debug_info,
3929	.slave_alloc =		scsi_debug_slave_alloc,
3930	.slave_configure =	scsi_debug_slave_configure,
3931	.slave_destroy =	scsi_debug_slave_destroy,
3932	.ioctl =		scsi_debug_ioctl,
3933	.queuecommand =		scsi_debug_queuecommand,
 
3934	.eh_abort_handler =	scsi_debug_abort,
3935	.eh_bus_reset_handler = scsi_debug_bus_reset,
3936	.eh_device_reset_handler = scsi_debug_device_reset,
 
 
3937	.eh_host_reset_handler = scsi_debug_host_reset,
3938	.bios_param =		scsi_debug_biosparam,
3939	.can_queue =		SCSI_DEBUG_CANQUEUE,
3940	.this_id =		7,
3941	.sg_tablesize =		256,
3942	.cmd_per_lun =		16,
3943	.max_sectors =		0xffff,
3944	.use_clustering = 	DISABLE_CLUSTERING,
3945	.module =		THIS_MODULE,
 
3946};
3947
3948static int sdebug_driver_probe(struct device * dev)
3949{
3950        int error = 0;
3951        struct sdebug_host_info *sdbg_host;
3952        struct Scsi_Host *hpnt;
3953	int host_prot;
3954
3955	sdbg_host = to_sdebug_host(dev);
3956
3957	sdebug_driver_template.can_queue = scsi_debug_max_queue;
3958	if (scsi_debug_clustering)
3959		sdebug_driver_template.use_clustering = ENABLE_CLUSTERING;
3960	hpnt = scsi_host_alloc(&sdebug_driver_template, sizeof(sdbg_host));
3961	if (NULL == hpnt) {
3962		printk(KERN_ERR "%s: scsi_register failed\n", __func__);
3963		error = -ENODEV;
3964		return error;
3965	}
 
 
 
 
 
 
 
 
 
 
3966
3967        sdbg_host->shost = hpnt;
3968	*((struct sdebug_host_info **)hpnt->hostdata) = sdbg_host;
3969	if ((hpnt->this_id >= 0) && (scsi_debug_num_tgts > hpnt->this_id))
3970		hpnt->max_id = scsi_debug_num_tgts + 1;
3971	else
3972		hpnt->max_id = scsi_debug_num_tgts;
3973	hpnt->max_lun = SAM2_WLUN_REPORT_LUNS;	/* = scsi_debug_max_luns; */
 
3974
3975	host_prot = 0;
3976
3977	switch (scsi_debug_dif) {
3978
3979	case SD_DIF_TYPE1_PROTECTION:
3980		host_prot = SHOST_DIF_TYPE1_PROTECTION;
3981		if (scsi_debug_dix)
3982			host_prot |= SHOST_DIX_TYPE1_PROTECTION;
3983		break;
3984
3985	case SD_DIF_TYPE2_PROTECTION:
3986		host_prot = SHOST_DIF_TYPE2_PROTECTION;
3987		if (scsi_debug_dix)
3988			host_prot |= SHOST_DIX_TYPE2_PROTECTION;
3989		break;
3990
3991	case SD_DIF_TYPE3_PROTECTION:
3992		host_prot = SHOST_DIF_TYPE3_PROTECTION;
3993		if (scsi_debug_dix)
3994			host_prot |= SHOST_DIX_TYPE3_PROTECTION;
3995		break;
3996
3997	default:
3998		if (scsi_debug_dix)
3999			host_prot |= SHOST_DIX_TYPE0_PROTECTION;
4000		break;
4001	}
4002
4003	scsi_host_set_prot(hpnt, host_prot);
4004
4005	printk(KERN_INFO "scsi_debug: host protection%s%s%s%s%s%s%s\n",
4006	       (host_prot & SHOST_DIF_TYPE1_PROTECTION) ? " DIF1" : "",
4007	       (host_prot & SHOST_DIF_TYPE2_PROTECTION) ? " DIF2" : "",
4008	       (host_prot & SHOST_DIF_TYPE3_PROTECTION) ? " DIF3" : "",
4009	       (host_prot & SHOST_DIX_TYPE0_PROTECTION) ? " DIX0" : "",
4010	       (host_prot & SHOST_DIX_TYPE1_PROTECTION) ? " DIX1" : "",
4011	       (host_prot & SHOST_DIX_TYPE2_PROTECTION) ? " DIX2" : "",
4012	       (host_prot & SHOST_DIX_TYPE3_PROTECTION) ? " DIX3" : "");
 
4013
4014	if (scsi_debug_guard == 1)
4015		scsi_host_set_guard(hpnt, SHOST_DIX_GUARD_IP);
4016	else
4017		scsi_host_set_guard(hpnt, SHOST_DIX_GUARD_CRC);
4018
 
 
 
 
4019        error = scsi_add_host(hpnt, &sdbg_host->dev);
4020        if (error) {
4021                printk(KERN_ERR "%s: scsi_add_host failed\n", __func__);
4022                error = -ENODEV;
4023		scsi_host_put(hpnt);
4024        } else
4025		scsi_scan_host(hpnt);
4026
4027
4028        return error;
4029}
4030
4031static int sdebug_driver_remove(struct device * dev)
4032{
4033        struct sdebug_host_info *sdbg_host;
4034	struct sdebug_dev_info *sdbg_devinfo, *tmp;
4035
4036	sdbg_host = to_sdebug_host(dev);
4037
4038	if (!sdbg_host) {
4039		printk(KERN_ERR "%s: Unable to locate host info\n",
4040		       __func__);
4041		return -ENODEV;
4042	}
4043
4044        scsi_remove_host(sdbg_host->shost);
4045
4046	list_for_each_entry_safe(sdbg_devinfo, tmp, &sdbg_host->dev_info_list,
4047				 dev_list) {
4048                list_del(&sdbg_devinfo->dev_list);
4049                kfree(sdbg_devinfo);
4050        }
4051
4052        scsi_host_put(sdbg_host->shost);
4053        return 0;
4054}
4055
4056static int pseudo_lld_bus_match(struct device *dev,
4057				struct device_driver *dev_driver)
4058{
4059	return 1;
4060}
4061
4062static struct bus_type pseudo_lld_bus = {
4063	.name = "pseudo",
4064	.match = pseudo_lld_bus_match,
4065	.probe = sdebug_driver_probe,
4066	.remove = sdebug_driver_remove,
4067	.drv_groups = sdebug_drv_groups,
4068};
v4.10.11
   1/*
   2 * vvvvvvvvvvvvvvvvvvvvvvv Original vvvvvvvvvvvvvvvvvvvvvvvvvvvvvvv
   3 *  Copyright (C) 1992  Eric Youngdale
   4 *  Simulate a host adapter with 2 disks attached.  Do a lot of checking
   5 *  to make sure that we are not getting blocks mixed up, and PANIC if
   6 *  anything out of the ordinary is seen.
   7 * ^^^^^^^^^^^^^^^^^^^^^^^ Original ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
   8 *
   9 * Copyright (C) 2001 - 2016 Douglas Gilbert
 
 
 
  10 *
  11 * This program is free software; you can redistribute it and/or modify
  12 * it under the terms of the GNU General Public License as published by
  13 * the Free Software Foundation; either version 2, or (at your option)
  14 * any later version.
  15 *
  16 *  For documentation see http://sg.danny.cz/sg/sdebug26.html
  17 *
 
 
 
 
 
 
 
 
 
  18 */
  19
  20
  21#define pr_fmt(fmt) KBUILD_MODNAME ":%s: " fmt, __func__
  22
  23#include <linux/module.h>
  24
  25#include <linux/kernel.h>
  26#include <linux/errno.h>
  27#include <linux/jiffies.h>
  28#include <linux/slab.h>
  29#include <linux/types.h>
  30#include <linux/string.h>
  31#include <linux/genhd.h>
  32#include <linux/fs.h>
  33#include <linux/init.h>
  34#include <linux/proc_fs.h>
  35#include <linux/vmalloc.h>
  36#include <linux/moduleparam.h>
  37#include <linux/scatterlist.h>
  38#include <linux/blkdev.h>
  39#include <linux/crc-t10dif.h>
  40#include <linux/spinlock.h>
  41#include <linux/interrupt.h>
  42#include <linux/atomic.h>
  43#include <linux/hrtimer.h>
  44#include <linux/uuid.h>
  45#include <linux/t10-pi.h>
  46
  47#include <net/checksum.h>
  48
  49#include <asm/unaligned.h>
  50
  51#include <scsi/scsi.h>
  52#include <scsi/scsi_cmnd.h>
  53#include <scsi/scsi_device.h>
  54#include <scsi/scsi_host.h>
  55#include <scsi/scsicam.h>
  56#include <scsi/scsi_eh.h>
  57#include <scsi/scsi_tcq.h>
  58#include <scsi/scsi_dbg.h>
  59
  60#include "sd.h"
  61#include "scsi_logging.h"
  62
  63/* make sure inq_product_rev string corresponds to this version */
  64#define SDEBUG_VERSION "1.86"
  65static const char *sdebug_version_date = "20160430";
  66
  67#define MY_NAME "scsi_debug"
  68
  69/* Additional Sense Code (ASC) */
  70#define NO_ADDITIONAL_SENSE 0x0
  71#define LOGICAL_UNIT_NOT_READY 0x4
  72#define LOGICAL_UNIT_COMMUNICATION_FAILURE 0x8
  73#define UNRECOVERED_READ_ERR 0x11
  74#define PARAMETER_LIST_LENGTH_ERR 0x1a
  75#define INVALID_OPCODE 0x20
  76#define LBA_OUT_OF_RANGE 0x21
 
  77#define INVALID_FIELD_IN_CDB 0x24
  78#define INVALID_FIELD_IN_PARAM_LIST 0x26
  79#define UA_RESET_ASC 0x29
  80#define UA_CHANGED_ASC 0x2a
  81#define TARGET_CHANGED_ASC 0x3f
  82#define LUNS_CHANGED_ASCQ 0x0e
  83#define INSUFF_RES_ASC 0x55
  84#define INSUFF_RES_ASCQ 0x3
  85#define POWER_ON_RESET_ASCQ 0x0
  86#define BUS_RESET_ASCQ 0x2	/* scsi bus reset occurred */
  87#define MODE_CHANGED_ASCQ 0x1	/* mode parameters changed */
  88#define CAPACITY_CHANGED_ASCQ 0x9
  89#define SAVING_PARAMS_UNSUP 0x39
  90#define TRANSPORT_PROBLEM 0x4b
  91#define THRESHOLD_EXCEEDED 0x5d
  92#define LOW_POWER_COND_ON 0x5e
  93#define MISCOMPARE_VERIFY_ASC 0x1d
  94#define MICROCODE_CHANGED_ASCQ 0x1	/* with TARGET_CHANGED_ASC */
  95#define MICROCODE_CHANGED_WO_RESET_ASCQ 0x16
  96
  97/* Additional Sense Code Qualifier (ASCQ) */
  98#define ACK_NAK_TO 0x3
  99
 
 
 100/* Default values for driver parameters */
 101#define DEF_NUM_HOST   1
 102#define DEF_NUM_TGTS   1
 103#define DEF_MAX_LUNS   1
 104/* With these defaults, this driver will make 1 host with 1 target
 105 * (id 0) containing 1 logical unit (lun 0). That is 1 device.
 106 */
 107#define DEF_ATO 1
 108#define DEF_JDELAY   1		/* if > 0 unit is a jiffy */
 109#define DEF_DEV_SIZE_MB   8
 110#define DEF_DIF 0
 111#define DEF_DIX 0
 112#define DEF_D_SENSE   0
 113#define DEF_EVERY_NTH   0
 114#define DEF_FAKE_RW	0
 115#define DEF_GUARD 0
 116#define DEF_HOST_LOCK 0
 117#define DEF_LBPU 0
 118#define DEF_LBPWS 0
 119#define DEF_LBPWS10 0
 120#define DEF_LBPRZ 1
 121#define DEF_LOWEST_ALIGNED 0
 122#define DEF_NDELAY   0		/* if > 0 unit is a nanosecond */
 123#define DEF_NO_LUN_0   0
 124#define DEF_NUM_PARTS   0
 125#define DEF_OPTS   0
 126#define DEF_OPT_BLKS 1024
 127#define DEF_PHYSBLK_EXP 0
 128#define DEF_PTYPE   TYPE_DISK
 129#define DEF_REMOVABLE false
 130#define DEF_SCSI_LEVEL   7    /* INQUIRY, byte2 [6->SPC-4; 7->SPC-5] */
 131#define DEF_SECTOR_SIZE 512
 132#define DEF_UNMAP_ALIGNMENT 0
 133#define DEF_UNMAP_GRANULARITY 1
 134#define DEF_UNMAP_MAX_BLOCKS 0xFFFFFFFF
 135#define DEF_UNMAP_MAX_DESC 256
 136#define DEF_VIRTUAL_GB   0
 137#define DEF_VPD_USE_HOSTNO 1
 138#define DEF_WRITESAME_LENGTH 0xFFFF
 139#define DEF_STRICT 0
 140#define DEF_STATISTICS false
 141#define DEF_SUBMIT_QUEUES 1
 142#define DEF_UUID_CTL 0
 143#define JDELAY_OVERRIDDEN -9999
 144
 145#define SDEBUG_LUN_0_VAL 0
 146
 147/* bit mask values for sdebug_opts */
 148#define SDEBUG_OPT_NOISE		1
 149#define SDEBUG_OPT_MEDIUM_ERR		2
 150#define SDEBUG_OPT_TIMEOUT		4
 151#define SDEBUG_OPT_RECOVERED_ERR	8
 152#define SDEBUG_OPT_TRANSPORT_ERR	16
 153#define SDEBUG_OPT_DIF_ERR		32
 154#define SDEBUG_OPT_DIX_ERR		64
 155#define SDEBUG_OPT_MAC_TIMEOUT		128
 156#define SDEBUG_OPT_SHORT_TRANSFER	0x100
 157#define SDEBUG_OPT_Q_NOISE		0x200
 158#define SDEBUG_OPT_ALL_TSF		0x400
 159#define SDEBUG_OPT_RARE_TSF		0x800
 160#define SDEBUG_OPT_N_WCE		0x1000
 161#define SDEBUG_OPT_RESET_NOISE		0x2000
 162#define SDEBUG_OPT_NO_CDB_NOISE		0x4000
 163#define SDEBUG_OPT_ALL_NOISE (SDEBUG_OPT_NOISE | SDEBUG_OPT_Q_NOISE | \
 164			      SDEBUG_OPT_RESET_NOISE)
 165#define SDEBUG_OPT_ALL_INJECTING (SDEBUG_OPT_RECOVERED_ERR | \
 166				  SDEBUG_OPT_TRANSPORT_ERR | \
 167				  SDEBUG_OPT_DIF_ERR | SDEBUG_OPT_DIX_ERR | \
 168				  SDEBUG_OPT_SHORT_TRANSFER)
 169/* When "every_nth" > 0 then modulo "every_nth" commands:
 170 *   - a missing response is simulated if SDEBUG_OPT_TIMEOUT is set
 171 *   - a RECOVERED_ERROR is simulated on successful read and write
 172 *     commands if SDEBUG_OPT_RECOVERED_ERR is set.
 173 *   - a TRANSPORT_ERROR is simulated on successful read and write
 174 *     commands if SDEBUG_OPT_TRANSPORT_ERR is set.
 175 *
 176 * When "every_nth" < 0 then after "- every_nth" commands:
 177 *   - a missing response is simulated if SDEBUG_OPT_TIMEOUT is set
 178 *   - a RECOVERED_ERROR is simulated on successful read and write
 179 *     commands if SDEBUG_OPT_RECOVERED_ERR is set.
 180 *   - a TRANSPORT_ERROR is simulated on successful read and write
 181 *     commands if _DEBUG_OPT_TRANSPORT_ERR is set.
 182 * This will continue on every subsequent command until some other action
 183 * occurs (e.g. the user * writing a new value (other than -1 or 1) to
 184 * every_nth via sysfs).
 185 */
 186
 187/* As indicated in SAM-5 and SPC-4 Unit Attentions (UAs) are returned in
 188 * priority order. In the subset implemented here lower numbers have higher
 189 * priority. The UA numbers should be a sequence starting from 0 with
 190 * SDEBUG_NUM_UAS being 1 higher than the highest numbered UA. */
 191#define SDEBUG_UA_POR 0		/* Power on, reset, or bus device reset */
 192#define SDEBUG_UA_BUS_RESET 1
 193#define SDEBUG_UA_MODE_CHANGED 2
 194#define SDEBUG_UA_CAPACITY_CHANGED 3
 195#define SDEBUG_UA_LUNS_CHANGED 4
 196#define SDEBUG_UA_MICROCODE_CHANGED 5	/* simulate firmware change */
 197#define SDEBUG_UA_MICROCODE_CHANGED_WO_RESET 6
 198#define SDEBUG_NUM_UAS 7
 199
 200/* when 1==SDEBUG_OPT_MEDIUM_ERR, a medium error is simulated at this
 201 * sector on read commands: */
 202#define OPT_MEDIUM_ERR_ADDR   0x1234 /* that's sector 4660 in decimal */
 203#define OPT_MEDIUM_ERR_NUM    10     /* number of consecutive medium errs */
 204
 205/* If REPORT LUNS has luns >= 256 it can choose "flat space" (value 1)
 206 * or "peripheral device" addressing (value 0) */
 207#define SAM2_LUN_ADDRESS_METHOD 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 208
 209/* SDEBUG_CANQUEUE is the maximum number of commands that can be queued
 210 * (for response) per submit queue at one time. Can be reduced by max_queue
 211 * option. Command responses are not queued when jdelay=0 and ndelay=0. The
 212 * per-device DEF_CMD_PER_LUN can be changed via sysfs:
 213 * /sys/class/scsi_device/<h:c:t:l>/device/queue_depth
 214 * but cannot exceed SDEBUG_CANQUEUE .
 215 */
 216#define SDEBUG_CANQUEUE_WORDS  3	/* a WORD is bits in a long */
 217#define SDEBUG_CANQUEUE  (SDEBUG_CANQUEUE_WORDS * BITS_PER_LONG)
 218#define DEF_CMD_PER_LUN  255
 219
 220#define F_D_IN			1
 221#define F_D_OUT			2
 222#define F_D_OUT_MAYBE		4	/* WRITE SAME, NDOB bit */
 223#define F_D_UNKN		8
 224#define F_RL_WLUN_OK		0x10
 225#define F_SKIP_UA		0x20
 226#define F_DELAY_OVERR		0x40
 227#define F_SA_LOW		0x80	/* cdb byte 1, bits 4 to 0 */
 228#define F_SA_HIGH		0x100	/* as used by variable length cdbs */
 229#define F_INV_OP		0x200
 230#define F_FAKE_RW		0x400
 231#define F_M_ACCESS		0x800	/* media access */
 232
 233#define FF_RESPOND (F_RL_WLUN_OK | F_SKIP_UA | F_DELAY_OVERR)
 234#define FF_DIRECT_IO (F_M_ACCESS | F_FAKE_RW)
 235#define FF_SA (F_SA_HIGH | F_SA_LOW)
 236
 237#define SDEBUG_MAX_PARTS 4
 238
 239#define SDEBUG_MAX_CMD_LEN 32
 240
 
 
 
 
 
 
 241
 242struct sdebug_dev_info {
 243	struct list_head dev_list;
 
 244	unsigned int channel;
 245	unsigned int target;
 246	u64 lun;
 247	uuid_be lu_name;
 248	struct sdebug_host_info *sdbg_host;
 249	unsigned long uas_bm[1];
 250	atomic_t num_in_q;
 251	atomic_t stopped;
 252	bool used;
 253};
 254
 255struct sdebug_host_info {
 256	struct list_head host_list;
 257	struct Scsi_Host *shost;
 258	struct device dev;
 259	struct list_head dev_info_list;
 260};
 261
 262#define to_sdebug_host(d)	\
 263	container_of(d, struct sdebug_host_info, dev)
 264
 265struct sdebug_defer {
 266	struct hrtimer hrt;
 267	struct execute_work ew;
 268	int sqa_idx;	/* index of sdebug_queue array */
 269	int qc_idx;	/* index of sdebug_queued_cmd array within sqa_idx */
 270	int issuing_cpu;
 271};
 272
 273struct sdebug_queued_cmd {
 274	/* corresponding bit set in in_use_bm[] in owning struct sdebug_queue
 275	 * instance indicates this slot is in use.
 276	 */
 277	struct sdebug_defer *sd_dp;
 278	struct scsi_cmnd *a_cmnd;
 279	unsigned int inj_recovered:1;
 280	unsigned int inj_transport:1;
 281	unsigned int inj_dif:1;
 282	unsigned int inj_dix:1;
 283	unsigned int inj_short:1;
 284};
 285
 286struct sdebug_queue {
 287	struct sdebug_queued_cmd qc_arr[SDEBUG_CANQUEUE];
 288	unsigned long in_use_bm[SDEBUG_CANQUEUE_WORDS];
 289	spinlock_t qc_lock;
 290	atomic_t blocked;	/* to temporarily stop more being queued */
 291};
 292
 293static atomic_t sdebug_cmnd_count;   /* number of incoming commands */
 294static atomic_t sdebug_completions;  /* count of deferred completions */
 295static atomic_t sdebug_miss_cpus;    /* submission + completion cpus differ */
 296static atomic_t sdebug_a_tsf;	     /* 'almost task set full' counter */
 297
 298struct opcode_info_t {
 299	u8 num_attached;	/* 0 if this is it (i.e. a leaf); use 0xff */
 300				/* for terminating element */
 301	u8 opcode;		/* if num_attached > 0, preferred */
 302	u16 sa;			/* service action */
 303	u32 flags;		/* OR-ed set of SDEB_F_* */
 304	int (*pfp)(struct scsi_cmnd *, struct sdebug_dev_info *);
 305	const struct opcode_info_t *arrp;  /* num_attached elements or NULL */
 306	u8 len_mask[16];	/* len=len_mask[0], then mask for cdb[1]... */
 307				/* ignore cdb bytes after position 15 */
 308};
 309
 310/* SCSI opcodes (first byte of cdb) of interest mapped onto these indexes */
 311enum sdeb_opcode_index {
 312	SDEB_I_INVALID_OPCODE =	0,
 313	SDEB_I_INQUIRY = 1,
 314	SDEB_I_REPORT_LUNS = 2,
 315	SDEB_I_REQUEST_SENSE = 3,
 316	SDEB_I_TEST_UNIT_READY = 4,
 317	SDEB_I_MODE_SENSE = 5,		/* 6, 10 */
 318	SDEB_I_MODE_SELECT = 6,		/* 6, 10 */
 319	SDEB_I_LOG_SENSE = 7,
 320	SDEB_I_READ_CAPACITY = 8,	/* 10; 16 is in SA_IN(16) */
 321	SDEB_I_READ = 9,		/* 6, 10, 12, 16 */
 322	SDEB_I_WRITE = 10,		/* 6, 10, 12, 16 */
 323	SDEB_I_START_STOP = 11,
 324	SDEB_I_SERV_ACT_IN = 12,	/* 12, 16 */
 325	SDEB_I_SERV_ACT_OUT = 13,	/* 12, 16 */
 326	SDEB_I_MAINT_IN = 14,
 327	SDEB_I_MAINT_OUT = 15,
 328	SDEB_I_VERIFY = 16,		/* 10 only */
 329	SDEB_I_VARIABLE_LEN = 17,
 330	SDEB_I_RESERVE = 18,		/* 6, 10 */
 331	SDEB_I_RELEASE = 19,		/* 6, 10 */
 332	SDEB_I_ALLOW_REMOVAL = 20,	/* PREVENT ALLOW MEDIUM REMOVAL */
 333	SDEB_I_REZERO_UNIT = 21,	/* REWIND in SSC */
 334	SDEB_I_ATA_PT = 22,		/* 12, 16 */
 335	SDEB_I_SEND_DIAG = 23,
 336	SDEB_I_UNMAP = 24,
 337	SDEB_I_XDWRITEREAD = 25,	/* 10 only */
 338	SDEB_I_WRITE_BUFFER = 26,
 339	SDEB_I_WRITE_SAME = 27,		/* 10, 16 */
 340	SDEB_I_SYNC_CACHE = 28,		/* 10 only */
 341	SDEB_I_COMP_WRITE = 29,
 342	SDEB_I_LAST_ELEMENT = 30,	/* keep this last */
 343};
 344
 345
 346static const unsigned char opcode_ind_arr[256] = {
 347/* 0x0; 0x0->0x1f: 6 byte cdbs */
 348	SDEB_I_TEST_UNIT_READY, SDEB_I_REZERO_UNIT, 0, SDEB_I_REQUEST_SENSE,
 349	    0, 0, 0, 0,
 350	SDEB_I_READ, 0, SDEB_I_WRITE, 0, 0, 0, 0, 0,
 351	0, 0, SDEB_I_INQUIRY, 0, 0, SDEB_I_MODE_SELECT, SDEB_I_RESERVE,
 352	    SDEB_I_RELEASE,
 353	0, 0, SDEB_I_MODE_SENSE, SDEB_I_START_STOP, 0, SDEB_I_SEND_DIAG,
 354	    SDEB_I_ALLOW_REMOVAL, 0,
 355/* 0x20; 0x20->0x3f: 10 byte cdbs */
 356	0, 0, 0, 0, 0, SDEB_I_READ_CAPACITY, 0, 0,
 357	SDEB_I_READ, 0, SDEB_I_WRITE, 0, 0, 0, 0, SDEB_I_VERIFY,
 358	0, 0, 0, 0, 0, SDEB_I_SYNC_CACHE, 0, 0,
 359	0, 0, 0, SDEB_I_WRITE_BUFFER, 0, 0, 0, 0,
 360/* 0x40; 0x40->0x5f: 10 byte cdbs */
 361	0, SDEB_I_WRITE_SAME, SDEB_I_UNMAP, 0, 0, 0, 0, 0,
 362	0, 0, 0, 0, 0, SDEB_I_LOG_SENSE, 0, 0,
 363	0, 0, 0, SDEB_I_XDWRITEREAD, 0, SDEB_I_MODE_SELECT, SDEB_I_RESERVE,
 364	    SDEB_I_RELEASE,
 365	0, 0, SDEB_I_MODE_SENSE, 0, 0, 0, 0, 0,
 366/* 0x60; 0x60->0x7d are reserved, 0x7e is "extended cdb" */
 367	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 368	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 369	0, SDEB_I_VARIABLE_LEN,
 370/* 0x80; 0x80->0x9f: 16 byte cdbs */
 371	0, 0, 0, 0, 0, SDEB_I_ATA_PT, 0, 0,
 372	SDEB_I_READ, SDEB_I_COMP_WRITE, SDEB_I_WRITE, 0, 0, 0, 0, 0,
 373	0, 0, 0, SDEB_I_WRITE_SAME, 0, 0, 0, 0,
 374	0, 0, 0, 0, 0, 0, SDEB_I_SERV_ACT_IN, SDEB_I_SERV_ACT_OUT,
 375/* 0xa0; 0xa0->0xbf: 12 byte cdbs */
 376	SDEB_I_REPORT_LUNS, SDEB_I_ATA_PT, 0, SDEB_I_MAINT_IN,
 377	     SDEB_I_MAINT_OUT, 0, 0, 0,
 378	SDEB_I_READ, SDEB_I_SERV_ACT_OUT, SDEB_I_WRITE, SDEB_I_SERV_ACT_IN,
 379	     0, 0, 0, 0,
 380	0, 0, 0, 0, 0, 0, 0, 0,
 381	0, 0, 0, 0, 0, 0, 0, 0,
 382/* 0xc0; 0xc0->0xff: vendor specific */
 383	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 384	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 385	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 386	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 387};
 388
 389static int resp_inquiry(struct scsi_cmnd *, struct sdebug_dev_info *);
 390static int resp_report_luns(struct scsi_cmnd *, struct sdebug_dev_info *);
 391static int resp_requests(struct scsi_cmnd *, struct sdebug_dev_info *);
 392static int resp_mode_sense(struct scsi_cmnd *, struct sdebug_dev_info *);
 393static int resp_mode_select(struct scsi_cmnd *, struct sdebug_dev_info *);
 394static int resp_log_sense(struct scsi_cmnd *, struct sdebug_dev_info *);
 395static int resp_readcap(struct scsi_cmnd *, struct sdebug_dev_info *);
 396static int resp_read_dt0(struct scsi_cmnd *, struct sdebug_dev_info *);
 397static int resp_write_dt0(struct scsi_cmnd *, struct sdebug_dev_info *);
 398static int resp_start_stop(struct scsi_cmnd *, struct sdebug_dev_info *);
 399static int resp_readcap16(struct scsi_cmnd *, struct sdebug_dev_info *);
 400static int resp_get_lba_status(struct scsi_cmnd *, struct sdebug_dev_info *);
 401static int resp_report_tgtpgs(struct scsi_cmnd *, struct sdebug_dev_info *);
 402static int resp_unmap(struct scsi_cmnd *, struct sdebug_dev_info *);
 403static int resp_rsup_opcodes(struct scsi_cmnd *, struct sdebug_dev_info *);
 404static int resp_rsup_tmfs(struct scsi_cmnd *, struct sdebug_dev_info *);
 405static int resp_write_same_10(struct scsi_cmnd *, struct sdebug_dev_info *);
 406static int resp_write_same_16(struct scsi_cmnd *, struct sdebug_dev_info *);
 407static int resp_xdwriteread_10(struct scsi_cmnd *, struct sdebug_dev_info *);
 408static int resp_comp_write(struct scsi_cmnd *, struct sdebug_dev_info *);
 409static int resp_write_buffer(struct scsi_cmnd *, struct sdebug_dev_info *);
 410
 411static const struct opcode_info_t msense_iarr[1] = {
 412	{0, 0x1a, 0, F_D_IN, NULL, NULL,
 413	    {6,  0xe8, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 414};
 415
 416static const struct opcode_info_t mselect_iarr[1] = {
 417	{0, 0x15, 0, F_D_OUT, NULL, NULL,
 418	    {6,  0xf1, 0, 0, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 419};
 420
 421static const struct opcode_info_t read_iarr[3] = {
 422	{0, 0x28, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, NULL,/* READ(10) */
 423	    {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7, 0, 0,
 424	     0, 0, 0, 0} },
 425	{0, 0x8, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, NULL, /* READ(6) */
 426	    {6,  0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 427	{0, 0xa8, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, NULL,/* READ(12) */
 428	    {12,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0x9f,
 429	     0xc7, 0, 0, 0, 0} },
 430};
 431
 432static const struct opcode_info_t write_iarr[3] = {
 433	{0, 0x2a, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, NULL,   /* 10 */
 434	    {10,  0xfb, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7, 0, 0,
 435	     0, 0, 0, 0} },
 436	{0, 0xa, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, NULL,    /* 6 */
 437	    {6,  0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 438	{0, 0xaa, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, NULL,   /* 12 */
 439	    {12,  0xfb, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0x9f,
 440	     0xc7, 0, 0, 0, 0} },
 441};
 442
 443static const struct opcode_info_t sa_in_iarr[1] = {
 444	{0, 0x9e, 0x12, F_SA_LOW | F_D_IN, resp_get_lba_status, NULL,
 445	    {16,  0x12, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
 446	     0xff, 0xff, 0xff, 0, 0xc7} },
 447};
 448
 449static const struct opcode_info_t vl_iarr[1] = {	/* VARIABLE LENGTH */
 450	{0, 0x7f, 0xb, F_SA_HIGH | F_D_OUT | FF_DIRECT_IO, resp_write_dt0,
 451	    NULL, {32,  0xc7, 0, 0, 0, 0, 0x1f, 0x18, 0x0, 0xb, 0xfa,
 452		   0, 0xff, 0xff, 0xff, 0xff} },	/* WRITE(32) */
 453};
 454
 455static const struct opcode_info_t maint_in_iarr[2] = {
 456	{0, 0xa3, 0xc, F_SA_LOW | F_D_IN, resp_rsup_opcodes, NULL,
 457	    {12,  0xc, 0x87, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0,
 458	     0xc7, 0, 0, 0, 0} },
 459	{0, 0xa3, 0xd, F_SA_LOW | F_D_IN, resp_rsup_tmfs, NULL,
 460	    {12,  0xd, 0x80, 0, 0, 0, 0xff, 0xff, 0xff, 0xff, 0, 0xc7, 0, 0,
 461	     0, 0} },
 462};
 463
 464static const struct opcode_info_t write_same_iarr[1] = {
 465	{0, 0x93, 0, F_D_OUT_MAYBE | FF_DIRECT_IO, resp_write_same_16, NULL,
 466	    {16,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
 467	     0xff, 0xff, 0xff, 0x1f, 0xc7} },
 468};
 469
 470static const struct opcode_info_t reserve_iarr[1] = {
 471	{0, 0x16, 0, F_D_OUT, NULL, NULL,	/* RESERVE(6) */
 472	    {6,  0x1f, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 473};
 474
 475static const struct opcode_info_t release_iarr[1] = {
 476	{0, 0x17, 0, F_D_OUT, NULL, NULL,	/* RELEASE(6) */
 477	    {6,  0x1f, 0xff, 0, 0, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 478};
 479
 480
 481/* This array is accessed via SDEB_I_* values. Make sure all are mapped,
 482 * plus the terminating elements for logic that scans this table such as
 483 * REPORT SUPPORTED OPERATION CODES. */
 484static const struct opcode_info_t opcode_info_arr[SDEB_I_LAST_ELEMENT + 1] = {
 485/* 0 */
 486	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL,
 487	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 488	{0, 0x12, 0, FF_RESPOND | F_D_IN, resp_inquiry, NULL,
 489	    {6,  0xe3, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 490	{0, 0xa0, 0, FF_RESPOND | F_D_IN, resp_report_luns, NULL,
 491	    {12,  0xe3, 0xff, 0, 0, 0, 0xff, 0xff, 0xff, 0xff, 0, 0xc7, 0, 0,
 492	     0, 0} },
 493	{0, 0x3, 0, FF_RESPOND | F_D_IN, resp_requests, NULL,
 494	    {6,  0xe1, 0, 0, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 495	{0, 0x0, 0, F_M_ACCESS | F_RL_WLUN_OK, NULL, NULL,/* TEST UNIT READY */
 496	    {6,  0, 0, 0, 0, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 497	{1, 0x5a, 0, F_D_IN, resp_mode_sense, msense_iarr,
 498	    {10,  0xf8, 0xff, 0xff, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0,
 499	     0} },
 500	{1, 0x55, 0, F_D_OUT, resp_mode_select, mselect_iarr,
 501	    {10,  0xf1, 0, 0, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0} },
 502	{0, 0x4d, 0, F_D_IN, resp_log_sense, NULL,
 503	    {10,  0xe3, 0xff, 0xff, 0, 0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0,
 504	     0, 0, 0} },
 505	{0, 0x25, 0, F_D_IN, resp_readcap, NULL,
 506	    {10,  0xe1, 0xff, 0xff, 0xff, 0xff, 0, 0, 0x1, 0xc7, 0, 0, 0, 0,
 507	     0, 0} },
 508	{3, 0x88, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, read_iarr,
 509	    {16,  0xfe, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
 510	     0xff, 0xff, 0xff, 0x9f, 0xc7} },		/* READ(16) */
 511/* 10 */
 512	{3, 0x8a, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, write_iarr,
 513	    {16,  0xfa, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
 514	     0xff, 0xff, 0xff, 0x9f, 0xc7} },		/* WRITE(16) */
 515	{0, 0x1b, 0, 0, resp_start_stop, NULL,		/* START STOP UNIT */
 516	    {6,  0x1, 0, 0xf, 0xf7, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 517	{1, 0x9e, 0x10, F_SA_LOW | F_D_IN, resp_readcap16, sa_in_iarr,
 518	    {16,  0x10, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
 519	     0xff, 0xff, 0xff, 0x1, 0xc7} },	/* READ CAPACITY(16) */
 520	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* SA OUT */
 521	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 522	{2, 0xa3, 0xa, F_SA_LOW | F_D_IN, resp_report_tgtpgs, maint_in_iarr,
 523	    {12,  0xea, 0, 0, 0, 0, 0xff, 0xff, 0xff, 0xff, 0, 0xc7, 0, 0, 0,
 524	     0} },
 525	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* MAINT OUT */
 526	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 527	{0, 0x2f, 0, F_D_OUT_MAYBE | FF_DIRECT_IO, NULL, NULL, /* VERIFY(10) */
 528	    {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xc7,
 529	     0, 0, 0, 0, 0, 0} },
 530	{1, 0x7f, 0x9, F_SA_HIGH | F_D_IN | FF_DIRECT_IO, resp_read_dt0,
 531	    vl_iarr, {32,  0xc7, 0, 0, 0, 0, 0x1f, 0x18, 0x0, 0x9, 0xfe, 0,
 532		      0xff, 0xff, 0xff, 0xff} },/* VARIABLE LENGTH, READ(32) */
 533	{1, 0x56, 0, F_D_OUT, NULL, reserve_iarr, /* RESERVE(10) */
 534	    {10,  0xff, 0xff, 0xff, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0,
 535	     0} },
 536	{1, 0x57, 0, F_D_OUT, NULL, release_iarr, /* RELEASE(10) */
 537	    {10,  0x13, 0xff, 0xff, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0,
 538	     0} },
 539/* 20 */
 540	{0, 0x1e, 0, 0, NULL, NULL, /* ALLOW REMOVAL */
 541	    {6,  0, 0, 0, 0x3, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 542	{0, 0x1, 0, 0, resp_start_stop, NULL, /* REWIND ?? */
 543	    {6,  0x1, 0, 0, 0, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 544	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* ATA_PT */
 545	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 546	{0, 0x1d, F_D_OUT, 0, NULL, NULL,	/* SEND DIAGNOSTIC */
 547	    {6,  0xf7, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 548	{0, 0x42, 0, F_D_OUT | FF_DIRECT_IO, resp_unmap, NULL, /* UNMAP */
 549	    {10,  0x1, 0, 0, 0, 0, 0x1f, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0} },
 550	{0, 0x53, 0, F_D_IN | F_D_OUT | FF_DIRECT_IO, resp_xdwriteread_10,
 551	    NULL, {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7,
 552		   0, 0, 0, 0, 0, 0} },
 553	{0, 0x3b, 0, F_D_OUT_MAYBE, resp_write_buffer, NULL,
 554	    {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0,
 555	     0, 0, 0, 0} },			/* WRITE_BUFFER */
 556	{1, 0x41, 0, F_D_OUT_MAYBE | FF_DIRECT_IO, resp_write_same_10,
 557	    write_same_iarr, {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff,
 558			      0xff, 0xc7, 0, 0, 0, 0, 0, 0} },
 559	{0, 0x35, 0, F_DELAY_OVERR | FF_DIRECT_IO, NULL, NULL, /* SYNC_CACHE */
 560	    {10,  0x7, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7, 0, 0,
 561	     0, 0, 0, 0} },
 562	{0, 0x89, 0, F_D_OUT | FF_DIRECT_IO, resp_comp_write, NULL,
 563	    {16,  0xf8, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0, 0,
 564	     0, 0xff, 0x1f, 0xc7} },		/* COMPARE AND WRITE */
 565
 566/* 30 */
 567	{0xff, 0, 0, 0, NULL, NULL,		/* terminating element */
 568	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
 569};
 
 570
 571static int sdebug_add_host = DEF_NUM_HOST;
 572static int sdebug_ato = DEF_ATO;
 573static int sdebug_jdelay = DEF_JDELAY;	/* if > 0 then unit is jiffies */
 574static int sdebug_dev_size_mb = DEF_DEV_SIZE_MB;
 575static int sdebug_dif = DEF_DIF;
 576static int sdebug_dix = DEF_DIX;
 577static int sdebug_dsense = DEF_D_SENSE;
 578static int sdebug_every_nth = DEF_EVERY_NTH;
 579static int sdebug_fake_rw = DEF_FAKE_RW;
 580static unsigned int sdebug_guard = DEF_GUARD;
 581static int sdebug_lowest_aligned = DEF_LOWEST_ALIGNED;
 582static int sdebug_max_luns = DEF_MAX_LUNS;
 583static int sdebug_max_queue = SDEBUG_CANQUEUE;	/* per submit queue */
 584static atomic_t retired_max_queue;	/* if > 0 then was prior max_queue */
 585static int sdebug_ndelay = DEF_NDELAY;	/* if > 0 then unit is nanoseconds */
 586static int sdebug_no_lun_0 = DEF_NO_LUN_0;
 587static int sdebug_no_uld;
 588static int sdebug_num_parts = DEF_NUM_PARTS;
 589static int sdebug_num_tgts = DEF_NUM_TGTS; /* targets per host */
 590static int sdebug_opt_blks = DEF_OPT_BLKS;
 591static int sdebug_opts = DEF_OPTS;
 592static int sdebug_physblk_exp = DEF_PHYSBLK_EXP;
 593static int sdebug_ptype = DEF_PTYPE; /* SCSI peripheral device type */
 594static int sdebug_scsi_level = DEF_SCSI_LEVEL;
 595static int sdebug_sector_size = DEF_SECTOR_SIZE;
 596static int sdebug_virtual_gb = DEF_VIRTUAL_GB;
 597static int sdebug_vpd_use_hostno = DEF_VPD_USE_HOSTNO;
 598static unsigned int sdebug_lbpu = DEF_LBPU;
 599static unsigned int sdebug_lbpws = DEF_LBPWS;
 600static unsigned int sdebug_lbpws10 = DEF_LBPWS10;
 601static unsigned int sdebug_lbprz = DEF_LBPRZ;
 602static unsigned int sdebug_unmap_alignment = DEF_UNMAP_ALIGNMENT;
 603static unsigned int sdebug_unmap_granularity = DEF_UNMAP_GRANULARITY;
 604static unsigned int sdebug_unmap_max_blocks = DEF_UNMAP_MAX_BLOCKS;
 605static unsigned int sdebug_unmap_max_desc = DEF_UNMAP_MAX_DESC;
 606static unsigned int sdebug_write_same_length = DEF_WRITESAME_LENGTH;
 607static int sdebug_uuid_ctl = DEF_UUID_CTL;
 608static bool sdebug_removable = DEF_REMOVABLE;
 609static bool sdebug_clustering;
 610static bool sdebug_host_lock = DEF_HOST_LOCK;
 611static bool sdebug_strict = DEF_STRICT;
 612static bool sdebug_any_injecting_opt;
 613static bool sdebug_verbose;
 614static bool have_dif_prot;
 615static bool sdebug_statistics = DEF_STATISTICS;
 616static bool sdebug_mq_active;
 617
 618static unsigned int sdebug_store_sectors;
 619static sector_t sdebug_capacity;	/* in sectors */
 620
 621/* old BIOS stuff, kernel may get rid of them but some mode sense pages
 622   may still need them */
 623static int sdebug_heads;		/* heads per disk */
 624static int sdebug_cylinders_per;	/* cylinders per surface */
 625static int sdebug_sectors_per;		/* sectors per cylinder */
 626
 627static LIST_HEAD(sdebug_host_list);
 628static DEFINE_SPINLOCK(sdebug_host_list_lock);
 629
 630static unsigned char *fake_storep;	/* ramdisk storage */
 631static struct t10_pi_tuple *dif_storep;	/* protection info */
 632static void *map_storep;		/* provisioning map */
 633
 634static unsigned long map_size;
 635static int num_aborts;
 636static int num_dev_resets;
 637static int num_target_resets;
 638static int num_bus_resets;
 639static int num_host_resets;
 640static int dix_writes;
 641static int dix_reads;
 642static int dif_errors;
 643
 644static int submit_queues = DEF_SUBMIT_QUEUES;  /* > 1 for multi-queue (mq) */
 645static struct sdebug_queue *sdebug_q_arr;  /* ptr to array of submit queues */
 646
 647static DEFINE_RWLOCK(atomic_rw);
 648
 649static char sdebug_proc_name[] = MY_NAME;
 650static const char *my_name = MY_NAME;
 651
 652static struct bus_type pseudo_lld_bus;
 653
 654static struct device_driver sdebug_driverfs_driver = {
 655	.name 		= sdebug_proc_name,
 656	.bus		= &pseudo_lld_bus,
 657};
 658
 659static const int check_condition_result =
 660		(DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION;
 661
 662static const int illegal_condition_result =
 663	(DRIVER_SENSE << 24) | (DID_ABORT << 16) | SAM_STAT_CHECK_CONDITION;
 664
 665static const int device_qfull_result =
 666	(DID_OK << 16) | (COMMAND_COMPLETE << 8) | SAM_STAT_TASK_SET_FULL;
 667
 668
 669/* Only do the extra work involved in logical block provisioning if one or
 670 * more of the lbpu, lbpws or lbpws10 parameters are given and we are doing
 671 * real reads and writes (i.e. not skipping them for speed).
 672 */
 673static inline bool scsi_debug_lbp(void)
 674{
 675	return 0 == sdebug_fake_rw &&
 676		(sdebug_lbpu || sdebug_lbpws || sdebug_lbpws10);
 677}
 678
 679static void *fake_store(unsigned long long lba)
 680{
 681	lba = do_div(lba, sdebug_store_sectors);
 682
 683	return fake_storep + lba * sdebug_sector_size;
 684}
 685
 686static struct t10_pi_tuple *dif_store(sector_t sector)
 687{
 688	sector = sector_div(sector, sdebug_store_sectors);
 689
 690	return dif_storep + sector;
 691}
 692
 
 
 
 693static void sdebug_max_tgts_luns(void)
 694{
 695	struct sdebug_host_info *sdbg_host;
 696	struct Scsi_Host *hpnt;
 697
 698	spin_lock(&sdebug_host_list_lock);
 699	list_for_each_entry(sdbg_host, &sdebug_host_list, host_list) {
 700		hpnt = sdbg_host->shost;
 701		if ((hpnt->this_id >= 0) &&
 702		    (sdebug_num_tgts > hpnt->this_id))
 703			hpnt->max_id = sdebug_num_tgts + 1;
 704		else
 705			hpnt->max_id = sdebug_num_tgts;
 706		/* sdebug_max_luns; */
 707		hpnt->max_lun = SCSI_W_LUN_REPORT_LUNS + 1;
 708	}
 709	spin_unlock(&sdebug_host_list_lock);
 710}
 711
 712enum sdeb_cmd_data {SDEB_IN_DATA = 0, SDEB_IN_CDB = 1};
 713
 714/* Set in_bit to -1 to indicate no bit position of invalid field */
 715static void mk_sense_invalid_fld(struct scsi_cmnd *scp,
 716				 enum sdeb_cmd_data c_d,
 717				 int in_byte, int in_bit)
 718{
 719	unsigned char *sbuff;
 720	u8 sks[4];
 721	int sl, asc;
 722
 723	sbuff = scp->sense_buffer;
 724	if (!sbuff) {
 725		sdev_printk(KERN_ERR, scp->device,
 726			    "%s: sense_buffer is NULL\n", __func__);
 727		return;
 728	}
 729	asc = c_d ? INVALID_FIELD_IN_CDB : INVALID_FIELD_IN_PARAM_LIST;
 730	memset(sbuff, 0, SCSI_SENSE_BUFFERSIZE);
 731	scsi_build_sense_buffer(sdebug_dsense, sbuff, ILLEGAL_REQUEST, asc, 0);
 732	memset(sks, 0, sizeof(sks));
 733	sks[0] = 0x80;
 734	if (c_d)
 735		sks[0] |= 0x40;
 736	if (in_bit >= 0) {
 737		sks[0] |= 0x8;
 738		sks[0] |= 0x7 & in_bit;
 739	}
 740	put_unaligned_be16(in_byte, sks + 1);
 741	if (sdebug_dsense) {
 742		sl = sbuff[7] + 8;
 743		sbuff[7] = sl;
 744		sbuff[sl] = 0x2;
 745		sbuff[sl + 1] = 0x6;
 746		memcpy(sbuff + sl + 4, sks, 3);
 747	} else
 748		memcpy(sbuff + 15, sks, 3);
 749	if (sdebug_verbose)
 750		sdev_printk(KERN_INFO, scp->device, "%s:  [sense_key,asc,ascq"
 751			    "]: [0x5,0x%x,0x0] %c byte=%d, bit=%d\n",
 752			    my_name, asc, c_d ? 'C' : 'D', in_byte, in_bit);
 753}
 754
 755static void mk_sense_buffer(struct scsi_cmnd *scp, int key, int asc, int asq)
 
 
 756{
 757	unsigned char *sbuff;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 758
 759	sbuff = scp->sense_buffer;
 760	if (!sbuff) {
 761		sdev_printk(KERN_ERR, scp->device,
 762			    "%s: sense_buffer is NULL\n", __func__);
 763		return;
 764	}
 765	memset(sbuff, 0, SCSI_SENSE_BUFFERSIZE);
 766
 767	scsi_build_sense_buffer(sdebug_dsense, sbuff, key, asc, asq);
 
 
 
 
 
 
 768
 769	if (sdebug_verbose)
 770		sdev_printk(KERN_INFO, scp->device,
 771			    "%s:  [sense_key,asc,ascq]: [0x%x,0x%x,0x%x]\n",
 772			    my_name, key, asc, asq);
 773}
 
 
 
 
 774
 775static void mk_sense_invalid_opcode(struct scsi_cmnd *scp)
 776{
 777	mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_OPCODE, 0);
 
 
 
 
 
 
 
 
 778}
 779
 780static int scsi_debug_ioctl(struct scsi_device *dev, int cmd, void __user *arg)
 781{
 782	if (sdebug_verbose) {
 783		if (0x1261 == cmd)
 784			sdev_printk(KERN_INFO, dev,
 785				    "%s: BLKFLSBUF [0x1261]\n", __func__);
 786		else if (0x5331 == cmd)
 787			sdev_printk(KERN_INFO, dev,
 788				    "%s: CDROM_GET_CAPABILITY [0x5331]\n",
 789				    __func__);
 790		else
 791			sdev_printk(KERN_INFO, dev, "%s: cmd=0x%x\n",
 792				    __func__, cmd);
 793	}
 794	return -EINVAL;
 795	/* return -ENOTTY; // correct return but upsets fdisk */
 796}
 797
 798static void clear_luns_changed_on_target(struct sdebug_dev_info *devip)
 
 799{
 800	struct sdebug_host_info *sdhp;
 801	struct sdebug_dev_info *dp;
 802
 803	spin_lock(&sdebug_host_list_lock);
 804	list_for_each_entry(sdhp, &sdebug_host_list, host_list) {
 805		list_for_each_entry(dp, &sdhp->dev_info_list, dev_list) {
 806			if ((devip->sdbg_host == dp->sdbg_host) &&
 807			    (devip->target == dp->target))
 808				clear_bit(SDEBUG_UA_LUNS_CHANGED, dp->uas_bm);
 809		}
 810	}
 811	spin_unlock(&sdebug_host_list_lock);
 812}
 813
 814static int make_ua(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
 815{
 816	int k;
 817
 818	k = find_first_bit(devip->uas_bm, SDEBUG_NUM_UAS);
 819	if (k != SDEBUG_NUM_UAS) {
 820		const char *cp = NULL;
 821
 822		switch (k) {
 823		case SDEBUG_UA_POR:
 824			mk_sense_buffer(scp, UNIT_ATTENTION, UA_RESET_ASC,
 825					POWER_ON_RESET_ASCQ);
 826			if (sdebug_verbose)
 827				cp = "power on reset";
 828			break;
 829		case SDEBUG_UA_BUS_RESET:
 830			mk_sense_buffer(scp, UNIT_ATTENTION, UA_RESET_ASC,
 831					BUS_RESET_ASCQ);
 832			if (sdebug_verbose)
 833				cp = "bus reset";
 834			break;
 835		case SDEBUG_UA_MODE_CHANGED:
 836			mk_sense_buffer(scp, UNIT_ATTENTION, UA_CHANGED_ASC,
 837					MODE_CHANGED_ASCQ);
 838			if (sdebug_verbose)
 839				cp = "mode parameters changed";
 840			break;
 841		case SDEBUG_UA_CAPACITY_CHANGED:
 842			mk_sense_buffer(scp, UNIT_ATTENTION, UA_CHANGED_ASC,
 843					CAPACITY_CHANGED_ASCQ);
 844			if (sdebug_verbose)
 845				cp = "capacity data changed";
 846			break;
 847		case SDEBUG_UA_MICROCODE_CHANGED:
 848			mk_sense_buffer(scp, UNIT_ATTENTION,
 849					TARGET_CHANGED_ASC,
 850					MICROCODE_CHANGED_ASCQ);
 851			if (sdebug_verbose)
 852				cp = "microcode has been changed";
 853			break;
 854		case SDEBUG_UA_MICROCODE_CHANGED_WO_RESET:
 855			mk_sense_buffer(scp, UNIT_ATTENTION,
 856					TARGET_CHANGED_ASC,
 857					MICROCODE_CHANGED_WO_RESET_ASCQ);
 858			if (sdebug_verbose)
 859				cp = "microcode has been changed without reset";
 860			break;
 861		case SDEBUG_UA_LUNS_CHANGED:
 862			/*
 863			 * SPC-3 behavior is to report a UNIT ATTENTION with
 864			 * ASC/ASCQ REPORTED LUNS DATA HAS CHANGED on every LUN
 865			 * on the target, until a REPORT LUNS command is
 866			 * received.  SPC-4 behavior is to report it only once.
 867			 * NOTE:  sdebug_scsi_level does not use the same
 868			 * values as struct scsi_device->scsi_level.
 869			 */
 870			if (sdebug_scsi_level >= 6)	/* SPC-4 and above */
 871				clear_luns_changed_on_target(devip);
 872			mk_sense_buffer(scp, UNIT_ATTENTION,
 873					TARGET_CHANGED_ASC,
 874					LUNS_CHANGED_ASCQ);
 875			if (sdebug_verbose)
 876				cp = "reported luns data has changed";
 877			break;
 878		default:
 879			pr_warn("unexpected unit attention code=%d\n", k);
 880			if (sdebug_verbose)
 881				cp = "unknown";
 882			break;
 883		}
 884		clear_bit(k, devip->uas_bm);
 885		if (sdebug_verbose)
 886			sdev_printk(KERN_INFO, scp->device,
 887				   "%s reports: Unit attention: %s\n",
 888				   my_name, cp);
 889		return check_condition_result;
 890	}
 891	return 0;
 892}
 893
 894/* Build SCSI "data-in" buffer. Returns 0 if ok else (DID_ERROR << 16). */
 895static int fill_from_dev_buffer(struct scsi_cmnd *scp, unsigned char *arr,
 896				int arr_len)
 897{
 898	int act_len;
 899	struct scsi_data_buffer *sdb = scsi_in(scp);
 900
 901	if (!sdb->length)
 902		return 0;
 903	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_FROM_DEVICE))
 904		return DID_ERROR << 16;
 905
 906	act_len = sg_copy_from_buffer(sdb->table.sgl, sdb->table.nents,
 907				      arr, arr_len);
 908	sdb->resid = scsi_bufflen(scp) - act_len;
 909
 910	return 0;
 911}
 912
 913/* Partial build of SCSI "data-in" buffer. Returns 0 if ok else
 914 * (DID_ERROR << 16). Can write to offset in data-in buffer. If multiple
 915 * calls, not required to write in ascending offset order. Assumes resid
 916 * set to scsi_bufflen() prior to any calls.
 917 */
 918static int p_fill_from_dev_buffer(struct scsi_cmnd *scp, const void *arr,
 919				  int arr_len, unsigned int off_dst)
 920{
 921	int act_len, n;
 922	struct scsi_data_buffer *sdb = scsi_in(scp);
 923	off_t skip = off_dst;
 924
 925	if (sdb->length <= off_dst)
 926		return 0;
 927	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_FROM_DEVICE))
 928		return DID_ERROR << 16;
 929
 930	act_len = sg_pcopy_from_buffer(sdb->table.sgl, sdb->table.nents,
 931				       arr, arr_len, skip);
 932	pr_debug("%s: off_dst=%u, scsi_bufflen=%u, act_len=%u, resid=%d\n",
 933		 __func__, off_dst, scsi_bufflen(scp), act_len, sdb->resid);
 934	n = (int)scsi_bufflen(scp) - ((int)off_dst + act_len);
 935	sdb->resid = min(sdb->resid, n);
 936	return 0;
 937}
 938
 939/* Fetches from SCSI "data-out" buffer. Returns number of bytes fetched into
 940 * 'arr' or -1 if error.
 941 */
 942static int fetch_to_dev_buffer(struct scsi_cmnd *scp, unsigned char *arr,
 943			       int arr_len)
 944{
 945	if (!scsi_bufflen(scp))
 946		return 0;
 947	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_TO_DEVICE))
 948		return -1;
 949
 950	return scsi_sg_copy_to_buffer(scp, arr, arr_len);
 951}
 952
 953
 954static const char * inq_vendor_id = "Linux   ";
 955static const char * inq_product_id = "scsi_debug      ";
 956static const char *inq_product_rev = "0186";	/* version less '.' */
 957/* Use some locally assigned NAAs for SAS addresses. */
 958static const u64 naa3_comp_a = 0x3222222000000000ULL;
 959static const u64 naa3_comp_b = 0x3333333000000000ULL;
 960static const u64 naa3_comp_c = 0x3111111000000000ULL;
 961
 962/* Device identification VPD page. Returns number of bytes placed in arr */
 963static int inquiry_vpd_83(unsigned char *arr, int port_group_id,
 964			  int target_dev_id, int dev_id_num,
 965			  const char *dev_id_str, int dev_id_str_len,
 966			  const uuid_be *lu_name)
 967{
 968	int num, port_a;
 969	char b[32];
 970
 971	port_a = target_dev_id + 1;
 972	/* T10 vendor identifier field format (faked) */
 973	arr[0] = 0x2;	/* ASCII */
 974	arr[1] = 0x1;
 975	arr[2] = 0x0;
 976	memcpy(&arr[4], inq_vendor_id, 8);
 977	memcpy(&arr[12], inq_product_id, 16);
 978	memcpy(&arr[28], dev_id_str, dev_id_str_len);
 979	num = 8 + 16 + dev_id_str_len;
 980	arr[3] = num;
 981	num += 4;
 982	if (dev_id_num >= 0) {
 983		if (sdebug_uuid_ctl) {
 984			/* Locally assigned UUID */
 985			arr[num++] = 0x1;  /* binary (not necessarily sas) */
 986			arr[num++] = 0xa;  /* PIV=0, lu, naa */
 987			arr[num++] = 0x0;
 988			arr[num++] = 0x12;
 989			arr[num++] = 0x10; /* uuid type=1, locally assigned */
 990			arr[num++] = 0x0;
 991			memcpy(arr + num, lu_name, 16);
 992			num += 16;
 993		} else {
 994			/* NAA-3, Logical unit identifier (binary) */
 995			arr[num++] = 0x1;  /* binary (not necessarily sas) */
 996			arr[num++] = 0x3;  /* PIV=0, lu, naa */
 997			arr[num++] = 0x0;
 998			arr[num++] = 0x8;
 999			put_unaligned_be64(naa3_comp_b + dev_id_num, arr + num);
1000			num += 8;
1001		}
1002		/* Target relative port number */
1003		arr[num++] = 0x61;	/* proto=sas, binary */
1004		arr[num++] = 0x94;	/* PIV=1, target port, rel port */
1005		arr[num++] = 0x0;	/* reserved */
1006		arr[num++] = 0x4;	/* length */
1007		arr[num++] = 0x0;	/* reserved */
1008		arr[num++] = 0x0;	/* reserved */
1009		arr[num++] = 0x0;
1010		arr[num++] = 0x1;	/* relative port A */
1011	}
1012	/* NAA-3, Target port identifier */
1013	arr[num++] = 0x61;	/* proto=sas, binary */
1014	arr[num++] = 0x93;	/* piv=1, target port, naa */
1015	arr[num++] = 0x0;
1016	arr[num++] = 0x8;
1017	put_unaligned_be64(naa3_comp_a + port_a, arr + num);
1018	num += 8;
1019	/* NAA-3, Target port group identifier */
 
 
 
 
 
 
1020	arr[num++] = 0x61;	/* proto=sas, binary */
1021	arr[num++] = 0x95;	/* piv=1, target port group id */
1022	arr[num++] = 0x0;
1023	arr[num++] = 0x4;
1024	arr[num++] = 0;
1025	arr[num++] = 0;
1026	put_unaligned_be16(port_group_id, arr + num);
1027	num += 2;
1028	/* NAA-3, Target device identifier */
1029	arr[num++] = 0x61;	/* proto=sas, binary */
1030	arr[num++] = 0xa3;	/* piv=1, target device, naa */
1031	arr[num++] = 0x0;
1032	arr[num++] = 0x8;
1033	put_unaligned_be64(naa3_comp_a + target_dev_id, arr + num);
1034	num += 8;
 
 
 
 
 
 
1035	/* SCSI name string: Target device identifier */
1036	arr[num++] = 0x63;	/* proto=sas, UTF-8 */
1037	arr[num++] = 0xa8;	/* piv=1, target device, SCSI name string */
1038	arr[num++] = 0x0;
1039	arr[num++] = 24;
1040	memcpy(arr + num, "naa.32222220", 12);
1041	num += 12;
1042	snprintf(b, sizeof(b), "%08X", target_dev_id);
1043	memcpy(arr + num, b, 8);
1044	num += 8;
1045	memset(arr + num, 0, 4);
1046	num += 4;
1047	return num;
1048}
1049
 
1050static unsigned char vpd84_data[] = {
1051/* from 4th byte */ 0x22,0x22,0x22,0x0,0xbb,0x0,
1052    0x22,0x22,0x22,0x0,0xbb,0x1,
1053    0x22,0x22,0x22,0x0,0xbb,0x2,
1054};
1055
1056/*  Software interface identification VPD page */
1057static int inquiry_vpd_84(unsigned char *arr)
1058{
1059	memcpy(arr, vpd84_data, sizeof(vpd84_data));
1060	return sizeof(vpd84_data);
1061}
1062
1063/* Management network addresses VPD page */
1064static int inquiry_vpd_85(unsigned char *arr)
1065{
1066	int num = 0;
1067	const char * na1 = "https://www.kernel.org/config";
1068	const char * na2 = "http://www.kernel.org/log";
1069	int plen, olen;
1070
1071	arr[num++] = 0x1;	/* lu, storage config */
1072	arr[num++] = 0x0;	/* reserved */
1073	arr[num++] = 0x0;
1074	olen = strlen(na1);
1075	plen = olen + 1;
1076	if (plen % 4)
1077		plen = ((plen / 4) + 1) * 4;
1078	arr[num++] = plen;	/* length, null termianted, padded */
1079	memcpy(arr + num, na1, olen);
1080	memset(arr + num + olen, 0, plen - olen);
1081	num += plen;
1082
1083	arr[num++] = 0x4;	/* lu, logging */
1084	arr[num++] = 0x0;	/* reserved */
1085	arr[num++] = 0x0;
1086	olen = strlen(na2);
1087	plen = olen + 1;
1088	if (plen % 4)
1089		plen = ((plen / 4) + 1) * 4;
1090	arr[num++] = plen;	/* length, null terminated, padded */
1091	memcpy(arr + num, na2, olen);
1092	memset(arr + num + olen, 0, plen - olen);
1093	num += plen;
1094
1095	return num;
1096}
1097
1098/* SCSI ports VPD page */
1099static int inquiry_vpd_88(unsigned char *arr, int target_dev_id)
1100{
1101	int num = 0;
1102	int port_a, port_b;
1103
1104	port_a = target_dev_id + 1;
1105	port_b = port_a + 1;
1106	arr[num++] = 0x0;	/* reserved */
1107	arr[num++] = 0x0;	/* reserved */
1108	arr[num++] = 0x0;
1109	arr[num++] = 0x1;	/* relative port 1 (primary) */
1110	memset(arr + num, 0, 6);
1111	num += 6;
1112	arr[num++] = 0x0;
1113	arr[num++] = 12;	/* length tp descriptor */
1114	/* naa-5 target port identifier (A) */
1115	arr[num++] = 0x61;	/* proto=sas, binary */
1116	arr[num++] = 0x93;	/* PIV=1, target port, NAA */
1117	arr[num++] = 0x0;	/* reserved */
1118	arr[num++] = 0x8;	/* length */
1119	put_unaligned_be64(naa3_comp_a + port_a, arr + num);
1120	num += 8;
 
 
 
 
 
 
 
1121	arr[num++] = 0x0;	/* reserved */
1122	arr[num++] = 0x0;	/* reserved */
1123	arr[num++] = 0x0;
1124	arr[num++] = 0x2;	/* relative port 2 (secondary) */
1125	memset(arr + num, 0, 6);
1126	num += 6;
1127	arr[num++] = 0x0;
1128	arr[num++] = 12;	/* length tp descriptor */
1129	/* naa-5 target port identifier (B) */
1130	arr[num++] = 0x61;	/* proto=sas, binary */
1131	arr[num++] = 0x93;	/* PIV=1, target port, NAA */
1132	arr[num++] = 0x0;	/* reserved */
1133	arr[num++] = 0x8;	/* length */
1134	put_unaligned_be64(naa3_comp_a + port_b, arr + num);
1135	num += 8;
 
 
 
 
 
 
1136
1137	return num;
1138}
1139
1140
1141static unsigned char vpd89_data[] = {
1142/* from 4th byte */ 0,0,0,0,
1143'l','i','n','u','x',' ',' ',' ',
1144'S','A','T',' ','s','c','s','i','_','d','e','b','u','g',' ',' ',
1145'1','2','3','4',
11460x34,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,
11470xec,0,0,0,
11480x5a,0xc,0xff,0x3f,0x37,0xc8,0x10,0,0,0,0,0,0x3f,0,0,0,
11490,0,0,0,0x58,0x58,0x58,0x58,0x58,0x58,0x58,0x58,0x20,0x20,0x20,0x20,
11500x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0,0,0,0x40,0x4,0,0x2e,0x33,
11510x38,0x31,0x20,0x20,0x20,0x20,0x54,0x53,0x38,0x33,0x30,0x30,0x33,0x31,
11520x53,0x41,
11530x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,
11540x20,0x20,
11550x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,
11560x10,0x80,
11570,0,0,0x2f,0,0,0,0x2,0,0x2,0x7,0,0xff,0xff,0x1,0,
11580x3f,0,0xc1,0xff,0x3e,0,0x10,0x1,0xb0,0xf8,0x50,0x9,0,0,0x7,0,
11590x3,0,0x78,0,0x78,0,0xf0,0,0x78,0,0,0,0,0,0,0,
11600,0,0,0,0,0,0,0,0x2,0,0,0,0,0,0,0,
11610x7e,0,0x1b,0,0x6b,0x34,0x1,0x7d,0x3,0x40,0x69,0x34,0x1,0x3c,0x3,0x40,
11620x7f,0x40,0,0,0,0,0xfe,0xfe,0,0,0,0,0,0xfe,0,0,
11630,0,0,0,0,0,0,0,0xb0,0xf8,0x50,0x9,0,0,0,0,
11640,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11650,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11660,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11670x1,0,0xb0,0xf8,0x50,0x9,0xb0,0xf8,0x50,0x9,0x20,0x20,0x2,0,0xb6,0x42,
11680,0x80,0x8a,0,0x6,0x3c,0xa,0x3c,0xff,0xff,0xc6,0x7,0,0x1,0,0x8,
11690xf0,0xf,0,0x10,0x2,0,0x30,0,0,0,0,0,0,0,0x6,0xfe,
11700,0,0x2,0,0x50,0,0x8a,0,0x4f,0x95,0,0,0x21,0,0xb,0,
11710,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11720,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11730,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11740,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11750,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11760,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11770,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11780,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11790,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11800,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11810,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
11820,0,0,0,0,0,0,0,0,0,0,0,0,0,0xa5,0x51,
1183};
1184
1185/* ATA Information VPD page */
1186static int inquiry_vpd_89(unsigned char *arr)
1187{
1188	memcpy(arr, vpd89_data, sizeof(vpd89_data));
1189	return sizeof(vpd89_data);
1190}
1191
1192
 
1193static unsigned char vpdb0_data[] = {
1194	/* from 4th byte */ 0,0,0,4, 0,0,0x4,0, 0,0,0,64,
1195	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1196	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1197	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1198};
1199
1200/* Block limits VPD page (SBC-3) */
1201static int inquiry_vpd_b0(unsigned char *arr)
1202{
1203	unsigned int gran;
1204
1205	memcpy(arr, vpdb0_data, sizeof(vpdb0_data));
1206
1207	/* Optimal transfer length granularity */
1208	gran = 1 << sdebug_physblk_exp;
1209	put_unaligned_be16(gran, arr + 2);
 
1210
1211	/* Maximum Transfer Length */
1212	if (sdebug_store_sectors > 0x400)
1213		put_unaligned_be32(sdebug_store_sectors, arr + 4);
 
 
 
 
1214
1215	/* Optimal Transfer Length */
1216	put_unaligned_be32(sdebug_opt_blks, &arr[8]);
1217
1218	if (sdebug_lbpu) {
1219		/* Maximum Unmap LBA Count */
1220		put_unaligned_be32(sdebug_unmap_max_blocks, &arr[16]);
1221
1222		/* Maximum Unmap Block Descriptor Count */
1223		put_unaligned_be32(sdebug_unmap_max_desc, &arr[20]);
1224	}
1225
1226	/* Unmap Granularity Alignment */
1227	if (sdebug_unmap_alignment) {
1228		put_unaligned_be32(sdebug_unmap_alignment, &arr[28]);
1229		arr[28] |= 0x80; /* UGAVALID */
1230	}
1231
1232	/* Optimal Unmap Granularity */
1233	put_unaligned_be32(sdebug_unmap_granularity, &arr[24]);
1234
1235	/* Maximum WRITE SAME Length */
1236	put_unaligned_be64(sdebug_write_same_length, &arr[32]);
1237
1238	return 0x3c; /* Mandatory page length for Logical Block Provisioning */
1239
1240	return sizeof(vpdb0_data);
1241}
1242
1243/* Block device characteristics VPD page (SBC-3) */
1244static int inquiry_vpd_b1(unsigned char *arr)
1245{
1246	memset(arr, 0, 0x3c);
1247	arr[0] = 0;
1248	arr[1] = 1;	/* non rotating medium (e.g. solid state) */
1249	arr[2] = 0;
1250	arr[3] = 5;	/* less than 1.8" */
1251
1252	return 0x3c;
1253}
1254
1255/* Logical block provisioning VPD page (SBC-4) */
1256static int inquiry_vpd_b2(unsigned char *arr)
1257{
1258	memset(arr, 0, 0x4);
1259	arr[0] = 0;			/* threshold exponent */
1260	if (sdebug_lbpu)
 
1261		arr[1] = 1 << 7;
1262	if (sdebug_lbpws)
 
1263		arr[1] |= 1 << 6;
1264	if (sdebug_lbpws10)
 
1265		arr[1] |= 1 << 5;
1266	if (sdebug_lbprz && scsi_debug_lbp())
1267		arr[1] |= (sdebug_lbprz & 0x7) << 2;  /* sbc4r07 and later */
1268	/* anc_sup=0; dp=0 (no provisioning group descriptor) */
1269	/* minimum_percentage=0; provisioning_type=0 (unknown) */
1270	/* threshold_percentage=0 */
1271	return 0x4;
1272}
1273
1274#define SDEBUG_LONG_INQ_SZ 96
1275#define SDEBUG_MAX_INQ_ARR_SZ 584
1276
1277static int resp_inquiry(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
 
1278{
1279	unsigned char pq_pdt;
1280	unsigned char * arr;
1281	unsigned char *cmd = scp->cmnd;
1282	int alloc_len, n, ret;
1283	bool have_wlun, is_disk;
1284
1285	alloc_len = get_unaligned_be16(cmd + 3);
1286	arr = kzalloc(SDEBUG_MAX_INQ_ARR_SZ, GFP_ATOMIC);
1287	if (! arr)
1288		return DID_REQUEUE << 16;
1289	is_disk = (sdebug_ptype == TYPE_DISK);
1290	have_wlun = scsi_is_wlun(scp->device->lun);
1291	if (have_wlun)
1292		pq_pdt = TYPE_WLUN;	/* present, wlun */
1293	else if (sdebug_no_lun_0 && (devip->lun == SDEBUG_LUN_0_VAL))
1294		pq_pdt = 0x7f;	/* not present, PQ=3, PDT=0x1f */
1295	else
1296		pq_pdt = (sdebug_ptype & 0x1f);
1297	arr[0] = pq_pdt;
1298	if (0x2 & cmd[1]) {  /* CMDDT bit set */
1299		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 1);
 
1300		kfree(arr);
1301		return check_condition_result;
1302	} else if (0x1 & cmd[1]) {  /* EVPD bit set */
1303		int lu_id_num, port_group_id, target_dev_id, len;
1304		char lu_id_str[6];
1305		int host_no = devip->sdbg_host->shost->host_no;
1306		
1307		port_group_id = (((host_no + 1) & 0x7f) << 8) +
1308		    (devip->channel & 0x7f);
1309		if (sdebug_vpd_use_hostno == 0)
1310			host_no = 0;
1311		lu_id_num = have_wlun ? -1 : (((host_no + 1) * 2000) +
1312			    (devip->target * 1000) + devip->lun);
1313		target_dev_id = ((host_no + 1) * 2000) +
1314				 (devip->target * 1000) - 3;
1315		len = scnprintf(lu_id_str, 6, "%d", lu_id_num);
1316		if (0 == cmd[2]) { /* supported vital product data pages */
1317			arr[1] = cmd[2];	/*sanity */
1318			n = 4;
1319			arr[n++] = 0x0;   /* this page */
1320			arr[n++] = 0x80;  /* unit serial number */
1321			arr[n++] = 0x83;  /* device identification */
1322			arr[n++] = 0x84;  /* software interface ident. */
1323			arr[n++] = 0x85;  /* management network addresses */
1324			arr[n++] = 0x86;  /* extended inquiry */
1325			arr[n++] = 0x87;  /* mode page policy */
1326			arr[n++] = 0x88;  /* SCSI ports */
1327			if (is_disk) {	  /* SBC only */
1328				arr[n++] = 0x89;  /* ATA information */
1329				arr[n++] = 0xb0;  /* Block limits */
1330				arr[n++] = 0xb1;  /* Block characteristics */
1331				arr[n++] = 0xb2;  /* Logical Block Prov */
1332			}
1333			arr[3] = n - 4;	  /* number of supported VPD pages */
1334		} else if (0x80 == cmd[2]) { /* unit serial number */
1335			arr[1] = cmd[2];	/*sanity */
1336			arr[3] = len;
1337			memcpy(&arr[4], lu_id_str, len);
1338		} else if (0x83 == cmd[2]) { /* device identification */
1339			arr[1] = cmd[2];	/*sanity */
1340			arr[3] = inquiry_vpd_83(&arr[4], port_group_id,
1341						target_dev_id, lu_id_num,
1342						lu_id_str, len,
1343						&devip->lu_name);
1344		} else if (0x84 == cmd[2]) { /* Software interface ident. */
1345			arr[1] = cmd[2];	/*sanity */
1346			arr[3] = inquiry_vpd_84(&arr[4]);
1347		} else if (0x85 == cmd[2]) { /* Management network addresses */
1348			arr[1] = cmd[2];	/*sanity */
1349			arr[3] = inquiry_vpd_85(&arr[4]);
1350		} else if (0x86 == cmd[2]) { /* extended inquiry */
1351			arr[1] = cmd[2];	/*sanity */
1352			arr[3] = 0x3c;	/* number of following entries */
1353			if (sdebug_dif == T10_PI_TYPE3_PROTECTION)
1354				arr[4] = 0x4;	/* SPT: GRD_CHK:1 */
1355			else if (have_dif_prot)
1356				arr[4] = 0x5;   /* SPT: GRD_CHK:1, REF_CHK:1 */
1357			else
1358				arr[4] = 0x0;   /* no protection stuff */
1359			arr[5] = 0x7;   /* head of q, ordered + simple q's */
1360		} else if (0x87 == cmd[2]) { /* mode page policy */
1361			arr[1] = cmd[2];	/*sanity */
1362			arr[3] = 0x8;	/* number of following entries */
1363			arr[4] = 0x2;	/* disconnect-reconnect mp */
1364			arr[6] = 0x80;	/* mlus, shared */
1365			arr[8] = 0x18;	 /* protocol specific lu */
1366			arr[10] = 0x82;	 /* mlus, per initiator port */
1367		} else if (0x88 == cmd[2]) { /* SCSI Ports */
1368			arr[1] = cmd[2];	/*sanity */
1369			arr[3] = inquiry_vpd_88(&arr[4], target_dev_id);
1370		} else if (is_disk && 0x89 == cmd[2]) { /* ATA information */
1371			arr[1] = cmd[2];        /*sanity */
1372			n = inquiry_vpd_89(&arr[4]);
1373			put_unaligned_be16(n, arr + 2);
1374		} else if (is_disk && 0xb0 == cmd[2]) { /* Block limits */
 
1375			arr[1] = cmd[2];        /*sanity */
1376			arr[3] = inquiry_vpd_b0(&arr[4]);
1377		} else if (is_disk && 0xb1 == cmd[2]) { /* Block char. */
1378			arr[1] = cmd[2];        /*sanity */
1379			arr[3] = inquiry_vpd_b1(&arr[4]);
1380		} else if (is_disk && 0xb2 == cmd[2]) { /* LB Prov. */
1381			arr[1] = cmd[2];        /*sanity */
1382			arr[3] = inquiry_vpd_b2(&arr[4]);
1383		} else {
1384			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, -1);
 
 
1385			kfree(arr);
1386			return check_condition_result;
1387		}
1388		len = min(get_unaligned_be16(arr + 2) + 4, alloc_len);
1389		ret = fill_from_dev_buffer(scp, arr,
1390			    min(len, SDEBUG_MAX_INQ_ARR_SZ));
1391		kfree(arr);
1392		return ret;
1393	}
1394	/* drops through here for a standard inquiry */
1395	arr[1] = sdebug_removable ? 0x80 : 0;	/* Removable disk */
1396	arr[2] = sdebug_scsi_level;
1397	arr[3] = 2;    /* response_data_format==2 */
1398	arr[4] = SDEBUG_LONG_INQ_SZ - 5;
1399	arr[5] = (int)have_dif_prot;	/* PROTECT bit */
1400	if (sdebug_vpd_use_hostno == 0)
1401		arr[5] = 0x10; /* claim: implicit TGPS */
1402	arr[6] = 0x10; /* claim: MultiP */
1403	/* arr[6] |= 0x40; ... claim: EncServ (enclosure services) */
1404	arr[7] = 0xa; /* claim: LINKED + CMDQUE */
1405	memcpy(&arr[8], inq_vendor_id, 8);
1406	memcpy(&arr[16], inq_product_id, 16);
1407	memcpy(&arr[32], inq_product_rev, 4);
1408	/* version descriptors (2 bytes each) follow */
1409	put_unaligned_be16(0xc0, arr + 58);   /* SAM-6 no version claimed */
1410	put_unaligned_be16(0x5c0, arr + 60);  /* SPC-5 no version claimed */
1411	n = 62;
1412	if (is_disk) {		/* SBC-4 no version claimed */
1413		put_unaligned_be16(0x600, arr + n);
1414		n += 2;
1415	} else if (sdebug_ptype == TYPE_TAPE) {	/* SSC-4 rev 3 */
1416		put_unaligned_be16(0x525, arr + n);
1417		n += 2;
1418	}
1419	put_unaligned_be16(0x2100, arr + n);	/* SPL-4 no version claimed */
1420	ret = fill_from_dev_buffer(scp, arr,
1421			    min(alloc_len, SDEBUG_LONG_INQ_SZ));
1422	kfree(arr);
1423	return ret;
1424}
1425
1426static unsigned char iec_m_pg[] = {0x1c, 0xa, 0x08, 0, 0, 0, 0, 0,
1427				   0, 0, 0x0, 0x0};
1428
1429static int resp_requests(struct scsi_cmnd * scp,
1430			 struct sdebug_dev_info * devip)
1431{
1432	unsigned char * sbuff;
1433	unsigned char *cmd = scp->cmnd;
1434	unsigned char arr[SCSI_SENSE_BUFFERSIZE];
1435	bool dsense;
1436	int len = 18;
1437
1438	memset(arr, 0, sizeof(arr));
1439	dsense = !!(cmd[1] & 1);
1440	sbuff = scp->sense_buffer;
 
 
1441	if ((iec_m_pg[2] & 0x4) && (6 == (iec_m_pg[3] & 0xf))) {
1442		if (dsense) {
1443			arr[0] = 0x72;
1444			arr[1] = 0x0;		/* NO_SENSE in sense_key */
1445			arr[2] = THRESHOLD_EXCEEDED;
1446			arr[3] = 0xff;		/* TEST set and MRIE==6 */
1447			len = 8;
1448		} else {
1449			arr[0] = 0x70;
1450			arr[2] = 0x0;		/* NO_SENSE in sense_key */
1451			arr[7] = 0xa;   	/* 18 byte sense buffer */
1452			arr[12] = THRESHOLD_EXCEEDED;
1453			arr[13] = 0xff;		/* TEST set and MRIE==6 */
1454		}
1455	} else {
1456		memcpy(arr, sbuff, SCSI_SENSE_BUFFERSIZE);
1457		if (arr[0] >= 0x70 && dsense == sdebug_dsense)
1458			;	/* have sense and formats match */
1459		else if (arr[0] <= 0x70) {
1460			if (dsense) {
1461				memset(arr, 0, 8);
1462				arr[0] = 0x72;
1463				len = 8;
1464			} else {
1465				memset(arr, 0, 18);
1466				arr[0] = 0x70;
1467				arr[7] = 0xa;
1468			}
1469		} else if (dsense) {
1470			memset(arr, 0, 8);
1471			arr[0] = 0x72;
1472			arr[1] = sbuff[2];     /* sense key */
1473			arr[2] = sbuff[12];    /* asc */
1474			arr[3] = sbuff[13];    /* ascq */
1475			len = 8;
1476		} else {
1477			memset(arr, 0, 18);
1478			arr[0] = 0x70;
1479			arr[2] = sbuff[1];
1480			arr[7] = 0xa;
1481			arr[12] = sbuff[1];
1482			arr[13] = sbuff[3];
1483		}
1484
1485	}
1486	mk_sense_buffer(scp, 0, NO_ADDITIONAL_SENSE, 0);
1487	return fill_from_dev_buffer(scp, arr, len);
1488}
1489
1490static int resp_start_stop(struct scsi_cmnd * scp,
1491			   struct sdebug_dev_info * devip)
1492{
1493	unsigned char *cmd = scp->cmnd;
1494	int power_cond, stop;
1495
 
 
1496	power_cond = (cmd[4] & 0xf0) >> 4;
1497	if (power_cond) {
1498		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 4, 7);
 
1499		return check_condition_result;
1500	}
1501	stop = !(cmd[4] & 1);
1502	atomic_xchg(&devip->stopped, stop);
 
1503	return 0;
1504}
1505
1506static sector_t get_sdebug_capacity(void)
1507{
1508	static const unsigned int gibibyte = 1073741824;
1509
1510	if (sdebug_virtual_gb > 0)
1511		return (sector_t)sdebug_virtual_gb *
1512			(gibibyte / sdebug_sector_size);
1513	else
1514		return sdebug_store_sectors;
1515}
1516
1517#define SDEBUG_READCAP_ARR_SZ 8
1518static int resp_readcap(struct scsi_cmnd * scp,
1519			struct sdebug_dev_info * devip)
1520{
1521	unsigned char arr[SDEBUG_READCAP_ARR_SZ];
1522	unsigned int capac;
 
1523
 
 
1524	/* following just in case virtual_gb changed */
1525	sdebug_capacity = get_sdebug_capacity();
1526	memset(arr, 0, SDEBUG_READCAP_ARR_SZ);
1527	if (sdebug_capacity < 0xffffffff) {
1528		capac = (unsigned int)sdebug_capacity - 1;
1529		put_unaligned_be32(capac, arr + 0);
1530	} else
1531		put_unaligned_be32(0xffffffff, arr + 0);
1532	put_unaligned_be16(sdebug_sector_size, arr + 6);
 
 
 
 
 
 
 
 
1533	return fill_from_dev_buffer(scp, arr, SDEBUG_READCAP_ARR_SZ);
1534}
1535
1536#define SDEBUG_READCAP16_ARR_SZ 32
1537static int resp_readcap16(struct scsi_cmnd * scp,
1538			  struct sdebug_dev_info * devip)
1539{
1540	unsigned char *cmd = scp->cmnd;
1541	unsigned char arr[SDEBUG_READCAP16_ARR_SZ];
1542	int alloc_len;
 
1543
1544	alloc_len = get_unaligned_be32(cmd + 10);
 
 
 
1545	/* following just in case virtual_gb changed */
1546	sdebug_capacity = get_sdebug_capacity();
1547	memset(arr, 0, SDEBUG_READCAP16_ARR_SZ);
1548	put_unaligned_be64((u64)(sdebug_capacity - 1), arr + 0);
1549	put_unaligned_be32(sdebug_sector_size, arr + 8);
1550	arr[13] = sdebug_physblk_exp & 0xf;
1551	arr[14] = (sdebug_lowest_aligned >> 8) & 0x3f;
 
 
 
 
 
1552
1553	if (scsi_debug_lbp()) {
1554		arr[14] |= 0x80; /* LBPME */
1555		/* from sbc4r07, this LBPRZ field is 1 bit, but the LBPRZ in
1556		 * the LB Provisioning VPD page is 3 bits. Note that lbprz=2
1557		 * in the wider field maps to 0 in this field.
1558		 */
1559		if (sdebug_lbprz & 1)	/* precisely what the draft requires */
1560			arr[14] |= 0x40;
1561	}
1562
1563	arr[15] = sdebug_lowest_aligned & 0xff;
1564
1565	if (have_dif_prot) {
1566		arr[12] = (sdebug_dif - 1) << 1; /* P_TYPE */
1567		arr[12] |= 1; /* PROT_EN */
1568	}
1569
1570	return fill_from_dev_buffer(scp, arr,
1571				    min(alloc_len, SDEBUG_READCAP16_ARR_SZ));
1572}
1573
1574#define SDEBUG_MAX_TGTPGS_ARR_SZ 1412
1575
1576static int resp_report_tgtpgs(struct scsi_cmnd * scp,
1577			      struct sdebug_dev_info * devip)
1578{
1579	unsigned char *cmd = scp->cmnd;
1580	unsigned char * arr;
1581	int host_no = devip->sdbg_host->shost->host_no;
1582	int n, ret, alen, rlen;
1583	int port_group_a, port_group_b, port_a, port_b;
1584
1585	alen = get_unaligned_be32(cmd + 6);
 
 
1586	arr = kzalloc(SDEBUG_MAX_TGTPGS_ARR_SZ, GFP_ATOMIC);
1587	if (! arr)
1588		return DID_REQUEUE << 16;
1589	/*
1590	 * EVPD page 0x88 states we have two ports, one
1591	 * real and a fake port with no device connected.
1592	 * So we create two port groups with one port each
1593	 * and set the group with port B to unavailable.
1594	 */
1595	port_a = 0x1; /* relative port A */
1596	port_b = 0x2; /* relative port B */
1597	port_group_a = (((host_no + 1) & 0x7f) << 8) +
1598			(devip->channel & 0x7f);
1599	port_group_b = (((host_no + 1) & 0x7f) << 8) +
1600			(devip->channel & 0x7f) + 0x80;
1601
1602	/*
1603	 * The asymmetric access state is cycled according to the host_id.
1604	 */
1605	n = 4;
1606	if (sdebug_vpd_use_hostno == 0) {
1607		arr[n++] = host_no % 3; /* Asymm access state */
1608		arr[n++] = 0x0F; /* claim: all states are supported */
1609	} else {
1610		arr[n++] = 0x0; /* Active/Optimized path */
1611		arr[n++] = 0x01; /* only support active/optimized paths */
1612	}
1613	put_unaligned_be16(port_group_a, arr + n);
1614	n += 2;
1615	arr[n++] = 0;    /* Reserved */
1616	arr[n++] = 0;    /* Status code */
1617	arr[n++] = 0;    /* Vendor unique */
1618	arr[n++] = 0x1;  /* One port per group */
1619	arr[n++] = 0;    /* Reserved */
1620	arr[n++] = 0;    /* Reserved */
1621	put_unaligned_be16(port_a, arr + n);
1622	n += 2;
1623	arr[n++] = 3;    /* Port unavailable */
1624	arr[n++] = 0x08; /* claim: only unavailalbe paths are supported */
1625	put_unaligned_be16(port_group_b, arr + n);
1626	n += 2;
1627	arr[n++] = 0;    /* Reserved */
1628	arr[n++] = 0;    /* Status code */
1629	arr[n++] = 0;    /* Vendor unique */
1630	arr[n++] = 0x1;  /* One port per group */
1631	arr[n++] = 0;    /* Reserved */
1632	arr[n++] = 0;    /* Reserved */
1633	put_unaligned_be16(port_b, arr + n);
1634	n += 2;
1635
1636	rlen = n - 4;
1637	put_unaligned_be32(rlen, arr + 0);
 
 
 
1638
1639	/*
1640	 * Return the smallest value of either
1641	 * - The allocated length
1642	 * - The constructed command length
1643	 * - The maximum array size
1644	 */
1645	rlen = min(alen,n);
1646	ret = fill_from_dev_buffer(scp, arr,
1647				   min(rlen, SDEBUG_MAX_TGTPGS_ARR_SZ));
1648	kfree(arr);
1649	return ret;
1650}
1651
1652static int resp_rsup_opcodes(struct scsi_cmnd *scp,
1653			     struct sdebug_dev_info *devip)
1654{
1655	bool rctd;
1656	u8 reporting_opts, req_opcode, sdeb_i, supp;
1657	u16 req_sa, u;
1658	u32 alloc_len, a_len;
1659	int k, offset, len, errsts, count, bump, na;
1660	const struct opcode_info_t *oip;
1661	const struct opcode_info_t *r_oip;
1662	u8 *arr;
1663	u8 *cmd = scp->cmnd;
1664
1665	rctd = !!(cmd[2] & 0x80);
1666	reporting_opts = cmd[2] & 0x7;
1667	req_opcode = cmd[3];
1668	req_sa = get_unaligned_be16(cmd + 4);
1669	alloc_len = get_unaligned_be32(cmd + 6);
1670	if (alloc_len < 4 || alloc_len > 0xffff) {
1671		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 6, -1);
1672		return check_condition_result;
1673	}
1674	if (alloc_len > 8192)
1675		a_len = 8192;
1676	else
1677		a_len = alloc_len;
1678	arr = kzalloc((a_len < 256) ? 320 : a_len + 64, GFP_ATOMIC);
1679	if (NULL == arr) {
1680		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
1681				INSUFF_RES_ASCQ);
1682		return check_condition_result;
1683	}
1684	switch (reporting_opts) {
1685	case 0:	/* all commands */
1686		/* count number of commands */
1687		for (count = 0, oip = opcode_info_arr;
1688		     oip->num_attached != 0xff; ++oip) {
1689			if (F_INV_OP & oip->flags)
1690				continue;
1691			count += (oip->num_attached + 1);
1692		}
1693		bump = rctd ? 20 : 8;
1694		put_unaligned_be32(count * bump, arr);
1695		for (offset = 4, oip = opcode_info_arr;
1696		     oip->num_attached != 0xff && offset < a_len; ++oip) {
1697			if (F_INV_OP & oip->flags)
1698				continue;
1699			na = oip->num_attached;
1700			arr[offset] = oip->opcode;
1701			put_unaligned_be16(oip->sa, arr + offset + 2);
1702			if (rctd)
1703				arr[offset + 5] |= 0x2;
1704			if (FF_SA & oip->flags)
1705				arr[offset + 5] |= 0x1;
1706			put_unaligned_be16(oip->len_mask[0], arr + offset + 6);
1707			if (rctd)
1708				put_unaligned_be16(0xa, arr + offset + 8);
1709			r_oip = oip;
1710			for (k = 0, oip = oip->arrp; k < na; ++k, ++oip) {
1711				if (F_INV_OP & oip->flags)
1712					continue;
1713				offset += bump;
1714				arr[offset] = oip->opcode;
1715				put_unaligned_be16(oip->sa, arr + offset + 2);
1716				if (rctd)
1717					arr[offset + 5] |= 0x2;
1718				if (FF_SA & oip->flags)
1719					arr[offset + 5] |= 0x1;
1720				put_unaligned_be16(oip->len_mask[0],
1721						   arr + offset + 6);
1722				if (rctd)
1723					put_unaligned_be16(0xa,
1724							   arr + offset + 8);
1725			}
1726			oip = r_oip;
1727			offset += bump;
1728		}
1729		break;
1730	case 1:	/* one command: opcode only */
1731	case 2:	/* one command: opcode plus service action */
1732	case 3:	/* one command: if sa==0 then opcode only else opcode+sa */
1733		sdeb_i = opcode_ind_arr[req_opcode];
1734		oip = &opcode_info_arr[sdeb_i];
1735		if (F_INV_OP & oip->flags) {
1736			supp = 1;
1737			offset = 4;
1738		} else {
1739			if (1 == reporting_opts) {
1740				if (FF_SA & oip->flags) {
1741					mk_sense_invalid_fld(scp, SDEB_IN_CDB,
1742							     2, 2);
1743					kfree(arr);
1744					return check_condition_result;
1745				}
1746				req_sa = 0;
1747			} else if (2 == reporting_opts &&
1748				   0 == (FF_SA & oip->flags)) {
1749				mk_sense_invalid_fld(scp, SDEB_IN_CDB, 4, -1);
1750				kfree(arr);	/* point at requested sa */
1751				return check_condition_result;
1752			}
1753			if (0 == (FF_SA & oip->flags) &&
1754			    req_opcode == oip->opcode)
1755				supp = 3;
1756			else if (0 == (FF_SA & oip->flags)) {
1757				na = oip->num_attached;
1758				for (k = 0, oip = oip->arrp; k < na;
1759				     ++k, ++oip) {
1760					if (req_opcode == oip->opcode)
1761						break;
1762				}
1763				supp = (k >= na) ? 1 : 3;
1764			} else if (req_sa != oip->sa) {
1765				na = oip->num_attached;
1766				for (k = 0, oip = oip->arrp; k < na;
1767				     ++k, ++oip) {
1768					if (req_sa == oip->sa)
1769						break;
1770				}
1771				supp = (k >= na) ? 1 : 3;
1772			} else
1773				supp = 3;
1774			if (3 == supp) {
1775				u = oip->len_mask[0];
1776				put_unaligned_be16(u, arr + 2);
1777				arr[4] = oip->opcode;
1778				for (k = 1; k < u; ++k)
1779					arr[4 + k] = (k < 16) ?
1780						 oip->len_mask[k] : 0xff;
1781				offset = 4 + u;
1782			} else
1783				offset = 4;
1784		}
1785		arr[1] = (rctd ? 0x80 : 0) | supp;
1786		if (rctd) {
1787			put_unaligned_be16(0xa, arr + offset);
1788			offset += 12;
1789		}
1790		break;
1791	default:
1792		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 2);
1793		kfree(arr);
1794		return check_condition_result;
1795	}
1796	offset = (offset < a_len) ? offset : a_len;
1797	len = (offset < alloc_len) ? offset : alloc_len;
1798	errsts = fill_from_dev_buffer(scp, arr, len);
1799	kfree(arr);
1800	return errsts;
1801}
1802
1803static int resp_rsup_tmfs(struct scsi_cmnd *scp,
1804			  struct sdebug_dev_info *devip)
1805{
1806	bool repd;
1807	u32 alloc_len, len;
1808	u8 arr[16];
1809	u8 *cmd = scp->cmnd;
1810
1811	memset(arr, 0, sizeof(arr));
1812	repd = !!(cmd[2] & 0x80);
1813	alloc_len = get_unaligned_be32(cmd + 6);
1814	if (alloc_len < 4) {
1815		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 6, -1);
1816		return check_condition_result;
1817	}
1818	arr[0] = 0xc8;		/* ATS | ATSS | LURS */
1819	arr[1] = 0x1;		/* ITNRS */
1820	if (repd) {
1821		arr[3] = 0xc;
1822		len = 16;
1823	} else
1824		len = 4;
1825
1826	len = (len < alloc_len) ? len : alloc_len;
1827	return fill_from_dev_buffer(scp, arr, len);
1828}
1829
1830/* <<Following mode page info copied from ST318451LW>> */
1831
1832static int resp_err_recov_pg(unsigned char * p, int pcontrol, int target)
1833{	/* Read-Write Error Recovery page for mode_sense */
1834	unsigned char err_recov_pg[] = {0x1, 0xa, 0xc0, 11, 240, 0, 0, 0,
1835					5, 0, 0xff, 0xff};
1836
1837	memcpy(p, err_recov_pg, sizeof(err_recov_pg));
1838	if (1 == pcontrol)
1839		memset(p + 2, 0, sizeof(err_recov_pg) - 2);
1840	return sizeof(err_recov_pg);
1841}
1842
1843static int resp_disconnect_pg(unsigned char * p, int pcontrol, int target)
1844{ 	/* Disconnect-Reconnect page for mode_sense */
1845	unsigned char disconnect_pg[] = {0x2, 0xe, 128, 128, 0, 10, 0, 0,
1846					 0, 0, 0, 0, 0, 0, 0, 0};
1847
1848	memcpy(p, disconnect_pg, sizeof(disconnect_pg));
1849	if (1 == pcontrol)
1850		memset(p + 2, 0, sizeof(disconnect_pg) - 2);
1851	return sizeof(disconnect_pg);
1852}
1853
1854static int resp_format_pg(unsigned char * p, int pcontrol, int target)
1855{       /* Format device page for mode_sense */
1856	unsigned char format_pg[] = {0x3, 0x16, 0, 0, 0, 0, 0, 0,
1857				     0, 0, 0, 0, 0, 0, 0, 0,
1858				     0, 0, 0, 0, 0x40, 0, 0, 0};
1859
1860	memcpy(p, format_pg, sizeof(format_pg));
1861	put_unaligned_be16(sdebug_sectors_per, p + 10);
1862	put_unaligned_be16(sdebug_sector_size, p + 12);
1863	if (sdebug_removable)
 
 
1864		p[20] |= 0x20; /* should agree with INQUIRY */
1865	if (1 == pcontrol)
1866		memset(p + 2, 0, sizeof(format_pg) - 2);
1867	return sizeof(format_pg);
1868}
1869
1870static unsigned char caching_pg[] = {0x8, 18, 0x14, 0, 0xff, 0xff, 0, 0,
1871				     0xff, 0xff, 0xff, 0xff, 0x80, 0x14, 0, 0,
1872				     0, 0, 0, 0};
1873
1874static int resp_caching_pg(unsigned char * p, int pcontrol, int target)
1875{ 	/* Caching page for mode_sense */
1876	unsigned char ch_caching_pg[] = {/* 0x8, 18, */ 0x4, 0, 0, 0, 0, 0,
1877		0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0};
1878	unsigned char d_caching_pg[] = {0x8, 18, 0x14, 0, 0xff, 0xff, 0, 0,
1879		0xff, 0xff, 0xff, 0xff, 0x80, 0x14, 0, 0,     0, 0, 0, 0};
1880
1881	if (SDEBUG_OPT_N_WCE & sdebug_opts)
1882		caching_pg[2] &= ~0x4;	/* set WCE=0 (default WCE=1) */
1883	memcpy(p, caching_pg, sizeof(caching_pg));
1884	if (1 == pcontrol)
1885		memcpy(p + 2, ch_caching_pg, sizeof(ch_caching_pg));
1886	else if (2 == pcontrol)
1887		memcpy(p, d_caching_pg, sizeof(d_caching_pg));
1888	return sizeof(caching_pg);
1889}
1890
1891static unsigned char ctrl_m_pg[] = {0xa, 10, 2, 0, 0, 0, 0, 0,
1892				    0, 0, 0x2, 0x4b};
1893
1894static int resp_ctrl_m_pg(unsigned char * p, int pcontrol, int target)
1895{ 	/* Control mode page for mode_sense */
1896	unsigned char ch_ctrl_m_pg[] = {/* 0xa, 10, */ 0x6, 0, 0, 0, 0, 0,
1897				        0, 0, 0, 0};
1898	unsigned char d_ctrl_m_pg[] = {0xa, 10, 2, 0, 0, 0, 0, 0,
1899				     0, 0, 0x2, 0x4b};
1900
1901	if (sdebug_dsense)
1902		ctrl_m_pg[2] |= 0x4;
1903	else
1904		ctrl_m_pg[2] &= ~0x4;
1905
1906	if (sdebug_ato)
1907		ctrl_m_pg[5] |= 0x80; /* ATO=1 */
1908
1909	memcpy(p, ctrl_m_pg, sizeof(ctrl_m_pg));
1910	if (1 == pcontrol)
1911		memcpy(p + 2, ch_ctrl_m_pg, sizeof(ch_ctrl_m_pg));
1912	else if (2 == pcontrol)
1913		memcpy(p, d_ctrl_m_pg, sizeof(d_ctrl_m_pg));
1914	return sizeof(ctrl_m_pg);
1915}
1916
1917
1918static int resp_iec_m_pg(unsigned char * p, int pcontrol, int target)
1919{	/* Informational Exceptions control mode page for mode_sense */
1920	unsigned char ch_iec_m_pg[] = {/* 0x1c, 0xa, */ 0x4, 0xf, 0, 0, 0, 0,
1921				       0, 0, 0x0, 0x0};
1922	unsigned char d_iec_m_pg[] = {0x1c, 0xa, 0x08, 0, 0, 0, 0, 0,
1923				      0, 0, 0x0, 0x0};
1924
1925	memcpy(p, iec_m_pg, sizeof(iec_m_pg));
1926	if (1 == pcontrol)
1927		memcpy(p + 2, ch_iec_m_pg, sizeof(ch_iec_m_pg));
1928	else if (2 == pcontrol)
1929		memcpy(p, d_iec_m_pg, sizeof(d_iec_m_pg));
1930	return sizeof(iec_m_pg);
1931}
1932
1933static int resp_sas_sf_m_pg(unsigned char * p, int pcontrol, int target)
1934{	/* SAS SSP mode page - short format for mode_sense */
1935	unsigned char sas_sf_m_pg[] = {0x19, 0x6,
1936		0x6, 0x0, 0x7, 0xd0, 0x0, 0x0};
1937
1938	memcpy(p, sas_sf_m_pg, sizeof(sas_sf_m_pg));
1939	if (1 == pcontrol)
1940		memset(p + 2, 0, sizeof(sas_sf_m_pg) - 2);
1941	return sizeof(sas_sf_m_pg);
1942}
1943
1944
1945static int resp_sas_pcd_m_spg(unsigned char * p, int pcontrol, int target,
1946			      int target_dev_id)
1947{	/* SAS phy control and discover mode page for mode_sense */
1948	unsigned char sas_pcd_m_pg[] = {0x59, 0x1, 0, 0x64, 0, 0x6, 0, 2,
1949		    0, 0, 0, 0, 0x10, 0x9, 0x8, 0x0,
1950		    0, 0, 0, 0, 0, 0, 0, 0,	/* insert SAS addr */
1951		    0, 0, 0, 0, 0, 0, 0, 0,	/* insert SAS addr */
1952		    0x2, 0, 0, 0, 0, 0, 0, 0,
1953		    0x88, 0x99, 0, 0, 0, 0, 0, 0,
1954		    0, 0, 0, 0, 0, 0, 0, 0,
1955		    0, 1, 0, 0, 0x10, 0x9, 0x8, 0x0,
1956		    0, 0, 0, 0, 0, 0, 0, 0,	/* insert SAS addr */
1957		    0, 0, 0, 0, 0, 0, 0, 0,	/* insert SAS addr */
1958		    0x3, 0, 0, 0, 0, 0, 0, 0,
1959		    0x88, 0x99, 0, 0, 0, 0, 0, 0,
1960		    0, 0, 0, 0, 0, 0, 0, 0,
1961		};
1962	int port_a, port_b;
1963
1964	put_unaligned_be64(naa3_comp_a, sas_pcd_m_pg + 16);
1965	put_unaligned_be64(naa3_comp_c + 1, sas_pcd_m_pg + 24);
1966	put_unaligned_be64(naa3_comp_a, sas_pcd_m_pg + 64);
1967	put_unaligned_be64(naa3_comp_c + 1, sas_pcd_m_pg + 72);
1968	port_a = target_dev_id + 1;
1969	port_b = port_a + 1;
1970	memcpy(p, sas_pcd_m_pg, sizeof(sas_pcd_m_pg));
1971	put_unaligned_be32(port_a, p + 20);
1972	put_unaligned_be32(port_b, p + 48 + 20);
 
 
 
 
 
 
1973	if (1 == pcontrol)
1974		memset(p + 4, 0, sizeof(sas_pcd_m_pg) - 4);
1975	return sizeof(sas_pcd_m_pg);
1976}
1977
1978static int resp_sas_sha_m_spg(unsigned char * p, int pcontrol)
1979{	/* SAS SSP shared protocol specific port mode subpage */
1980	unsigned char sas_sha_m_pg[] = {0x59, 0x2, 0, 0xc, 0, 0x6, 0x10, 0,
1981		    0, 0, 0, 0, 0, 0, 0, 0,
1982		};
1983
1984	memcpy(p, sas_sha_m_pg, sizeof(sas_sha_m_pg));
1985	if (1 == pcontrol)
1986		memset(p + 4, 0, sizeof(sas_sha_m_pg) - 4);
1987	return sizeof(sas_sha_m_pg);
1988}
1989
1990#define SDEBUG_MAX_MSENSE_SZ 256
1991
1992static int resp_mode_sense(struct scsi_cmnd *scp,
1993			   struct sdebug_dev_info *devip)
1994{
 
1995	int pcontrol, pcode, subpcode, bd_len;
1996	unsigned char dev_spec;
1997	int alloc_len, offset, len, target_dev_id;
1998	int target = scp->device->id;
1999	unsigned char * ap;
2000	unsigned char arr[SDEBUG_MAX_MSENSE_SZ];
2001	unsigned char *cmd = scp->cmnd;
2002	bool dbd, llbaa, msense_6, is_disk, bad_pcode;
2003
2004	dbd = !!(cmd[1] & 0x8);		/* disable block descriptors */
 
 
2005	pcontrol = (cmd[2] & 0xc0) >> 6;
2006	pcode = cmd[2] & 0x3f;
2007	subpcode = cmd[3];
2008	msense_6 = (MODE_SENSE == cmd[0]);
2009	llbaa = msense_6 ? false : !!(cmd[1] & 0x10);
2010	is_disk = (sdebug_ptype == TYPE_DISK);
2011	if (is_disk && !dbd)
2012		bd_len = llbaa ? 16 : 8;
2013	else
2014		bd_len = 0;
2015	alloc_len = msense_6 ? cmd[4] : get_unaligned_be16(cmd + 7);
2016	memset(arr, 0, SDEBUG_MAX_MSENSE_SZ);
2017	if (0x3 == pcontrol) {  /* Saving values not supported */
2018		mk_sense_buffer(scp, ILLEGAL_REQUEST, SAVING_PARAMS_UNSUP, 0);
 
2019		return check_condition_result;
2020	}
2021	target_dev_id = ((devip->sdbg_host->shost->host_no + 1) * 2000) +
2022			(devip->target * 1000) - 3;
2023	/* for disks set DPOFUA bit and clear write protect (WP) bit */
2024	if (is_disk)
2025		dev_spec = 0x10;	/* =0x90 if WP=1 implies read-only */
2026	else
2027		dev_spec = 0x0;
2028	if (msense_6) {
2029		arr[2] = dev_spec;
2030		arr[3] = bd_len;
2031		offset = 4;
2032	} else {
2033		arr[3] = dev_spec;
2034		if (16 == bd_len)
2035			arr[4] = 0x1;	/* set LONGLBA bit */
2036		arr[7] = bd_len;	/* assume 255 or less */
2037		offset = 8;
2038	}
2039	ap = arr + offset;
2040	if ((bd_len > 0) && (!sdebug_capacity))
2041		sdebug_capacity = get_sdebug_capacity();
2042
2043	if (8 == bd_len) {
2044		if (sdebug_capacity > 0xfffffffe)
2045			put_unaligned_be32(0xffffffff, ap + 0);
2046		else
2047			put_unaligned_be32(sdebug_capacity, ap + 0);
2048		put_unaligned_be16(sdebug_sector_size, ap + 6);
 
 
 
 
 
 
 
 
2049		offset += bd_len;
2050		ap = arr + offset;
2051	} else if (16 == bd_len) {
2052		put_unaligned_be64((u64)sdebug_capacity, ap + 0);
2053		put_unaligned_be32(sdebug_sector_size, ap + 12);
 
 
 
 
 
 
2054		offset += bd_len;
2055		ap = arr + offset;
2056	}
2057
2058	if ((subpcode > 0x0) && (subpcode < 0xff) && (0x19 != pcode)) {
2059		/* TODO: Control Extension page */
2060		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
 
2061		return check_condition_result;
2062	}
2063	bad_pcode = false;
2064
2065	switch (pcode) {
2066	case 0x1:	/* Read-Write error recovery page, direct access */
2067		len = resp_err_recov_pg(ap, pcontrol, target);
2068		offset += len;
2069		break;
2070	case 0x2:	/* Disconnect-Reconnect page, all devices */
2071		len = resp_disconnect_pg(ap, pcontrol, target);
2072		offset += len;
2073		break;
2074        case 0x3:       /* Format device page, direct access */
2075		if (is_disk) {
2076			len = resp_format_pg(ap, pcontrol, target);
2077			offset += len;
2078		} else
2079			bad_pcode = true;
2080                break;
2081	case 0x8:	/* Caching page, direct access */
2082		if (is_disk) {
2083			len = resp_caching_pg(ap, pcontrol, target);
2084			offset += len;
2085		} else
2086			bad_pcode = true;
2087		break;
2088	case 0xa:	/* Control Mode page, all devices */
2089		len = resp_ctrl_m_pg(ap, pcontrol, target);
2090		offset += len;
2091		break;
2092	case 0x19:	/* if spc==1 then sas phy, control+discover */
2093		if ((subpcode > 0x2) && (subpcode < 0xff)) {
2094			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
 
2095			return check_condition_result;
2096	        }
2097		len = 0;
2098		if ((0x0 == subpcode) || (0xff == subpcode))
2099			len += resp_sas_sf_m_pg(ap + len, pcontrol, target);
2100		if ((0x1 == subpcode) || (0xff == subpcode))
2101			len += resp_sas_pcd_m_spg(ap + len, pcontrol, target,
2102						  target_dev_id);
2103		if ((0x2 == subpcode) || (0xff == subpcode))
2104			len += resp_sas_sha_m_spg(ap + len, pcontrol);
2105		offset += len;
2106		break;
2107	case 0x1c:	/* Informational Exceptions Mode page, all devices */
2108		len = resp_iec_m_pg(ap, pcontrol, target);
2109		offset += len;
2110		break;
2111	case 0x3f:	/* Read all Mode pages */
2112		if ((0 == subpcode) || (0xff == subpcode)) {
2113			len = resp_err_recov_pg(ap, pcontrol, target);
2114			len += resp_disconnect_pg(ap + len, pcontrol, target);
2115			if (is_disk) {
2116				len += resp_format_pg(ap + len, pcontrol,
2117						      target);
2118				len += resp_caching_pg(ap + len, pcontrol,
2119						       target);
2120			}
2121			len += resp_ctrl_m_pg(ap + len, pcontrol, target);
2122			len += resp_sas_sf_m_pg(ap + len, pcontrol, target);
2123			if (0xff == subpcode) {
2124				len += resp_sas_pcd_m_spg(ap + len, pcontrol,
2125						  target, target_dev_id);
2126				len += resp_sas_sha_m_spg(ap + len, pcontrol);
2127			}
2128			len += resp_iec_m_pg(ap + len, pcontrol, target);
2129			offset += len;
2130		} else {
2131			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
 
2132			return check_condition_result;
2133                }
 
2134		break;
2135	default:
2136		bad_pcode = true;
2137		break;
2138	}
2139	if (bad_pcode) {
2140		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 5);
2141		return check_condition_result;
2142	}
2143	if (msense_6)
2144		arr[0] = offset - 1;
2145	else
2146		put_unaligned_be16((offset - 2), arr + 0);
 
 
2147	return fill_from_dev_buffer(scp, arr, min(alloc_len, offset));
2148}
2149
2150#define SDEBUG_MAX_MSELECT_SZ 512
2151
2152static int resp_mode_select(struct scsi_cmnd *scp,
2153			    struct sdebug_dev_info *devip)
2154{
2155	int pf, sp, ps, md_len, bd_len, off, spf, pg_len;
2156	int param_len, res, mpage;
2157	unsigned char arr[SDEBUG_MAX_MSELECT_SZ];
2158	unsigned char *cmd = scp->cmnd;
2159	int mselect6 = (MODE_SELECT == cmd[0]);
2160
 
 
2161	memset(arr, 0, sizeof(arr));
2162	pf = cmd[1] & 0x10;
2163	sp = cmd[1] & 0x1;
2164	param_len = mselect6 ? cmd[4] : get_unaligned_be16(cmd + 7);
2165	if ((0 == pf) || sp || (param_len > SDEBUG_MAX_MSELECT_SZ)) {
2166		mk_sense_invalid_fld(scp, SDEB_IN_CDB, mselect6 ? 4 : 7, -1);
 
2167		return check_condition_result;
2168	}
2169        res = fetch_to_dev_buffer(scp, arr, param_len);
2170        if (-1 == res)
2171		return DID_ERROR << 16;
2172	else if (sdebug_verbose && (res < param_len))
2173		sdev_printk(KERN_INFO, scp->device,
2174			    "%s: cdb indicated=%d, IO sent=%d bytes\n",
2175			    __func__, param_len, res);
2176	md_len = mselect6 ? (arr[0] + 1) : (get_unaligned_be16(arr + 0) + 2);
2177	bd_len = mselect6 ? arr[3] : get_unaligned_be16(arr + 6);
2178	if (md_len > 2) {
2179		mk_sense_invalid_fld(scp, SDEB_IN_DATA, 0, -1);
 
2180		return check_condition_result;
2181	}
2182	off = bd_len + (mselect6 ? 4 : 8);
2183	mpage = arr[off] & 0x3f;
2184	ps = !!(arr[off] & 0x80);
2185	if (ps) {
2186		mk_sense_invalid_fld(scp, SDEB_IN_DATA, off, 7);
 
2187		return check_condition_result;
2188	}
2189	spf = !!(arr[off] & 0x40);
2190	pg_len = spf ? (get_unaligned_be16(arr + off + 2) + 4) :
2191		       (arr[off + 1] + 2);
2192	if ((pg_len + off) > param_len) {
2193		mk_sense_buffer(scp, ILLEGAL_REQUEST,
2194				PARAMETER_LIST_LENGTH_ERR, 0);
2195		return check_condition_result;
2196	}
2197	switch (mpage) {
2198	case 0x8:      /* Caching Mode page */
2199		if (caching_pg[1] == arr[off + 1]) {
2200			memcpy(caching_pg + 2, arr + off + 2,
2201			       sizeof(caching_pg) - 2);
2202			goto set_mode_changed_ua;
2203		}
2204		break;
2205	case 0xa:      /* Control Mode page */
2206		if (ctrl_m_pg[1] == arr[off + 1]) {
2207			memcpy(ctrl_m_pg + 2, arr + off + 2,
2208			       sizeof(ctrl_m_pg) - 2);
2209			sdebug_dsense = !!(ctrl_m_pg[2] & 0x4);
2210			goto set_mode_changed_ua;
2211		}
2212		break;
2213	case 0x1c:      /* Informational Exceptions Mode page */
2214		if (iec_m_pg[1] == arr[off + 1]) {
2215			memcpy(iec_m_pg + 2, arr + off + 2,
2216			       sizeof(iec_m_pg) - 2);
2217			goto set_mode_changed_ua;
2218		}
2219		break;
2220	default:
2221		break;
2222	}
2223	mk_sense_invalid_fld(scp, SDEB_IN_DATA, off, 5);
 
2224	return check_condition_result;
2225set_mode_changed_ua:
2226	set_bit(SDEBUG_UA_MODE_CHANGED, devip->uas_bm);
2227	return 0;
2228}
2229
2230static int resp_temp_l_pg(unsigned char * arr)
2231{
2232	unsigned char temp_l_pg[] = {0x0, 0x0, 0x3, 0x2, 0x0, 38,
2233				     0x0, 0x1, 0x3, 0x2, 0x0, 65,
2234		};
2235
2236        memcpy(arr, temp_l_pg, sizeof(temp_l_pg));
2237        return sizeof(temp_l_pg);
2238}
2239
2240static int resp_ie_l_pg(unsigned char * arr)
2241{
2242	unsigned char ie_l_pg[] = {0x0, 0x0, 0x3, 0x3, 0x0, 0x0, 38,
2243		};
2244
2245        memcpy(arr, ie_l_pg, sizeof(ie_l_pg));
2246	if (iec_m_pg[2] & 0x4) {	/* TEST bit set */
2247		arr[4] = THRESHOLD_EXCEEDED;
2248		arr[5] = 0xff;
2249	}
2250        return sizeof(ie_l_pg);
2251}
2252
2253#define SDEBUG_MAX_LSENSE_SZ 512
2254
2255static int resp_log_sense(struct scsi_cmnd * scp,
2256                          struct sdebug_dev_info * devip)
2257{
2258	int ppc, sp, pcontrol, pcode, subpcode, alloc_len, len, n;
2259	unsigned char arr[SDEBUG_MAX_LSENSE_SZ];
2260	unsigned char *cmd = scp->cmnd;
2261
 
 
2262	memset(arr, 0, sizeof(arr));
2263	ppc = cmd[1] & 0x2;
2264	sp = cmd[1] & 0x1;
2265	if (ppc || sp) {
2266		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, ppc ? 1 : 0);
 
2267		return check_condition_result;
2268	}
2269	pcontrol = (cmd[2] & 0xc0) >> 6;
2270	pcode = cmd[2] & 0x3f;
2271	subpcode = cmd[3] & 0xff;
2272	alloc_len = get_unaligned_be16(cmd + 7);
2273	arr[0] = pcode;
2274	if (0 == subpcode) {
2275		switch (pcode) {
2276		case 0x0:	/* Supported log pages log page */
2277			n = 4;
2278			arr[n++] = 0x0;		/* this page */
2279			arr[n++] = 0xd;		/* Temperature */
2280			arr[n++] = 0x2f;	/* Informational exceptions */
2281			arr[3] = n - 4;
2282			break;
2283		case 0xd:	/* Temperature log page */
2284			arr[3] = resp_temp_l_pg(arr + 4);
2285			break;
2286		case 0x2f:	/* Informational exceptions log page */
2287			arr[3] = resp_ie_l_pg(arr + 4);
2288			break;
2289		default:
2290			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 5);
 
2291			return check_condition_result;
2292		}
2293	} else if (0xff == subpcode) {
2294		arr[0] |= 0x40;
2295		arr[1] = subpcode;
2296		switch (pcode) {
2297		case 0x0:	/* Supported log pages and subpages log page */
2298			n = 4;
2299			arr[n++] = 0x0;
2300			arr[n++] = 0x0;		/* 0,0 page */
2301			arr[n++] = 0x0;
2302			arr[n++] = 0xff;	/* this page */
2303			arr[n++] = 0xd;
2304			arr[n++] = 0x0;		/* Temperature */
2305			arr[n++] = 0x2f;
2306			arr[n++] = 0x0;	/* Informational exceptions */
2307			arr[3] = n - 4;
2308			break;
2309		case 0xd:	/* Temperature subpages */
2310			n = 4;
2311			arr[n++] = 0xd;
2312			arr[n++] = 0x0;		/* Temperature */
2313			arr[3] = n - 4;
2314			break;
2315		case 0x2f:	/* Informational exceptions subpages */
2316			n = 4;
2317			arr[n++] = 0x2f;
2318			arr[n++] = 0x0;		/* Informational exceptions */
2319			arr[3] = n - 4;
2320			break;
2321		default:
2322			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 5);
 
2323			return check_condition_result;
2324		}
2325	} else {
2326		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
 
2327		return check_condition_result;
2328	}
2329	len = min(get_unaligned_be16(arr + 2) + 4, alloc_len);
2330	return fill_from_dev_buffer(scp, arr,
2331		    min(len, SDEBUG_MAX_INQ_ARR_SZ));
2332}
2333
2334static int check_device_access_params(struct scsi_cmnd *scp,
2335				      unsigned long long lba, unsigned int num)
2336{
2337	if (lba + num > sdebug_capacity) {
2338		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
2339		return check_condition_result;
2340	}
2341	/* transfer length excessive (tie in to block limits VPD page) */
2342	if (num > sdebug_store_sectors) {
2343		/* needs work to find which cdb byte 'num' comes from */
2344		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
2345		return check_condition_result;
2346	}
2347	return 0;
2348}
2349
2350/* Returns number of bytes copied or -1 if error. */
2351static int do_device_access(struct scsi_cmnd *scmd, u64 lba, u32 num,
2352			    bool do_write)
 
2353{
2354	int ret;
2355	u64 block, rest = 0;
2356	struct scsi_data_buffer *sdb;
2357	enum dma_data_direction dir;
 
 
2358
2359	if (do_write) {
2360		sdb = scsi_out(scmd);
2361		dir = DMA_TO_DEVICE;
 
2362	} else {
2363		sdb = scsi_in(scmd);
2364		dir = DMA_FROM_DEVICE;
 
2365	}
2366
2367	if (!sdb->length)
2368		return 0;
2369	if (!(scsi_bidi_cmnd(scmd) || scmd->sc_data_direction == dir))
2370		return -1;
2371
2372	block = do_div(lba, sdebug_store_sectors);
2373	if (block + num > sdebug_store_sectors)
2374		rest = block + num - sdebug_store_sectors;
2375
2376	ret = sg_copy_buffer(sdb->table.sgl, sdb->table.nents,
2377		   fake_storep + (block * sdebug_sector_size),
2378		   (num - rest) * sdebug_sector_size, 0, do_write);
2379	if (ret != (num - rest) * sdebug_sector_size)
2380		return ret;
2381
2382	if (rest) {
2383		ret += sg_copy_buffer(sdb->table.sgl, sdb->table.nents,
2384			    fake_storep, rest * sdebug_sector_size,
2385			    (num - rest) * sdebug_sector_size, do_write);
2386	}
2387
2388	return ret;
2389}
2390
2391/* If fake_store(lba,num) compares equal to arr(num), then copy top half of
2392 * arr into fake_store(lba,num) and return true. If comparison fails then
2393 * return false. */
2394static bool comp_write_worker(u64 lba, u32 num, const u8 *arr)
2395{
2396	bool res;
2397	u64 block, rest = 0;
2398	u32 store_blks = sdebug_store_sectors;
2399	u32 lb_size = sdebug_sector_size;
2400
2401	block = do_div(lba, store_blks);
2402	if (block + num > store_blks)
2403		rest = block + num - store_blks;
2404
2405	res = !memcmp(fake_storep + (block * lb_size), arr,
2406		      (num - rest) * lb_size);
2407	if (!res)
2408		return res;
2409	if (rest)
2410		res = memcmp(fake_storep, arr + ((num - rest) * lb_size),
2411			     rest * lb_size);
2412	if (!res)
2413		return res;
2414	arr += num * lb_size;
2415	memcpy(fake_storep + (block * lb_size), arr, (num - rest) * lb_size);
2416	if (rest)
2417		memcpy(fake_storep, arr + ((num - rest) * lb_size),
2418		       rest * lb_size);
2419	return res;
2420}
2421
2422static __be16 dif_compute_csum(const void *buf, int len)
2423{
2424	__be16 csum;
2425
2426	if (sdebug_guard)
2427		csum = (__force __be16)ip_compute_csum(buf, len);
2428	else
2429		csum = cpu_to_be16(crc_t10dif(buf, len));
2430
2431	return csum;
2432}
2433
2434static int dif_verify(struct t10_pi_tuple *sdt, const void *data,
2435		      sector_t sector, u32 ei_lba)
2436{
2437	__be16 csum = dif_compute_csum(data, sdebug_sector_size);
2438
2439	if (sdt->guard_tag != csum) {
2440		pr_err("GUARD check failed on sector %lu rcvd 0x%04x, data 0x%04x\n",
 
2441			(unsigned long)sector,
2442			be16_to_cpu(sdt->guard_tag),
2443			be16_to_cpu(csum));
2444		return 0x01;
2445	}
2446	if (sdebug_dif == T10_PI_TYPE1_PROTECTION &&
2447	    be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) {
2448		pr_err("REF check failed on sector %lu\n",
2449			(unsigned long)sector);
2450		return 0x03;
2451	}
2452	if (sdebug_dif == T10_PI_TYPE2_PROTECTION &&
2453	    be32_to_cpu(sdt->ref_tag) != ei_lba) {
2454		pr_err("REF check failed on sector %lu\n",
2455			(unsigned long)sector);
2456		return 0x03;
2457	}
2458	return 0;
2459}
2460
2461static void dif_copy_prot(struct scsi_cmnd *SCpnt, sector_t sector,
2462			  unsigned int sectors, bool read)
2463{
2464	size_t resid;
2465	void *paddr;
2466	const void *dif_store_end = dif_storep + sdebug_store_sectors;
2467	struct sg_mapping_iter miter;
2468
2469	/* Bytes of protection data to copy into sgl */
2470	resid = sectors * sizeof(*dif_storep);
2471
2472	sg_miter_start(&miter, scsi_prot_sglist(SCpnt),
2473			scsi_prot_sg_count(SCpnt), SG_MITER_ATOMIC |
2474			(read ? SG_MITER_TO_SG : SG_MITER_FROM_SG));
2475
2476	while (sg_miter_next(&miter) && resid > 0) {
2477		size_t len = min(miter.length, resid);
2478		void *start = dif_store(sector);
2479		size_t rest = 0;
2480
2481		if (dif_store_end < start + len)
2482			rest = start + len - dif_store_end;
2483
2484		paddr = miter.addr;
2485
2486		if (read)
2487			memcpy(paddr, start, len - rest);
2488		else
2489			memcpy(start, paddr, len - rest);
2490
2491		if (rest) {
2492			if (read)
2493				memcpy(paddr + len - rest, dif_storep, rest);
2494			else
2495				memcpy(dif_storep, paddr + len - rest, rest);
2496		}
2497
2498		sector += len / sizeof(*dif_storep);
2499		resid -= len;
2500	}
2501	sg_miter_stop(&miter);
2502}
2503
2504static int prot_verify_read(struct scsi_cmnd *SCpnt, sector_t start_sec,
2505			    unsigned int sectors, u32 ei_lba)
2506{
2507	unsigned int i;
2508	struct t10_pi_tuple *sdt;
2509	sector_t sector;
2510
2511	for (i = 0; i < sectors; i++, ei_lba++) {
2512		int ret;
2513
2514		sector = start_sec + i;
2515		sdt = dif_store(sector);
2516
2517		if (sdt->app_tag == cpu_to_be16(0xffff))
2518			continue;
2519
2520		ret = dif_verify(sdt, fake_store(sector), sector, ei_lba);
2521		if (ret) {
2522			dif_errors++;
2523			return ret;
2524		}
2525	}
2526
2527	dif_copy_prot(SCpnt, start_sec, sectors, true);
2528	dix_reads++;
2529
2530	return 0;
2531}
2532
2533static int resp_read_dt0(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
 
 
2534{
2535	u8 *cmd = scp->cmnd;
2536	struct sdebug_queued_cmd *sqcp;
2537	u64 lba;
2538	u32 num;
2539	u32 ei_lba;
2540	unsigned long iflags;
2541	int ret;
2542	bool check_prot;
2543
2544	switch (cmd[0]) {
2545	case READ_16:
2546		ei_lba = 0;
2547		lba = get_unaligned_be64(cmd + 2);
2548		num = get_unaligned_be32(cmd + 10);
2549		check_prot = true;
2550		break;
2551	case READ_10:
2552		ei_lba = 0;
2553		lba = get_unaligned_be32(cmd + 2);
2554		num = get_unaligned_be16(cmd + 7);
2555		check_prot = true;
2556		break;
2557	case READ_6:
2558		ei_lba = 0;
2559		lba = (u32)cmd[3] | (u32)cmd[2] << 8 |
2560		      (u32)(cmd[1] & 0x1f) << 16;
2561		num = (0 == cmd[4]) ? 256 : cmd[4];
2562		check_prot = true;
2563		break;
2564	case READ_12:
2565		ei_lba = 0;
2566		lba = get_unaligned_be32(cmd + 2);
2567		num = get_unaligned_be32(cmd + 6);
2568		check_prot = true;
2569		break;
2570	case XDWRITEREAD_10:
2571		ei_lba = 0;
2572		lba = get_unaligned_be32(cmd + 2);
2573		num = get_unaligned_be16(cmd + 7);
2574		check_prot = false;
2575		break;
2576	default:	/* assume READ(32) */
2577		lba = get_unaligned_be64(cmd + 12);
2578		ei_lba = get_unaligned_be32(cmd + 20);
2579		num = get_unaligned_be32(cmd + 28);
2580		check_prot = false;
2581		break;
2582	}
2583	if (unlikely(have_dif_prot && check_prot)) {
2584		if (sdebug_dif == T10_PI_TYPE2_PROTECTION &&
2585		    (cmd[1] & 0xe0)) {
2586			mk_sense_invalid_opcode(scp);
2587			return check_condition_result;
2588		}
2589		if ((sdebug_dif == T10_PI_TYPE1_PROTECTION ||
2590		     sdebug_dif == T10_PI_TYPE3_PROTECTION) &&
2591		    (cmd[1] & 0xe0) == 0)
2592			sdev_printk(KERN_ERR, scp->device, "Unprotected RD "
2593				    "to DIF device\n");
2594	}
2595	if (unlikely(sdebug_any_injecting_opt)) {
2596		sqcp = (struct sdebug_queued_cmd *)scp->host_scribble;
2597
2598		if (sqcp) {
2599			if (sqcp->inj_short)
2600				num /= 2;
2601		}
2602	} else
2603		sqcp = NULL;
2604
2605	/* inline check_device_access_params() */
2606	if (unlikely(lba + num > sdebug_capacity)) {
2607		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
2608		return check_condition_result;
2609	}
2610	/* transfer length excessive (tie in to block limits VPD page) */
2611	if (unlikely(num > sdebug_store_sectors)) {
2612		/* needs work to find which cdb byte 'num' comes from */
2613		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
2614		return check_condition_result;
2615	}
2616
2617	if (unlikely((SDEBUG_OPT_MEDIUM_ERR & sdebug_opts) &&
2618		     (lba <= (OPT_MEDIUM_ERR_ADDR + OPT_MEDIUM_ERR_NUM - 1)) &&
2619		     ((lba + num) > OPT_MEDIUM_ERR_ADDR))) {
2620		/* claim unrecoverable read error */
2621		mk_sense_buffer(scp, MEDIUM_ERROR, UNRECOVERED_READ_ERR, 0);
2622		/* set info field and valid bit for fixed descriptor */
2623		if (0x70 == (scp->sense_buffer[0] & 0x7f)) {
2624			scp->sense_buffer[0] |= 0x80;	/* Valid bit */
2625			ret = (lba < OPT_MEDIUM_ERR_ADDR)
2626			      ? OPT_MEDIUM_ERR_ADDR : (int)lba;
2627			put_unaligned_be32(ret, scp->sense_buffer + 3);
 
 
 
2628		}
2629		scsi_set_resid(scp, scsi_bufflen(scp));
2630		return check_condition_result;
2631	}
2632
2633	read_lock_irqsave(&atomic_rw, iflags);
2634
2635	/* DIX + T10 DIF */
2636	if (unlikely(sdebug_dix && scsi_prot_sg_count(scp))) {
2637		int prot_ret = prot_verify_read(scp, lba, num, ei_lba);
2638
2639		if (prot_ret) {
2640			read_unlock_irqrestore(&atomic_rw, iflags);
2641			mk_sense_buffer(scp, ABORTED_COMMAND, 0x10, prot_ret);
2642			return illegal_condition_result;
2643		}
2644	}
2645
2646	ret = do_device_access(scp, lba, num, false);
2647	read_unlock_irqrestore(&atomic_rw, iflags);
2648	if (unlikely(ret == -1))
2649		return DID_ERROR << 16;
2650
2651	scsi_in(scp)->resid = scsi_bufflen(scp) - ret;
2652
2653	if (unlikely(sqcp)) {
2654		if (sqcp->inj_recovered) {
2655			mk_sense_buffer(scp, RECOVERED_ERROR,
2656					THRESHOLD_EXCEEDED, 0);
2657			return check_condition_result;
2658		} else if (sqcp->inj_transport) {
2659			mk_sense_buffer(scp, ABORTED_COMMAND,
2660					TRANSPORT_PROBLEM, ACK_NAK_TO);
2661			return check_condition_result;
2662		} else if (sqcp->inj_dif) {
2663			/* Logical block guard check failed */
2664			mk_sense_buffer(scp, ABORTED_COMMAND, 0x10, 1);
2665			return illegal_condition_result;
2666		} else if (sqcp->inj_dix) {
2667			mk_sense_buffer(scp, ILLEGAL_REQUEST, 0x10, 1);
2668			return illegal_condition_result;
2669		}
2670	}
2671	return 0;
2672}
2673
2674static void dump_sector(unsigned char *buf, int len)
2675{
2676	int i, j, n;
 
 
2677
2678	pr_err(">>> Sector Dump <<<\n");
2679	for (i = 0 ; i < len ; i += 16) {
2680		char b[128];
2681
2682		for (j = 0, n = 0; j < 16; j++) {
2683			unsigned char c = buf[i+j];
2684
2685			if (c >= 0x20 && c < 0x7e)
2686				n += scnprintf(b + n, sizeof(b) - n,
2687					       " %c ", buf[i+j]);
2688			else
2689				n += scnprintf(b + n, sizeof(b) - n,
2690					       "%02x ", buf[i+j]);
2691		}
2692		pr_err("%04d: %s\n", i, b);
 
2693	}
2694}
2695
2696static int prot_verify_write(struct scsi_cmnd *SCpnt, sector_t start_sec,
2697			     unsigned int sectors, u32 ei_lba)
2698{
2699	int ret;
2700	struct t10_pi_tuple *sdt;
2701	void *daddr;
2702	sector_t sector = start_sec;
2703	int ppage_offset;
2704	int dpage_offset;
2705	struct sg_mapping_iter diter;
2706	struct sg_mapping_iter piter;
2707
2708	BUG_ON(scsi_sg_count(SCpnt) == 0);
2709	BUG_ON(scsi_prot_sg_count(SCpnt) == 0);
2710
2711	sg_miter_start(&piter, scsi_prot_sglist(SCpnt),
2712			scsi_prot_sg_count(SCpnt),
2713			SG_MITER_ATOMIC | SG_MITER_FROM_SG);
2714	sg_miter_start(&diter, scsi_sglist(SCpnt), scsi_sg_count(SCpnt),
2715			SG_MITER_ATOMIC | SG_MITER_FROM_SG);
2716
2717	/* For each protection page */
2718	while (sg_miter_next(&piter)) {
2719		dpage_offset = 0;
2720		if (WARN_ON(!sg_miter_next(&diter))) {
2721			ret = 0x01;
2722			goto out;
2723		}
2724
2725		for (ppage_offset = 0; ppage_offset < piter.length;
2726		     ppage_offset += sizeof(struct t10_pi_tuple)) {
2727			/* If we're at the end of the current
2728			 * data page advance to the next one
2729			 */
2730			if (dpage_offset >= diter.length) {
2731				if (WARN_ON(!sg_miter_next(&diter))) {
2732					ret = 0x01;
2733					goto out;
2734				}
2735				dpage_offset = 0;
2736			}
2737
2738			sdt = piter.addr + ppage_offset;
2739			daddr = diter.addr + dpage_offset;
2740
2741			ret = dif_verify(sdt, daddr, sector, ei_lba);
2742			if (ret) {
2743				dump_sector(daddr, sdebug_sector_size);
2744				goto out;
2745			}
2746
2747			sector++;
2748			ei_lba++;
2749			dpage_offset += sdebug_sector_size;
2750		}
2751		diter.consumed = dpage_offset;
2752		sg_miter_stop(&diter);
2753	}
2754	sg_miter_stop(&piter);
2755
2756	dif_copy_prot(SCpnt, start_sec, sectors, false);
2757	dix_writes++;
2758
2759	return 0;
2760
2761out:
2762	dif_errors++;
2763	sg_miter_stop(&diter);
2764	sg_miter_stop(&piter);
2765	return ret;
2766}
2767
2768static unsigned long lba_to_map_index(sector_t lba)
2769{
2770	if (sdebug_unmap_alignment)
2771		lba += sdebug_unmap_granularity - sdebug_unmap_alignment;
2772	sector_div(lba, sdebug_unmap_granularity);
 
 
 
2773	return lba;
2774}
2775
2776static sector_t map_index_to_lba(unsigned long index)
2777{
2778	sector_t lba = index * sdebug_unmap_granularity;
 
 
 
 
 
2779
2780	if (sdebug_unmap_alignment)
2781		lba -= sdebug_unmap_granularity - sdebug_unmap_alignment;
2782	return lba;
2783}
2784
2785static unsigned int map_state(sector_t lba, unsigned int *num)
2786{
2787	sector_t end;
2788	unsigned int mapped;
2789	unsigned long index;
2790	unsigned long next;
2791
2792	index = lba_to_map_index(lba);
2793	mapped = test_bit(index, map_storep);
2794
2795	if (mapped)
2796		next = find_next_zero_bit(map_storep, map_size, index);
2797	else
2798		next = find_next_bit(map_storep, map_size, index);
2799
2800	end = min_t(sector_t, sdebug_store_sectors,  map_index_to_lba(next));
2801	*num = end - lba;
 
2802	return mapped;
2803}
2804
2805static void map_region(sector_t lba, unsigned int len)
2806{
2807	sector_t end = lba + len;
2808
2809	while (lba < end) {
2810		unsigned long index = lba_to_map_index(lba);
2811
2812		if (index < map_size)
2813			set_bit(index, map_storep);
2814
2815		lba = map_index_to_lba(index + 1);
2816	}
2817}
2818
2819static void unmap_region(sector_t lba, unsigned int len)
2820{
2821	sector_t end = lba + len;
2822
2823	while (lba < end) {
2824		unsigned long index = lba_to_map_index(lba);
2825
2826		if (lba == map_index_to_lba(index) &&
2827		    lba + sdebug_unmap_granularity <= end &&
2828		    index < map_size) {
2829			clear_bit(index, map_storep);
2830			if (sdebug_lbprz) {  /* for LBPRZ=2 return 0xff_s */
2831				memset(fake_storep +
2832				       lba * sdebug_sector_size,
2833				       (sdebug_lbprz & 1) ? 0 : 0xff,
2834				       sdebug_sector_size *
2835				       sdebug_unmap_granularity);
2836			}
2837			if (dif_storep) {
2838				memset(dif_storep + lba, 0xff,
2839				       sizeof(*dif_storep) *
2840				       sdebug_unmap_granularity);
2841			}
2842		}
2843		lba = map_index_to_lba(index + 1);
2844	}
2845}
2846
2847static int resp_write_dt0(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
 
 
2848{
2849	u8 *cmd = scp->cmnd;
2850	u64 lba;
2851	u32 num;
2852	u32 ei_lba;
2853	unsigned long iflags;
2854	int ret;
2855	bool check_prot;
2856
2857	switch (cmd[0]) {
2858	case WRITE_16:
2859		ei_lba = 0;
2860		lba = get_unaligned_be64(cmd + 2);
2861		num = get_unaligned_be32(cmd + 10);
2862		check_prot = true;
2863		break;
2864	case WRITE_10:
2865		ei_lba = 0;
2866		lba = get_unaligned_be32(cmd + 2);
2867		num = get_unaligned_be16(cmd + 7);
2868		check_prot = true;
2869		break;
2870	case WRITE_6:
2871		ei_lba = 0;
2872		lba = (u32)cmd[3] | (u32)cmd[2] << 8 |
2873		      (u32)(cmd[1] & 0x1f) << 16;
2874		num = (0 == cmd[4]) ? 256 : cmd[4];
2875		check_prot = true;
2876		break;
2877	case WRITE_12:
2878		ei_lba = 0;
2879		lba = get_unaligned_be32(cmd + 2);
2880		num = get_unaligned_be32(cmd + 6);
2881		check_prot = true;
2882		break;
2883	case 0x53:	/* XDWRITEREAD(10) */
2884		ei_lba = 0;
2885		lba = get_unaligned_be32(cmd + 2);
2886		num = get_unaligned_be16(cmd + 7);
2887		check_prot = false;
2888		break;
2889	default:	/* assume WRITE(32) */
2890		lba = get_unaligned_be64(cmd + 12);
2891		ei_lba = get_unaligned_be32(cmd + 20);
2892		num = get_unaligned_be32(cmd + 28);
2893		check_prot = false;
2894		break;
2895	}
2896	if (unlikely(have_dif_prot && check_prot)) {
2897		if (sdebug_dif == T10_PI_TYPE2_PROTECTION &&
2898		    (cmd[1] & 0xe0)) {
2899			mk_sense_invalid_opcode(scp);
2900			return check_condition_result;
2901		}
2902		if ((sdebug_dif == T10_PI_TYPE1_PROTECTION ||
2903		     sdebug_dif == T10_PI_TYPE3_PROTECTION) &&
2904		    (cmd[1] & 0xe0) == 0)
2905			sdev_printk(KERN_ERR, scp->device, "Unprotected WR "
2906				    "to DIF device\n");
2907	}
2908
2909	/* inline check_device_access_params() */
2910	if (unlikely(lba + num > sdebug_capacity)) {
2911		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
2912		return check_condition_result;
2913	}
2914	/* transfer length excessive (tie in to block limits VPD page) */
2915	if (unlikely(num > sdebug_store_sectors)) {
2916		/* needs work to find which cdb byte 'num' comes from */
2917		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
2918		return check_condition_result;
2919	}
2920
2921	write_lock_irqsave(&atomic_rw, iflags);
2922
2923	/* DIX + T10 DIF */
2924	if (unlikely(sdebug_dix && scsi_prot_sg_count(scp))) {
2925		int prot_ret = prot_verify_write(scp, lba, num, ei_lba);
2926
2927		if (prot_ret) {
2928			write_unlock_irqrestore(&atomic_rw, iflags);
2929			mk_sense_buffer(scp, ILLEGAL_REQUEST, 0x10, prot_ret);
2930			return illegal_condition_result;
2931		}
2932	}
2933
2934	ret = do_device_access(scp, lba, num, true);
2935	if (unlikely(scsi_debug_lbp()))
2936		map_region(lba, num);
2937	write_unlock_irqrestore(&atomic_rw, iflags);
2938	if (unlikely(-1 == ret))
2939		return DID_ERROR << 16;
2940	else if (unlikely(sdebug_verbose &&
2941			  (ret < (num * sdebug_sector_size))))
2942		sdev_printk(KERN_INFO, scp->device,
2943			    "%s: write: cdb indicated=%u, IO sent=%d bytes\n",
2944			    my_name, num * sdebug_sector_size, ret);
2945
2946	if (unlikely(sdebug_any_injecting_opt)) {
2947		struct sdebug_queued_cmd *sqcp =
2948				(struct sdebug_queued_cmd *)scp->host_scribble;
2949
2950		if (sqcp) {
2951			if (sqcp->inj_recovered) {
2952				mk_sense_buffer(scp, RECOVERED_ERROR,
2953						THRESHOLD_EXCEEDED, 0);
2954				return check_condition_result;
2955			} else if (sqcp->inj_dif) {
2956				/* Logical block guard check failed */
2957				mk_sense_buffer(scp, ABORTED_COMMAND, 0x10, 1);
2958				return illegal_condition_result;
2959			} else if (sqcp->inj_dix) {
2960				mk_sense_buffer(scp, ILLEGAL_REQUEST, 0x10, 1);
2961				return illegal_condition_result;
2962			}
2963		}
2964	}
2965	return 0;
2966}
2967
2968static int resp_write_same(struct scsi_cmnd *scp, u64 lba, u32 num,
2969			   u32 ei_lba, bool unmap, bool ndob)
 
2970{
2971	unsigned long iflags;
2972	unsigned long long i;
2973	int ret;
2974	u64 lba_off;
2975
2976	ret = check_device_access_params(scp, lba, num);
2977	if (ret)
2978		return ret;
2979
 
 
 
 
 
 
2980	write_lock_irqsave(&atomic_rw, iflags);
2981
2982	if (unmap && scsi_debug_lbp()) {
2983		unmap_region(lba, num);
2984		goto out;
2985	}
2986
2987	lba_off = lba * sdebug_sector_size;
2988	/* if ndob then zero 1 logical block, else fetch 1 logical block */
2989	if (ndob) {
2990		memset(fake_storep + lba_off, 0, sdebug_sector_size);
2991		ret = 0;
2992	} else
2993		ret = fetch_to_dev_buffer(scp, fake_storep + lba_off,
2994					  sdebug_sector_size);
2995
2996	if (-1 == ret) {
2997		write_unlock_irqrestore(&atomic_rw, iflags);
2998		return DID_ERROR << 16;
2999	} else if (sdebug_verbose && (ret < (num * sdebug_sector_size)))
3000		sdev_printk(KERN_INFO, scp->device,
3001			    "%s: %s: cdb indicated=%u, IO sent=%d bytes\n",
3002			    my_name, "write same",
3003			    num * sdebug_sector_size, ret);
3004
3005	/* Copy first sector to remaining blocks */
3006	for (i = 1 ; i < num ; i++)
3007		memcpy(fake_storep + ((lba + i) * sdebug_sector_size),
3008		       fake_storep + lba_off,
3009		       sdebug_sector_size);
3010
3011	if (scsi_debug_lbp())
3012		map_region(lba, num);
3013out:
3014	write_unlock_irqrestore(&atomic_rw, iflags);
3015
3016	return 0;
3017}
3018
3019static int resp_write_same_10(struct scsi_cmnd *scp,
3020			      struct sdebug_dev_info *devip)
3021{
3022	u8 *cmd = scp->cmnd;
3023	u32 lba;
3024	u16 num;
3025	u32 ei_lba = 0;
3026	bool unmap = false;
3027
3028	if (cmd[1] & 0x8) {
3029		if (sdebug_lbpws10 == 0) {
3030			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 3);
3031			return check_condition_result;
3032		} else
3033			unmap = true;
3034	}
3035	lba = get_unaligned_be32(cmd + 2);
3036	num = get_unaligned_be16(cmd + 7);
3037	if (num > sdebug_write_same_length) {
3038		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 7, -1);
3039		return check_condition_result;
3040	}
3041	return resp_write_same(scp, lba, num, ei_lba, unmap, false);
3042}
3043
3044static int resp_write_same_16(struct scsi_cmnd *scp,
3045			      struct sdebug_dev_info *devip)
3046{
3047	u8 *cmd = scp->cmnd;
3048	u64 lba;
3049	u32 num;
3050	u32 ei_lba = 0;
3051	bool unmap = false;
3052	bool ndob = false;
3053
3054	if (cmd[1] & 0x8) {	/* UNMAP */
3055		if (sdebug_lbpws == 0) {
3056			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 3);
3057			return check_condition_result;
3058		} else
3059			unmap = true;
3060	}
3061	if (cmd[1] & 0x1)  /* NDOB (no data-out buffer, assumes zeroes) */
3062		ndob = true;
3063	lba = get_unaligned_be64(cmd + 2);
3064	num = get_unaligned_be32(cmd + 10);
3065	if (num > sdebug_write_same_length) {
3066		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 10, -1);
3067		return check_condition_result;
3068	}
3069	return resp_write_same(scp, lba, num, ei_lba, unmap, ndob);
3070}
3071
3072/* Note the mode field is in the same position as the (lower) service action
3073 * field. For the Report supported operation codes command, SPC-4 suggests
3074 * each mode of this command should be reported separately; for future. */
3075static int resp_write_buffer(struct scsi_cmnd *scp,
3076			     struct sdebug_dev_info *devip)
3077{
3078	u8 *cmd = scp->cmnd;
3079	struct scsi_device *sdp = scp->device;
3080	struct sdebug_dev_info *dp;
3081	u8 mode;
3082
3083	mode = cmd[1] & 0x1f;
3084	switch (mode) {
3085	case 0x4:	/* download microcode (MC) and activate (ACT) */
3086		/* set UAs on this device only */
3087		set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3088		set_bit(SDEBUG_UA_MICROCODE_CHANGED, devip->uas_bm);
3089		break;
3090	case 0x5:	/* download MC, save and ACT */
3091		set_bit(SDEBUG_UA_MICROCODE_CHANGED_WO_RESET, devip->uas_bm);
3092		break;
3093	case 0x6:	/* download MC with offsets and ACT */
3094		/* set UAs on most devices (LUs) in this target */
3095		list_for_each_entry(dp,
3096				    &devip->sdbg_host->dev_info_list,
3097				    dev_list)
3098			if (dp->target == sdp->id) {
3099				set_bit(SDEBUG_UA_BUS_RESET, dp->uas_bm);
3100				if (devip != dp)
3101					set_bit(SDEBUG_UA_MICROCODE_CHANGED,
3102						dp->uas_bm);
3103			}
3104		break;
3105	case 0x7:	/* download MC with offsets, save, and ACT */
3106		/* set UA on all devices (LUs) in this target */
3107		list_for_each_entry(dp,
3108				    &devip->sdbg_host->dev_info_list,
3109				    dev_list)
3110			if (dp->target == sdp->id)
3111				set_bit(SDEBUG_UA_MICROCODE_CHANGED_WO_RESET,
3112					dp->uas_bm);
3113		break;
3114	default:
3115		/* do nothing for this command for other mode values */
3116		break;
3117	}
3118	return 0;
3119}
3120
3121static int resp_comp_write(struct scsi_cmnd *scp,
3122			   struct sdebug_dev_info *devip)
3123{
3124	u8 *cmd = scp->cmnd;
3125	u8 *arr;
3126	u8 *fake_storep_hold;
3127	u64 lba;
3128	u32 dnum;
3129	u32 lb_size = sdebug_sector_size;
3130	u8 num;
3131	unsigned long iflags;
3132	int ret;
3133	int retval = 0;
3134
3135	lba = get_unaligned_be64(cmd + 2);
3136	num = cmd[13];		/* 1 to a maximum of 255 logical blocks */
3137	if (0 == num)
3138		return 0;	/* degenerate case, not an error */
3139	if (sdebug_dif == T10_PI_TYPE2_PROTECTION &&
3140	    (cmd[1] & 0xe0)) {
3141		mk_sense_invalid_opcode(scp);
3142		return check_condition_result;
3143	}
3144	if ((sdebug_dif == T10_PI_TYPE1_PROTECTION ||
3145	     sdebug_dif == T10_PI_TYPE3_PROTECTION) &&
3146	    (cmd[1] & 0xe0) == 0)
3147		sdev_printk(KERN_ERR, scp->device, "Unprotected WR "
3148			    "to DIF device\n");
3149
3150	/* inline check_device_access_params() */
3151	if (lba + num > sdebug_capacity) {
3152		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
3153		return check_condition_result;
3154	}
3155	/* transfer length excessive (tie in to block limits VPD page) */
3156	if (num > sdebug_store_sectors) {
3157		/* needs work to find which cdb byte 'num' comes from */
3158		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
3159		return check_condition_result;
3160	}
3161	dnum = 2 * num;
3162	arr = kzalloc(dnum * lb_size, GFP_ATOMIC);
3163	if (NULL == arr) {
3164		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3165				INSUFF_RES_ASCQ);
3166		return check_condition_result;
3167	}
3168
3169	write_lock_irqsave(&atomic_rw, iflags);
3170
3171	/* trick do_device_access() to fetch both compare and write buffers
3172	 * from data-in into arr. Safe (atomic) since write_lock held. */
3173	fake_storep_hold = fake_storep;
3174	fake_storep = arr;
3175	ret = do_device_access(scp, 0, dnum, true);
3176	fake_storep = fake_storep_hold;
3177	if (ret == -1) {
3178		retval = DID_ERROR << 16;
3179		goto cleanup;
3180	} else if (sdebug_verbose && (ret < (dnum * lb_size)))
3181		sdev_printk(KERN_INFO, scp->device, "%s: compare_write: cdb "
3182			    "indicated=%u, IO sent=%d bytes\n", my_name,
3183			    dnum * lb_size, ret);
3184	if (!comp_write_worker(lba, num, arr)) {
3185		mk_sense_buffer(scp, MISCOMPARE, MISCOMPARE_VERIFY_ASC, 0);
3186		retval = check_condition_result;
3187		goto cleanup;
3188	}
3189	if (scsi_debug_lbp())
3190		map_region(lba, num);
3191cleanup:
3192	write_unlock_irqrestore(&atomic_rw, iflags);
3193	kfree(arr);
3194	return retval;
3195}
3196
3197struct unmap_block_desc {
3198	__be64	lba;
3199	__be32	blocks;
3200	__be32	__reserved;
3201};
3202
3203static int resp_unmap(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3204{
3205	unsigned char *buf;
3206	struct unmap_block_desc *desc;
3207	unsigned int i, payload_len, descriptors;
3208	int ret;
3209	unsigned long iflags;
3210
 
 
 
3211
3212	if (!scsi_debug_lbp())
3213		return 0;	/* fib and say its done */
3214	payload_len = get_unaligned_be16(scp->cmnd + 7);
3215	BUG_ON(scsi_bufflen(scp) != payload_len);
3216
3217	descriptors = (payload_len - 8) / 16;
3218	if (descriptors > sdebug_unmap_max_desc) {
3219		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 7, -1);
3220		return check_condition_result;
3221	}
3222
3223	buf = kzalloc(scsi_bufflen(scp), GFP_ATOMIC);
3224	if (!buf) {
3225		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3226				INSUFF_RES_ASCQ);
3227		return check_condition_result;
3228	}
3229
3230	scsi_sg_copy_to_buffer(scp, buf, scsi_bufflen(scp));
3231
3232	BUG_ON(get_unaligned_be16(&buf[0]) != payload_len - 2);
3233	BUG_ON(get_unaligned_be16(&buf[2]) != descriptors * 16);
3234
3235	desc = (void *)&buf[8];
3236
3237	write_lock_irqsave(&atomic_rw, iflags);
3238
3239	for (i = 0 ; i < descriptors ; i++) {
3240		unsigned long long lba = get_unaligned_be64(&desc[i].lba);
3241		unsigned int num = get_unaligned_be32(&desc[i].blocks);
3242
3243		ret = check_device_access_params(scp, lba, num);
3244		if (ret)
3245			goto out;
3246
3247		unmap_region(lba, num);
3248	}
3249
3250	ret = 0;
3251
3252out:
3253	write_unlock_irqrestore(&atomic_rw, iflags);
3254	kfree(buf);
3255
3256	return ret;
3257}
3258
3259#define SDEBUG_GET_LBA_STATUS_LEN 32
3260
3261static int resp_get_lba_status(struct scsi_cmnd *scp,
3262			       struct sdebug_dev_info *devip)
3263{
3264	u8 *cmd = scp->cmnd;
3265	u64 lba;
3266	u32 alloc_len, mapped, num;
3267	u8 arr[SDEBUG_GET_LBA_STATUS_LEN];
3268	int ret;
3269
3270	lba = get_unaligned_be64(cmd + 2);
3271	alloc_len = get_unaligned_be32(cmd + 10);
 
 
 
 
3272
3273	if (alloc_len < 24)
3274		return 0;
3275
3276	ret = check_device_access_params(scp, lba, 1);
3277	if (ret)
3278		return ret;
3279
3280	if (scsi_debug_lbp())
3281		mapped = map_state(lba, &num);
3282	else {
3283		mapped = 1;
3284		/* following just in case virtual_gb changed */
3285		sdebug_capacity = get_sdebug_capacity();
3286		if (sdebug_capacity - lba <= 0xffffffff)
3287			num = sdebug_capacity - lba;
3288		else
3289			num = 0xffffffff;
3290	}
3291
3292	memset(arr, 0, SDEBUG_GET_LBA_STATUS_LEN);
3293	put_unaligned_be32(20, arr);		/* Parameter Data Length */
3294	put_unaligned_be64(lba, arr + 8);	/* LBA */
3295	put_unaligned_be32(num, arr + 16);	/* Number of blocks */
3296	arr[20] = !mapped;		/* prov_stat=0: mapped; 1: dealloc */
3297
3298	return fill_from_dev_buffer(scp, arr, SDEBUG_GET_LBA_STATUS_LEN);
3299}
3300
3301#define RL_BUCKET_ELEMS 8
3302
3303/* Even though each pseudo target has a REPORT LUNS "well known logical unit"
3304 * (W-LUN), the normal Linux scanning logic does not associate it with a
3305 * device (e.g. /dev/sg7). The following magic will make that association:
3306 *   "cd /sys/class/scsi_host/host<n> ; echo '- - 49409' > scan"
3307 * where <n> is a host number. If there are multiple targets in a host then
3308 * the above will associate a W-LUN to each target. To only get a W-LUN
3309 * for target 2, then use "echo '- 2 49409' > scan" .
3310 */
3311static int resp_report_luns(struct scsi_cmnd *scp,
3312			    struct sdebug_dev_info *devip)
3313{
3314	unsigned char *cmd = scp->cmnd;
3315	unsigned int alloc_len;
3316	unsigned char select_report;
3317	u64 lun;
3318	struct scsi_lun *lun_p;
3319	u8 arr[RL_BUCKET_ELEMS * sizeof(struct scsi_lun)];
3320	unsigned int lun_cnt;	/* normal LUN count (max: 256) */
3321	unsigned int wlun_cnt;	/* report luns W-LUN count */
3322	unsigned int tlun_cnt;	/* total LUN count */
3323	unsigned int rlen;	/* response length (in bytes) */
3324	int k, j, n, res;
3325	unsigned int off_rsp = 0;
3326	const int sz_lun = sizeof(struct scsi_lun);
3327
3328	clear_luns_changed_on_target(devip);
3329
3330	select_report = cmd[2];
3331	alloc_len = get_unaligned_be32(cmd + 6);
3332
3333	if (alloc_len < 4) {
3334		pr_err("alloc len too small %d\n", alloc_len);
3335		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 6, -1);
3336		return check_condition_result;
3337	}
3338
3339	switch (select_report) {
3340	case 0:		/* all LUNs apart from W-LUNs */
3341		lun_cnt = sdebug_max_luns;
3342		wlun_cnt = 0;
3343		break;
3344	case 1:		/* only W-LUNs */
3345		lun_cnt = 0;
3346		wlun_cnt = 1;
3347		break;
3348	case 2:		/* all LUNs */
3349		lun_cnt = sdebug_max_luns;
3350		wlun_cnt = 1;
3351		break;
3352	case 0x10:	/* only administrative LUs */
3353	case 0x11:	/* see SPC-5 */
3354	case 0x12:	/* only subsiduary LUs owned by referenced LU */
3355	default:
3356		pr_debug("select report invalid %d\n", select_report);
3357		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, -1);
3358		return check_condition_result;
3359	}
3360
3361	if (sdebug_no_lun_0 && (lun_cnt > 0))
3362		--lun_cnt;
3363
3364	tlun_cnt = lun_cnt + wlun_cnt;
3365	rlen = tlun_cnt * sz_lun;	/* excluding 8 byte header */
3366	scsi_set_resid(scp, scsi_bufflen(scp));
3367	pr_debug("select_report %d luns = %d wluns = %d no_lun0 %d\n",
3368		 select_report, lun_cnt, wlun_cnt, sdebug_no_lun_0);
3369
3370	/* loops rely on sizeof response header same as sizeof lun (both 8) */
3371	lun = sdebug_no_lun_0 ? 1 : 0;
3372	for (k = 0, j = 0, res = 0; true; ++k, j = 0) {
3373		memset(arr, 0, sizeof(arr));
3374		lun_p = (struct scsi_lun *)&arr[0];
3375		if (k == 0) {
3376			put_unaligned_be32(rlen, &arr[0]);
3377			++lun_p;
3378			j = 1;
3379		}
3380		for ( ; j < RL_BUCKET_ELEMS; ++j, ++lun_p) {
3381			if ((k * RL_BUCKET_ELEMS) + j > lun_cnt)
3382				break;
3383			int_to_scsilun(lun++, lun_p);
3384		}
3385		if (j < RL_BUCKET_ELEMS)
3386			break;
3387		n = j * sz_lun;
3388		res = p_fill_from_dev_buffer(scp, arr, n, off_rsp);
3389		if (res)
3390			return res;
3391		off_rsp += n;
3392	}
3393	if (wlun_cnt) {
3394		int_to_scsilun(SCSI_W_LUN_REPORT_LUNS, lun_p);
3395		++j;
3396	}
3397	if (j > 0)
3398		res = p_fill_from_dev_buffer(scp, arr, j * sz_lun, off_rsp);
3399	return res;
3400}
3401
3402static int resp_xdwriteread(struct scsi_cmnd *scp, unsigned long long lba,
3403			    unsigned int num, struct sdebug_dev_info *devip)
3404{
3405	int j;
3406	unsigned char *kaddr, *buf;
3407	unsigned int offset;
3408	struct scsi_data_buffer *sdb = scsi_in(scp);
3409	struct sg_mapping_iter miter;
3410
3411	/* better not to use temporary buffer. */
3412	buf = kzalloc(scsi_bufflen(scp), GFP_ATOMIC);
3413	if (!buf) {
3414		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3415				INSUFF_RES_ASCQ);
3416		return check_condition_result;
3417	}
3418
3419	scsi_sg_copy_to_buffer(scp, buf, scsi_bufflen(scp));
3420
3421	offset = 0;
3422	sg_miter_start(&miter, sdb->table.sgl, sdb->table.nents,
3423			SG_MITER_ATOMIC | SG_MITER_TO_SG);
3424
3425	while (sg_miter_next(&miter)) {
3426		kaddr = miter.addr;
3427		for (j = 0; j < miter.length; j++)
3428			*(kaddr + j) ^= *(buf + offset + j);
3429
3430		offset += miter.length;
3431	}
3432	sg_miter_stop(&miter);
3433	kfree(buf);
3434
3435	return 0;
3436}
3437
3438static int resp_xdwriteread_10(struct scsi_cmnd *scp,
3439			       struct sdebug_dev_info *devip)
3440{
3441	u8 *cmd = scp->cmnd;
3442	u64 lba;
3443	u32 num;
3444	int errsts;
3445
3446	if (!scsi_bidi_cmnd(scp)) {
3447		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3448				INSUFF_RES_ASCQ);
3449		return check_condition_result;
3450	}
3451	errsts = resp_read_dt0(scp, devip);
3452	if (errsts)
3453		return errsts;
3454	if (!(cmd[1] & 0x4)) {		/* DISABLE_WRITE is not set */
3455		errsts = resp_write_dt0(scp, devip);
3456		if (errsts)
3457			return errsts;
3458	}
3459	lba = get_unaligned_be32(cmd + 2);
3460	num = get_unaligned_be16(cmd + 7);
3461	return resp_xdwriteread(scp, lba, num, devip);
3462}
3463
3464static struct sdebug_queue *get_queue(struct scsi_cmnd *cmnd)
3465{
3466	struct sdebug_queue *sqp = sdebug_q_arr;
3467
3468	if (sdebug_mq_active) {
3469		u32 tag = blk_mq_unique_tag(cmnd->request);
3470		u16 hwq = blk_mq_unique_tag_to_hwq(tag);
3471
3472		if (unlikely(hwq >= submit_queues)) {
3473			pr_warn("Unexpected hwq=%d, apply modulo\n", hwq);
3474			hwq %= submit_queues;
3475		}
3476		pr_debug("tag=%u, hwq=%d\n", tag, hwq);
3477		return sqp + hwq;
3478	} else
3479		return sqp;
3480}
3481
3482/* Queued (deferred) command completions converge here. */
3483static void sdebug_q_cmd_complete(struct sdebug_defer *sd_dp)
3484{
3485	int qc_idx;
3486	int retiring = 0;
3487	unsigned long iflags;
3488	struct sdebug_queue *sqp;
3489	struct sdebug_queued_cmd *sqcp;
3490	struct scsi_cmnd *scp;
3491	struct sdebug_dev_info *devip;
3492
3493	qc_idx = sd_dp->qc_idx;
3494	sqp = sdebug_q_arr + sd_dp->sqa_idx;
3495	if (sdebug_statistics) {
3496		atomic_inc(&sdebug_completions);
3497		if (raw_smp_processor_id() != sd_dp->issuing_cpu)
3498			atomic_inc(&sdebug_miss_cpus);
3499	}
3500	if (unlikely((qc_idx < 0) || (qc_idx >= SDEBUG_CANQUEUE))) {
3501		pr_err("wild qc_idx=%d\n", qc_idx);
3502		return;
3503	}
3504	spin_lock_irqsave(&sqp->qc_lock, iflags);
3505	sqcp = &sqp->qc_arr[qc_idx];
3506	scp = sqcp->a_cmnd;
3507	if (unlikely(scp == NULL)) {
3508		spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3509		pr_err("scp is NULL, sqa_idx=%d, qc_idx=%d\n",
3510		       sd_dp->sqa_idx, qc_idx);
3511		return;
3512	}
3513	devip = (struct sdebug_dev_info *)scp->device->hostdata;
3514	if (likely(devip))
3515		atomic_dec(&devip->num_in_q);
3516	else
3517		pr_err("devip=NULL\n");
3518	if (unlikely(atomic_read(&retired_max_queue) > 0))
3519		retiring = 1;
3520
3521	sqcp->a_cmnd = NULL;
3522	if (unlikely(!test_and_clear_bit(qc_idx, sqp->in_use_bm))) {
3523		spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3524		pr_err("Unexpected completion\n");
3525		return;
3526	}
3527
3528	if (unlikely(retiring)) {	/* user has reduced max_queue */
3529		int k, retval;
3530
3531		retval = atomic_read(&retired_max_queue);
3532		if (qc_idx >= retval) {
3533			spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3534			pr_err("index %d too large\n", retval);
3535			return;
3536		}
3537		k = find_last_bit(sqp->in_use_bm, retval);
3538		if ((k < sdebug_max_queue) || (k == retval))
3539			atomic_set(&retired_max_queue, 0);
3540		else
3541			atomic_set(&retired_max_queue, k + 1);
3542	}
3543	spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3544	scp->scsi_done(scp); /* callback to mid level */
3545}
3546
3547/* When high resolution timer goes off this function is called. */
3548static enum hrtimer_restart sdebug_q_cmd_hrt_complete(struct hrtimer *timer)
3549{
3550	struct sdebug_defer *sd_dp = container_of(timer, struct sdebug_defer,
3551						  hrt);
3552	sdebug_q_cmd_complete(sd_dp);
3553	return HRTIMER_NORESTART;
3554}
3555
3556/* When work queue schedules work, it calls this function. */
3557static void sdebug_q_cmd_wq_complete(struct work_struct *work)
3558{
3559	struct sdebug_defer *sd_dp = container_of(work, struct sdebug_defer,
3560						  ew.work);
3561	sdebug_q_cmd_complete(sd_dp);
3562}
3563
3564static bool got_shared_uuid;
3565static uuid_be shared_uuid;
3566
3567static struct sdebug_dev_info *sdebug_device_create(
3568			struct sdebug_host_info *sdbg_host, gfp_t flags)
3569{
3570	struct sdebug_dev_info *devip;
3571
3572	devip = kzalloc(sizeof(*devip), flags);
3573	if (devip) {
3574		if (sdebug_uuid_ctl == 1)
3575			uuid_be_gen(&devip->lu_name);
3576		else if (sdebug_uuid_ctl == 2) {
3577			if (got_shared_uuid)
3578				devip->lu_name = shared_uuid;
3579			else {
3580				uuid_be_gen(&shared_uuid);
3581				got_shared_uuid = true;
3582				devip->lu_name = shared_uuid;
3583			}
3584		}
3585		devip->sdbg_host = sdbg_host;
3586		list_add_tail(&devip->dev_list, &sdbg_host->dev_info_list);
3587	}
3588	return devip;
3589}
3590
3591static struct sdebug_dev_info *find_build_dev_info(struct scsi_device *sdev)
3592{
3593	struct sdebug_host_info *sdbg_host;
3594	struct sdebug_dev_info *open_devip = NULL;
3595	struct sdebug_dev_info *devip;
 
3596
 
 
3597	sdbg_host = *(struct sdebug_host_info **)shost_priv(sdev->host);
3598	if (!sdbg_host) {
3599		pr_err("Host info NULL\n");
3600		return NULL;
3601        }
3602	list_for_each_entry(devip, &sdbg_host->dev_info_list, dev_list) {
3603		if ((devip->used) && (devip->channel == sdev->channel) &&
3604                    (devip->target == sdev->id) &&
3605                    (devip->lun == sdev->lun))
3606                        return devip;
3607		else {
3608			if ((!devip->used) && (!open_devip))
3609				open_devip = devip;
3610		}
3611	}
3612	if (!open_devip) { /* try and make a new one */
3613		open_devip = sdebug_device_create(sdbg_host, GFP_ATOMIC);
3614		if (!open_devip) {
3615			pr_err("out of memory at line %d\n", __LINE__);
 
3616			return NULL;
3617		}
3618	}
3619
3620	open_devip->channel = sdev->channel;
3621	open_devip->target = sdev->id;
3622	open_devip->lun = sdev->lun;
3623	open_devip->sdbg_host = sdbg_host;
3624	atomic_set(&open_devip->num_in_q, 0);
3625	set_bit(SDEBUG_UA_POR, open_devip->uas_bm);
3626	open_devip->used = true;
 
 
 
 
 
 
 
 
 
3627	return open_devip;
3628}
3629
3630static int scsi_debug_slave_alloc(struct scsi_device *sdp)
3631{
3632	if (sdebug_verbose)
3633		pr_info("slave_alloc <%u %u %u %llu>\n",
3634		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
3635	queue_flag_set_unlocked(QUEUE_FLAG_BIDI, sdp->request_queue);
3636	return 0;
3637}
3638
3639static int scsi_debug_slave_configure(struct scsi_device *sdp)
3640{
3641	struct sdebug_dev_info *devip =
3642			(struct sdebug_dev_info *)sdp->hostdata;
3643
3644	if (sdebug_verbose)
3645		pr_info("slave_configure <%u %u %u %llu>\n",
3646		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
3647	if (sdp->host->max_cmd_len != SDEBUG_MAX_CMD_LEN)
3648		sdp->host->max_cmd_len = SDEBUG_MAX_CMD_LEN;
3649	if (devip == NULL) {
3650		devip = find_build_dev_info(sdp);
3651		if (devip == NULL)
3652			return 1;  /* no resources, will be marked offline */
3653	}
3654	sdp->hostdata = devip;
3655	blk_queue_max_segment_size(sdp->request_queue, -1U);
3656	if (sdebug_no_uld)
 
 
 
3657		sdp->no_uld_attach = 1;
3658	return 0;
3659}
3660
3661static void scsi_debug_slave_destroy(struct scsi_device *sdp)
3662{
3663	struct sdebug_dev_info *devip =
3664		(struct sdebug_dev_info *)sdp->hostdata;
3665
3666	if (sdebug_verbose)
3667		pr_info("slave_destroy <%u %u %u %llu>\n",
3668		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
3669	if (devip) {
3670		/* make this slot available for re-use */
3671		devip->used = false;
3672		sdp->hostdata = NULL;
3673	}
3674}
3675
3676static void stop_qc_helper(struct sdebug_defer *sd_dp)
3677{
3678	if (!sd_dp)
3679		return;
3680	if ((sdebug_jdelay > 0) || (sdebug_ndelay > 0))
3681		hrtimer_cancel(&sd_dp->hrt);
3682	else if (sdebug_jdelay < 0)
3683		cancel_work_sync(&sd_dp->ew.work);
3684}
3685
3686/* If @cmnd found deletes its timer or work queue and returns true; else
3687   returns false */
3688static bool stop_queued_cmnd(struct scsi_cmnd *cmnd)
3689{
3690	unsigned long iflags;
3691	int j, k, qmax, r_qmax;
3692	struct sdebug_queue *sqp;
3693	struct sdebug_queued_cmd *sqcp;
3694	struct sdebug_dev_info *devip;
3695	struct sdebug_defer *sd_dp;
3696
3697	for (j = 0, sqp = sdebug_q_arr; j < submit_queues; ++j, ++sqp) {
3698		spin_lock_irqsave(&sqp->qc_lock, iflags);
3699		qmax = sdebug_max_queue;
3700		r_qmax = atomic_read(&retired_max_queue);
3701		if (r_qmax > qmax)
3702			qmax = r_qmax;
3703		for (k = 0; k < qmax; ++k) {
3704			if (test_bit(k, sqp->in_use_bm)) {
3705				sqcp = &sqp->qc_arr[k];
3706				if (cmnd != sqcp->a_cmnd)
3707					continue;
3708				/* found */
3709				devip = (struct sdebug_dev_info *)
3710						cmnd->device->hostdata;
3711				if (devip)
3712					atomic_dec(&devip->num_in_q);
3713				sqcp->a_cmnd = NULL;
3714				sd_dp = sqcp->sd_dp;
3715				spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3716				stop_qc_helper(sd_dp);
3717				clear_bit(k, sqp->in_use_bm);
3718				return true;
3719			}
3720		}
3721		spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3722	}
3723	return false;
 
3724}
3725
3726/* Deletes (stops) timers or work queues of all queued commands */
3727static void stop_all_queued(void)
3728{
3729	unsigned long iflags;
3730	int j, k;
3731	struct sdebug_queue *sqp;
3732	struct sdebug_queued_cmd *sqcp;
3733	struct sdebug_dev_info *devip;
3734	struct sdebug_defer *sd_dp;
3735
3736	for (j = 0, sqp = sdebug_q_arr; j < submit_queues; ++j, ++sqp) {
3737		spin_lock_irqsave(&sqp->qc_lock, iflags);
3738		for (k = 0; k < SDEBUG_CANQUEUE; ++k) {
3739			if (test_bit(k, sqp->in_use_bm)) {
3740				sqcp = &sqp->qc_arr[k];
3741				if (sqcp->a_cmnd == NULL)
3742					continue;
3743				devip = (struct sdebug_dev_info *)
3744					sqcp->a_cmnd->device->hostdata;
3745				if (devip)
3746					atomic_dec(&devip->num_in_q);
3747				sqcp->a_cmnd = NULL;
3748				sd_dp = sqcp->sd_dp;
3749				spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3750				stop_qc_helper(sd_dp);
3751				clear_bit(k, sqp->in_use_bm);
3752				spin_lock_irqsave(&sqp->qc_lock, iflags);
3753			}
3754		}
3755		spin_unlock_irqrestore(&sqp->qc_lock, iflags);
3756	}
 
3757}
3758
3759/* Free queued command memory on heap */
3760static void free_all_queued(void)
3761{
3762	int j, k;
3763	struct sdebug_queue *sqp;
3764	struct sdebug_queued_cmd *sqcp;
3765
3766	for (j = 0, sqp = sdebug_q_arr; j < submit_queues; ++j, ++sqp) {
3767		for (k = 0; k < SDEBUG_CANQUEUE; ++k) {
3768			sqcp = &sqp->qc_arr[k];
3769			kfree(sqcp->sd_dp);
3770			sqcp->sd_dp = NULL;
3771		}
3772	}
3773}
3774
3775static int scsi_debug_abort(struct scsi_cmnd *SCpnt)
 
3776{
3777	bool ok;
 
3778
3779	++num_aborts;
3780	if (SCpnt) {
3781		ok = stop_queued_cmnd(SCpnt);
3782		if (SCpnt->device && (SDEBUG_OPT_ALL_NOISE & sdebug_opts))
3783			sdev_printk(KERN_INFO, SCpnt->device,
3784				    "%s: command%s found\n", __func__,
3785				    ok ? "" : " not");
 
 
3786	}
3787	return SUCCESS;
 
 
 
3788}
3789
3790static int scsi_debug_device_reset(struct scsi_cmnd * SCpnt)
3791{
 
 
 
 
3792	++num_dev_resets;
3793	if (SCpnt && SCpnt->device) {
3794		struct scsi_device *sdp = SCpnt->device;
3795		struct sdebug_dev_info *devip =
3796				(struct sdebug_dev_info *)sdp->hostdata;
3797
3798		if (SDEBUG_OPT_ALL_NOISE & sdebug_opts)
3799			sdev_printk(KERN_INFO, sdp, "%s\n", __func__);
3800		if (devip)
3801			set_bit(SDEBUG_UA_POR, devip->uas_bm);
3802	}
3803	return SUCCESS;
3804}
3805
3806static int scsi_debug_target_reset(struct scsi_cmnd *SCpnt)
3807{
3808	struct sdebug_host_info *sdbg_host;
3809	struct sdebug_dev_info *devip;
3810	struct scsi_device *sdp;
3811	struct Scsi_Host *hp;
3812	int k = 0;
3813
3814	++num_target_resets;
3815	if (!SCpnt)
3816		goto lie;
3817	sdp = SCpnt->device;
3818	if (!sdp)
3819		goto lie;
3820	if (SDEBUG_OPT_ALL_NOISE & sdebug_opts)
3821		sdev_printk(KERN_INFO, sdp, "%s\n", __func__);
3822	hp = sdp->host;
3823	if (!hp)
3824		goto lie;
3825	sdbg_host = *(struct sdebug_host_info **)shost_priv(hp);
3826	if (sdbg_host) {
3827		list_for_each_entry(devip,
3828				    &sdbg_host->dev_info_list,
3829				    dev_list)
3830			if (devip->target == sdp->id) {
3831				set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3832				++k;
3833			}
3834	}
3835	if (SDEBUG_OPT_RESET_NOISE & sdebug_opts)
3836		sdev_printk(KERN_INFO, sdp,
3837			    "%s: %d device(s) found in target\n", __func__, k);
3838lie:
3839	return SUCCESS;
3840}
3841
3842static int scsi_debug_bus_reset(struct scsi_cmnd * SCpnt)
3843{
3844	struct sdebug_host_info *sdbg_host;
3845	struct sdebug_dev_info *devip;
3846        struct scsi_device * sdp;
3847        struct Scsi_Host * hp;
3848	int k = 0;
3849
 
 
3850	++num_bus_resets;
3851	if (!(SCpnt && SCpnt->device))
3852		goto lie;
3853	sdp = SCpnt->device;
3854	if (SDEBUG_OPT_ALL_NOISE & sdebug_opts)
3855		sdev_printk(KERN_INFO, sdp, "%s\n", __func__);
3856	hp = sdp->host;
3857	if (hp) {
3858		sdbg_host = *(struct sdebug_host_info **)shost_priv(hp);
3859		if (sdbg_host) {
3860			list_for_each_entry(devip,
3861                                            &sdbg_host->dev_info_list,
3862					    dev_list) {
3863				set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3864				++k;
3865			}
3866		}
3867	}
3868	if (SDEBUG_OPT_RESET_NOISE & sdebug_opts)
3869		sdev_printk(KERN_INFO, sdp,
3870			    "%s: %d device(s) found in host\n", __func__, k);
3871lie:
3872	return SUCCESS;
3873}
3874
3875static int scsi_debug_host_reset(struct scsi_cmnd * SCpnt)
3876{
3877	struct sdebug_host_info * sdbg_host;
3878	struct sdebug_dev_info *devip;
3879	int k = 0;
3880
 
 
3881	++num_host_resets;
3882	if ((SCpnt->device) && (SDEBUG_OPT_ALL_NOISE & sdebug_opts))
3883		sdev_printk(KERN_INFO, SCpnt->device, "%s\n", __func__);
3884        spin_lock(&sdebug_host_list_lock);
3885        list_for_each_entry(sdbg_host, &sdebug_host_list, host_list) {
3886		list_for_each_entry(devip, &sdbg_host->dev_info_list,
3887				    dev_list) {
3888			set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3889			++k;
3890		}
3891        }
3892        spin_unlock(&sdebug_host_list_lock);
3893	stop_all_queued();
3894	if (SDEBUG_OPT_RESET_NOISE & sdebug_opts)
3895		sdev_printk(KERN_INFO, SCpnt->device,
3896			    "%s: %d device(s) found\n", __func__, k);
3897	return SUCCESS;
3898}
3899
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3900static void __init sdebug_build_parts(unsigned char *ramp,
3901				      unsigned long store_size)
3902{
3903	struct partition * pp;
3904	int starts[SDEBUG_MAX_PARTS + 2];
3905	int sectors_per_part, num_sectors, k;
3906	int heads_by_sects, start_sec, end_sec;
3907
3908	/* assume partition table already zeroed */
3909	if ((sdebug_num_parts < 1) || (store_size < 1048576))
3910		return;
3911	if (sdebug_num_parts > SDEBUG_MAX_PARTS) {
3912		sdebug_num_parts = SDEBUG_MAX_PARTS;
3913		pr_warn("reducing partitions to %d\n", SDEBUG_MAX_PARTS);
 
3914	}
3915	num_sectors = (int)sdebug_store_sectors;
3916	sectors_per_part = (num_sectors - sdebug_sectors_per)
3917			   / sdebug_num_parts;
3918	heads_by_sects = sdebug_heads * sdebug_sectors_per;
3919        starts[0] = sdebug_sectors_per;
3920	for (k = 1; k < sdebug_num_parts; ++k)
3921		starts[k] = ((k * sectors_per_part) / heads_by_sects)
3922			    * heads_by_sects;
3923	starts[sdebug_num_parts] = num_sectors;
3924	starts[sdebug_num_parts + 1] = 0;
3925
3926	ramp[510] = 0x55;	/* magic partition markings */
3927	ramp[511] = 0xAA;
3928	pp = (struct partition *)(ramp + 0x1be);
3929	for (k = 0; starts[k + 1]; ++k, ++pp) {
3930		start_sec = starts[k];
3931		end_sec = starts[k + 1] - 1;
3932		pp->boot_ind = 0;
3933
3934		pp->cyl = start_sec / heads_by_sects;
3935		pp->head = (start_sec - (pp->cyl * heads_by_sects))
3936			   / sdebug_sectors_per;
3937		pp->sector = (start_sec % sdebug_sectors_per) + 1;
3938
3939		pp->end_cyl = end_sec / heads_by_sects;
3940		pp->end_head = (end_sec - (pp->end_cyl * heads_by_sects))
3941			       / sdebug_sectors_per;
3942		pp->end_sector = (end_sec % sdebug_sectors_per) + 1;
3943
3944		pp->start_sect = cpu_to_le32(start_sec);
3945		pp->nr_sects = cpu_to_le32(end_sec - start_sec + 1);
3946		pp->sys_ind = 0x83;	/* plain Linux partition */
3947	}
3948}
3949
3950static void block_unblock_all_queues(bool block)
 
 
3951{
3952	int j;
3953	struct sdebug_queue *sqp;
 
3954
3955	for (j = 0, sqp = sdebug_q_arr; j < submit_queues; ++j, ++sqp)
3956		atomic_set(&sqp->blocked, (int)block);
3957}
3958
3959/* Adjust (by rounding down) the sdebug_cmnd_count so abs(every_nth)-1
3960 * commands will be processed normally before triggers occur.
3961 */
3962static void tweak_cmnd_count(void)
3963{
3964	int count, modulo;
3965
3966	modulo = abs(sdebug_every_nth);
3967	if (modulo < 2)
3968		return;
3969	block_unblock_all_queues(true);
3970	count = atomic_read(&sdebug_cmnd_count);
3971	atomic_set(&sdebug_cmnd_count, (count / modulo) * modulo);
3972	block_unblock_all_queues(false);
3973}
3974
3975static void clear_queue_stats(void)
3976{
3977	atomic_set(&sdebug_cmnd_count, 0);
3978	atomic_set(&sdebug_completions, 0);
3979	atomic_set(&sdebug_miss_cpus, 0);
3980	atomic_set(&sdebug_a_tsf, 0);
3981}
3982
3983static void setup_inject(struct sdebug_queue *sqp,
3984			 struct sdebug_queued_cmd *sqcp)
3985{
3986	if ((atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth)) > 0)
3987		return;
3988	sqcp->inj_recovered = !!(SDEBUG_OPT_RECOVERED_ERR & sdebug_opts);
3989	sqcp->inj_transport = !!(SDEBUG_OPT_TRANSPORT_ERR & sdebug_opts);
3990	sqcp->inj_dif = !!(SDEBUG_OPT_DIF_ERR & sdebug_opts);
3991	sqcp->inj_dix = !!(SDEBUG_OPT_DIX_ERR & sdebug_opts);
3992	sqcp->inj_short = !!(SDEBUG_OPT_SHORT_TRANSFER & sdebug_opts);
3993}
3994
3995/* Complete the processing of the thread that queued a SCSI command to this
3996 * driver. It either completes the command by calling cmnd_done() or
3997 * schedules a hr timer or work queue then returns 0. Returns
3998 * SCSI_MLQUEUE_HOST_BUSY if temporarily out of resources.
3999 */
4000static int schedule_resp(struct scsi_cmnd *cmnd, struct sdebug_dev_info *devip,
4001			 int scsi_result, int delta_jiff)
4002{
4003	unsigned long iflags;
4004	int k, num_in_q, qdepth, inject;
4005	struct sdebug_queue *sqp;
4006	struct sdebug_queued_cmd *sqcp;
4007	struct scsi_device *sdp;
4008	struct sdebug_defer *sd_dp;
4009
4010	if (unlikely(devip == NULL)) {
4011		if (scsi_result == 0)
4012			scsi_result = DID_NO_CONNECT << 16;
4013		goto respond_in_thread;
4014	}
4015	sdp = cmnd->device;
4016
4017	if (unlikely(sdebug_verbose && scsi_result))
4018		sdev_printk(KERN_INFO, sdp, "%s: non-zero result=0x%x\n",
4019			    __func__, scsi_result);
4020	if (delta_jiff == 0)
4021		goto respond_in_thread;
4022
4023	/* schedule the response at a later time if resources permit */
4024	sqp = get_queue(cmnd);
4025	spin_lock_irqsave(&sqp->qc_lock, iflags);
4026	if (unlikely(atomic_read(&sqp->blocked))) {
4027		spin_unlock_irqrestore(&sqp->qc_lock, iflags);
4028		return SCSI_MLQUEUE_HOST_BUSY;
4029	}
4030	num_in_q = atomic_read(&devip->num_in_q);
4031	qdepth = cmnd->device->queue_depth;
4032	inject = 0;
4033	if (unlikely((qdepth > 0) && (num_in_q >= qdepth))) {
4034		if (scsi_result) {
4035			spin_unlock_irqrestore(&sqp->qc_lock, iflags);
4036			goto respond_in_thread;
4037		} else
4038			scsi_result = device_qfull_result;
4039	} else if (unlikely(sdebug_every_nth &&
4040			    (SDEBUG_OPT_RARE_TSF & sdebug_opts) &&
4041			    (scsi_result == 0))) {
4042		if ((num_in_q == (qdepth - 1)) &&
4043		    (atomic_inc_return(&sdebug_a_tsf) >=
4044		     abs(sdebug_every_nth))) {
4045			atomic_set(&sdebug_a_tsf, 0);
4046			inject = 1;
4047			scsi_result = device_qfull_result;
4048		}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4049	}
4050
4051	k = find_first_zero_bit(sqp->in_use_bm, sdebug_max_queue);
4052	if (unlikely(k >= sdebug_max_queue)) {
4053		spin_unlock_irqrestore(&sqp->qc_lock, iflags);
4054		if (scsi_result)
4055			goto respond_in_thread;
4056		else if (SDEBUG_OPT_ALL_TSF & sdebug_opts)
4057			scsi_result = device_qfull_result;
4058		if (SDEBUG_OPT_Q_NOISE & sdebug_opts)
4059			sdev_printk(KERN_INFO, sdp,
4060				    "%s: max_queue=%d exceeded, %s\n",
4061				    __func__, sdebug_max_queue,
4062				    (scsi_result ?  "status: TASK SET FULL" :
4063						    "report: host busy"));
4064		if (scsi_result)
4065			goto respond_in_thread;
4066		else
4067			return SCSI_MLQUEUE_HOST_BUSY;
4068	}
4069	__set_bit(k, sqp->in_use_bm);
4070	atomic_inc(&devip->num_in_q);
4071	sqcp = &sqp->qc_arr[k];
4072	sqcp->a_cmnd = cmnd;
4073	cmnd->host_scribble = (unsigned char *)sqcp;
4074	cmnd->result = scsi_result;
4075	sd_dp = sqcp->sd_dp;
4076	spin_unlock_irqrestore(&sqp->qc_lock, iflags);
4077	if (unlikely(sdebug_every_nth && sdebug_any_injecting_opt))
4078		setup_inject(sqp, sqcp);
4079	if (delta_jiff > 0 || sdebug_ndelay > 0) {
4080		ktime_t kt;
4081
4082		if (delta_jiff > 0) {
4083			struct timespec ts;
4084
4085			jiffies_to_timespec(delta_jiff, &ts);
4086			kt = ktime_set(ts.tv_sec, ts.tv_nsec);
4087		} else
4088			kt = sdebug_ndelay;
4089		if (NULL == sd_dp) {
4090			sd_dp = kzalloc(sizeof(*sd_dp), GFP_ATOMIC);
4091			if (NULL == sd_dp)
4092				return SCSI_MLQUEUE_HOST_BUSY;
4093			sqcp->sd_dp = sd_dp;
4094			hrtimer_init(&sd_dp->hrt, CLOCK_MONOTONIC,
4095				     HRTIMER_MODE_REL_PINNED);
4096			sd_dp->hrt.function = sdebug_q_cmd_hrt_complete;
4097			sd_dp->sqa_idx = sqp - sdebug_q_arr;
4098			sd_dp->qc_idx = k;
4099		}
4100		if (sdebug_statistics)
4101			sd_dp->issuing_cpu = raw_smp_processor_id();
4102		hrtimer_start(&sd_dp->hrt, kt, HRTIMER_MODE_REL_PINNED);
4103	} else {	/* jdelay < 0, use work queue */
4104		if (NULL == sd_dp) {
4105			sd_dp = kzalloc(sizeof(*sqcp->sd_dp), GFP_ATOMIC);
4106			if (NULL == sd_dp)
4107				return SCSI_MLQUEUE_HOST_BUSY;
4108			sqcp->sd_dp = sd_dp;
4109			sd_dp->sqa_idx = sqp - sdebug_q_arr;
4110			sd_dp->qc_idx = k;
4111			INIT_WORK(&sd_dp->ew.work, sdebug_q_cmd_wq_complete);
4112		}
4113		if (sdebug_statistics)
4114			sd_dp->issuing_cpu = raw_smp_processor_id();
4115		schedule_work(&sd_dp->ew.work);
4116	}
4117	if (unlikely((SDEBUG_OPT_Q_NOISE & sdebug_opts) &&
4118		     (scsi_result == device_qfull_result)))
4119		sdev_printk(KERN_INFO, sdp,
4120			    "%s: num_in_q=%d +1, %s%s\n", __func__,
4121			    num_in_q, (inject ? "<inject> " : ""),
4122			    "status: TASK SET FULL");
4123	return 0;
4124
4125respond_in_thread:	/* call back to mid-layer using invocation thread */
4126	cmnd->result = scsi_result;
4127	cmnd->scsi_done(cmnd);
4128	return 0;
4129}
4130
4131/* Note: The following macros create attribute files in the
4132   /sys/module/scsi_debug/parameters directory. Unfortunately this
4133   driver is unaware of a change and cannot trigger auxiliary actions
4134   as it can when the corresponding attribute in the
4135   /sys/bus/pseudo/drivers/scsi_debug directory is changed.
4136 */
4137module_param_named(add_host, sdebug_add_host, int, S_IRUGO | S_IWUSR);
4138module_param_named(ato, sdebug_ato, int, S_IRUGO);
4139module_param_named(clustering, sdebug_clustering, bool, S_IRUGO | S_IWUSR);
4140module_param_named(delay, sdebug_jdelay, int, S_IRUGO | S_IWUSR);
4141module_param_named(dev_size_mb, sdebug_dev_size_mb, int, S_IRUGO);
4142module_param_named(dif, sdebug_dif, int, S_IRUGO);
4143module_param_named(dix, sdebug_dix, int, S_IRUGO);
4144module_param_named(dsense, sdebug_dsense, int, S_IRUGO | S_IWUSR);
4145module_param_named(every_nth, sdebug_every_nth, int, S_IRUGO | S_IWUSR);
4146module_param_named(fake_rw, sdebug_fake_rw, int, S_IRUGO | S_IWUSR);
4147module_param_named(guard, sdebug_guard, uint, S_IRUGO);
4148module_param_named(host_lock, sdebug_host_lock, bool, S_IRUGO | S_IWUSR);
4149module_param_named(lbpu, sdebug_lbpu, int, S_IRUGO);
4150module_param_named(lbpws, sdebug_lbpws, int, S_IRUGO);
4151module_param_named(lbpws10, sdebug_lbpws10, int, S_IRUGO);
4152module_param_named(lbprz, sdebug_lbprz, int, S_IRUGO);
4153module_param_named(lowest_aligned, sdebug_lowest_aligned, int, S_IRUGO);
4154module_param_named(max_luns, sdebug_max_luns, int, S_IRUGO | S_IWUSR);
4155module_param_named(max_queue, sdebug_max_queue, int, S_IRUGO | S_IWUSR);
4156module_param_named(ndelay, sdebug_ndelay, int, S_IRUGO | S_IWUSR);
4157module_param_named(no_lun_0, sdebug_no_lun_0, int, S_IRUGO | S_IWUSR);
4158module_param_named(no_uld, sdebug_no_uld, int, S_IRUGO);
4159module_param_named(num_parts, sdebug_num_parts, int, S_IRUGO);
4160module_param_named(num_tgts, sdebug_num_tgts, int, S_IRUGO | S_IWUSR);
4161module_param_named(opt_blks, sdebug_opt_blks, int, S_IRUGO);
4162module_param_named(opts, sdebug_opts, int, S_IRUGO | S_IWUSR);
4163module_param_named(physblk_exp, sdebug_physblk_exp, int, S_IRUGO);
4164module_param_named(ptype, sdebug_ptype, int, S_IRUGO | S_IWUSR);
4165module_param_named(removable, sdebug_removable, bool, S_IRUGO | S_IWUSR);
4166module_param_named(scsi_level, sdebug_scsi_level, int, S_IRUGO);
4167module_param_named(sector_size, sdebug_sector_size, int, S_IRUGO);
4168module_param_named(statistics, sdebug_statistics, bool, S_IRUGO | S_IWUSR);
4169module_param_named(strict, sdebug_strict, bool, S_IRUGO | S_IWUSR);
4170module_param_named(submit_queues, submit_queues, int, S_IRUGO);
4171module_param_named(unmap_alignment, sdebug_unmap_alignment, int, S_IRUGO);
4172module_param_named(unmap_granularity, sdebug_unmap_granularity, int, S_IRUGO);
4173module_param_named(unmap_max_blocks, sdebug_unmap_max_blocks, int, S_IRUGO);
4174module_param_named(unmap_max_desc, sdebug_unmap_max_desc, int, S_IRUGO);
4175module_param_named(virtual_gb, sdebug_virtual_gb, int, S_IRUGO | S_IWUSR);
4176module_param_named(uuid_ctl, sdebug_uuid_ctl, int, S_IRUGO);
4177module_param_named(vpd_use_hostno, sdebug_vpd_use_hostno, int,
4178		   S_IRUGO | S_IWUSR);
4179module_param_named(write_same_length, sdebug_write_same_length, int,
4180		   S_IRUGO | S_IWUSR);
4181
4182MODULE_AUTHOR("Eric Youngdale + Douglas Gilbert");
4183MODULE_DESCRIPTION("SCSI debug adapter driver");
4184MODULE_LICENSE("GPL");
4185MODULE_VERSION(SDEBUG_VERSION);
4186
4187MODULE_PARM_DESC(add_host, "0..127 hosts allowed(def=1)");
4188MODULE_PARM_DESC(ato, "application tag ownership: 0=disk 1=host (def=1)");
4189MODULE_PARM_DESC(clustering, "when set enables larger transfers (def=0)");
4190MODULE_PARM_DESC(delay, "response delay (def=1 jiffy); 0:imm, -1,-2:tiny");
4191MODULE_PARM_DESC(dev_size_mb, "size in MiB of ram shared by devs(def=8)");
4192MODULE_PARM_DESC(dif, "data integrity field type: 0-3 (def=0)");
4193MODULE_PARM_DESC(dix, "data integrity extensions mask (def=0)");
4194MODULE_PARM_DESC(dsense, "use descriptor sense format(def=0 -> fixed)");
4195MODULE_PARM_DESC(every_nth, "timeout every nth command(def=0)");
4196MODULE_PARM_DESC(fake_rw, "fake reads/writes instead of copying (def=0)");
4197MODULE_PARM_DESC(guard, "protection checksum: 0=crc, 1=ip (def=0)");
4198MODULE_PARM_DESC(host_lock, "host_lock is ignored (def=0)");
4199MODULE_PARM_DESC(lbpu, "enable LBP, support UNMAP command (def=0)");
4200MODULE_PARM_DESC(lbpws, "enable LBP, support WRITE SAME(16) with UNMAP bit (def=0)");
4201MODULE_PARM_DESC(lbpws10, "enable LBP, support WRITE SAME(10) with UNMAP bit (def=0)");
4202MODULE_PARM_DESC(lbprz,
4203	"on read unmapped LBs return 0 when 1 (def), return 0xff when 2");
4204MODULE_PARM_DESC(lowest_aligned, "lowest aligned lba (def=0)");
4205MODULE_PARM_DESC(max_luns, "number of LUNs per target to simulate(def=1)");
4206MODULE_PARM_DESC(max_queue, "max number of queued commands (1 to max(def))");
4207MODULE_PARM_DESC(ndelay, "response delay in nanoseconds (def=0 -> ignore)");
4208MODULE_PARM_DESC(no_lun_0, "no LU number 0 (def=0 -> have lun 0)");
4209MODULE_PARM_DESC(no_uld, "stop ULD (e.g. sd driver) attaching (def=0))");
4210MODULE_PARM_DESC(num_parts, "number of partitions(def=0)");
4211MODULE_PARM_DESC(num_tgts, "number of targets per host to simulate(def=1)");
4212MODULE_PARM_DESC(opt_blks, "optimal transfer length in blocks (def=1024)");
4213MODULE_PARM_DESC(opts, "1->noise, 2->medium_err, 4->timeout, 8->recovered_err... (def=0)");
4214MODULE_PARM_DESC(physblk_exp, "physical block exponent (def=0)");
4215MODULE_PARM_DESC(ptype, "SCSI peripheral type(def=0[disk])");
4216MODULE_PARM_DESC(removable, "claim to have removable media (def=0)");
4217MODULE_PARM_DESC(scsi_level, "SCSI level to simulate(def=7[SPC-5])");
4218MODULE_PARM_DESC(sector_size, "logical block size in bytes (def=512)");
4219MODULE_PARM_DESC(statistics, "collect statistics on commands, queues (def=0)");
4220MODULE_PARM_DESC(strict, "stricter checks: reserved field in cdb (def=0)");
4221MODULE_PARM_DESC(submit_queues, "support for block multi-queue (def=1)");
4222MODULE_PARM_DESC(unmap_alignment, "lowest aligned thin provisioning lba (def=0)");
4223MODULE_PARM_DESC(unmap_granularity, "thin provisioning granularity in blocks (def=1)");
4224MODULE_PARM_DESC(unmap_max_blocks, "max # of blocks can be unmapped in one cmd (def=0xffffffff)");
4225MODULE_PARM_DESC(unmap_max_desc, "max # of ranges that can be unmapped in one cmd (def=256)");
4226MODULE_PARM_DESC(uuid_ctl,
4227		 "1->use uuid for lu name, 0->don't, 2->all use same (def=0)");
4228MODULE_PARM_DESC(virtual_gb, "virtual gigabyte (GiB) size (def=0 -> use dev_size_mb)");
4229MODULE_PARM_DESC(vpd_use_hostno, "0 -> dev ids ignore hostno (def=1 -> unique dev ids)");
4230MODULE_PARM_DESC(write_same_length, "Maximum blocks per WRITE SAME cmd (def=0xffff)");
4231
4232#define SDEBUG_INFO_LEN 256
4233static char sdebug_info[SDEBUG_INFO_LEN];
4234
4235static const char * scsi_debug_info(struct Scsi_Host * shp)
4236{
4237	int k;
4238
4239	k = scnprintf(sdebug_info, SDEBUG_INFO_LEN, "%s: version %s [%s]\n",
4240		      my_name, SDEBUG_VERSION, sdebug_version_date);
4241	if (k >= (SDEBUG_INFO_LEN - 1))
4242		return sdebug_info;
4243	scnprintf(sdebug_info + k, SDEBUG_INFO_LEN - k,
4244		  "  dev_size_mb=%d, opts=0x%x, submit_queues=%d, %s=%d",
4245		  sdebug_dev_size_mb, sdebug_opts, submit_queues,
4246		  "statistics", (int)sdebug_statistics);
4247	return sdebug_info;
4248}
4249
4250/* 'echo <val> > /proc/scsi/scsi_debug/<host_id>' writes to opts */
4251static int scsi_debug_write_info(struct Scsi_Host *host, char *buffer,
4252				 int length)
 
4253{
4254	char arr[16];
4255	int opts;
4256	int minLen = length > 15 ? 15 : length;
4257
4258	if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO))
4259		return -EACCES;
4260	memcpy(arr, buffer, minLen);
4261	arr[minLen] = '\0';
4262	if (1 != sscanf(arr, "%d", &opts))
4263		return -EINVAL;
4264	sdebug_opts = opts;
4265	sdebug_verbose = !!(SDEBUG_OPT_NOISE & opts);
4266	sdebug_any_injecting_opt = !!(SDEBUG_OPT_ALL_INJECTING & opts);
4267	if (sdebug_every_nth != 0)
4268		tweak_cmnd_count();
4269	return length;
4270}
4271
4272/* Output seen with 'cat /proc/scsi/scsi_debug/<host_id>'. It will be the
4273 * same for each scsi_debug host (if more than one). Some of the counters
4274 * output are not atomics so might be inaccurate in a busy system. */
4275static int scsi_debug_show_info(struct seq_file *m, struct Scsi_Host *host)
4276{
4277	int f, j, l;
4278	struct sdebug_queue *sqp;
4279
4280	seq_printf(m, "scsi_debug adapter driver, version %s [%s]\n",
4281		   SDEBUG_VERSION, sdebug_version_date);
4282	seq_printf(m, "num_tgts=%d, %ssize=%d MB, opts=0x%x, every_nth=%d\n",
4283		   sdebug_num_tgts, "shared (ram) ", sdebug_dev_size_mb,
4284		   sdebug_opts, sdebug_every_nth);
4285	seq_printf(m, "delay=%d, ndelay=%d, max_luns=%d, sector_size=%d %s\n",
4286		   sdebug_jdelay, sdebug_ndelay, sdebug_max_luns,
4287		   sdebug_sector_size, "bytes");
4288	seq_printf(m, "cylinders=%d, heads=%d, sectors=%d, command aborts=%d\n",
4289		   sdebug_cylinders_per, sdebug_heads, sdebug_sectors_per,
4290		   num_aborts);
4291	seq_printf(m, "RESETs: device=%d, target=%d, bus=%d, host=%d\n",
4292		   num_dev_resets, num_target_resets, num_bus_resets,
4293		   num_host_resets);
4294	seq_printf(m, "dix_reads=%d, dix_writes=%d, dif_errors=%d\n",
4295		   dix_reads, dix_writes, dif_errors);
4296	seq_printf(m, "usec_in_jiffy=%lu, %s=%d, mq_active=%d\n",
4297		   TICK_NSEC / 1000, "statistics", sdebug_statistics,
4298		   sdebug_mq_active);
4299	seq_printf(m, "cmnd_count=%d, completions=%d, %s=%d, a_tsf=%d\n",
4300		   atomic_read(&sdebug_cmnd_count),
4301		   atomic_read(&sdebug_completions),
4302		   "miss_cpus", atomic_read(&sdebug_miss_cpus),
4303		   atomic_read(&sdebug_a_tsf));
4304
4305	seq_printf(m, "submit_queues=%d\n", submit_queues);
4306	for (j = 0, sqp = sdebug_q_arr; j < submit_queues; ++j, ++sqp) {
4307		seq_printf(m, "  queue %d:\n", j);
4308		f = find_first_bit(sqp->in_use_bm, sdebug_max_queue);
4309		if (f != sdebug_max_queue) {
4310			l = find_last_bit(sqp->in_use_bm, sdebug_max_queue);
4311			seq_printf(m, "    in_use_bm BUSY: %s: %d,%d\n",
4312				   "first,last bits", f, l);
4313		}
4314	}
4315	return 0;
4316}
4317
4318static ssize_t delay_show(struct device_driver *ddp, char *buf)
4319{
4320	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_jdelay);
4321}
4322/* Returns -EBUSY if jdelay is being changed and commands are queued. The unit
4323 * of delay is jiffies.
4324 */
4325static ssize_t delay_store(struct device_driver *ddp, const char *buf,
4326			   size_t count)
4327{
4328	int jdelay, res;
 
4329
4330	if (count > 0 && sscanf(buf, "%d", &jdelay) == 1) {
4331		res = count;
4332		if (sdebug_jdelay != jdelay) {
4333			int j, k;
4334			struct sdebug_queue *sqp;
4335
4336			block_unblock_all_queues(true);
4337			for (j = 0, sqp = sdebug_q_arr; j < submit_queues;
4338			     ++j, ++sqp) {
4339				k = find_first_bit(sqp->in_use_bm,
4340						   sdebug_max_queue);
4341				if (k != sdebug_max_queue) {
4342					res = -EBUSY;   /* queued commands */
4343					break;
4344				}
4345			}
4346			if (res > 0) {
4347				/* make sure sdebug_defer instances get
4348				 * re-allocated for new delay variant */
4349				free_all_queued();
4350				sdebug_jdelay = jdelay;
4351				sdebug_ndelay = 0;
4352			}
4353			block_unblock_all_queues(false);
4354		}
4355		return res;
4356	}
4357	return -EINVAL;
4358}
4359static DRIVER_ATTR_RW(delay);
4360
4361static ssize_t ndelay_show(struct device_driver *ddp, char *buf)
4362{
4363	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_ndelay);
4364}
4365/* Returns -EBUSY if ndelay is being changed and commands are queued */
4366/* If > 0 and accepted then sdebug_jdelay is set to JDELAY_OVERRIDDEN */
4367static ssize_t ndelay_store(struct device_driver *ddp, const char *buf,
4368			    size_t count)
4369{
4370	int ndelay, res;
4371
4372	if ((count > 0) && (1 == sscanf(buf, "%d", &ndelay)) &&
4373	    (ndelay >= 0) && (ndelay < (1000 * 1000 * 1000))) {
4374		res = count;
4375		if (sdebug_ndelay != ndelay) {
4376			int j, k;
4377			struct sdebug_queue *sqp;
4378
4379			block_unblock_all_queues(true);
4380			for (j = 0, sqp = sdebug_q_arr; j < submit_queues;
4381			     ++j, ++sqp) {
4382				k = find_first_bit(sqp->in_use_bm,
4383						   sdebug_max_queue);
4384				if (k != sdebug_max_queue) {
4385					res = -EBUSY;   /* queued commands */
4386					break;
4387				}
4388			}
4389			if (res > 0) {
4390				/* make sure sdebug_defer instances get
4391				 * re-allocated for new delay variant */
4392				free_all_queued();
4393				sdebug_ndelay = ndelay;
4394				sdebug_jdelay = ndelay  ? JDELAY_OVERRIDDEN
4395							: DEF_JDELAY;
4396			}
4397			block_unblock_all_queues(false);
4398		}
4399		return res;
4400	}
4401	return -EINVAL;
4402}
4403static DRIVER_ATTR_RW(ndelay);
4404
4405static ssize_t opts_show(struct device_driver *ddp, char *buf)
4406{
4407	return scnprintf(buf, PAGE_SIZE, "0x%x\n", sdebug_opts);
4408}
4409
4410static ssize_t opts_store(struct device_driver *ddp, const char *buf,
4411			  size_t count)
4412{
4413        int opts;
4414	char work[20];
4415
4416        if (1 == sscanf(buf, "%10s", work)) {
4417		if (0 == strncasecmp(work,"0x", 2)) {
4418			if (1 == sscanf(&work[2], "%x", &opts))
4419				goto opts_done;
4420		} else {
4421			if (1 == sscanf(work, "%d", &opts))
4422				goto opts_done;
4423		}
4424	}
4425	return -EINVAL;
4426opts_done:
4427	sdebug_opts = opts;
4428	sdebug_verbose = !!(SDEBUG_OPT_NOISE & opts);
4429	sdebug_any_injecting_opt = !!(SDEBUG_OPT_ALL_INJECTING & opts);
4430	tweak_cmnd_count();
4431	return count;
4432}
4433static DRIVER_ATTR_RW(opts);
4434
4435static ssize_t ptype_show(struct device_driver *ddp, char *buf)
4436{
4437	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_ptype);
4438}
4439static ssize_t ptype_store(struct device_driver *ddp, const char *buf,
4440			   size_t count)
4441{
4442        int n;
4443
4444	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4445		sdebug_ptype = n;
4446		return count;
4447	}
4448	return -EINVAL;
4449}
4450static DRIVER_ATTR_RW(ptype);
4451
4452static ssize_t dsense_show(struct device_driver *ddp, char *buf)
4453{
4454	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_dsense);
4455}
4456static ssize_t dsense_store(struct device_driver *ddp, const char *buf,
4457			    size_t count)
4458{
4459        int n;
4460
4461	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4462		sdebug_dsense = n;
4463		return count;
4464	}
4465	return -EINVAL;
4466}
4467static DRIVER_ATTR_RW(dsense);
4468
4469static ssize_t fake_rw_show(struct device_driver *ddp, char *buf)
4470{
4471	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_fake_rw);
4472}
4473static ssize_t fake_rw_store(struct device_driver *ddp, const char *buf,
4474			     size_t count)
4475{
4476        int n;
4477
4478	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4479		n = (n > 0);
4480		sdebug_fake_rw = (sdebug_fake_rw > 0);
4481		if (sdebug_fake_rw != n) {
4482			if ((0 == n) && (NULL == fake_storep)) {
4483				unsigned long sz =
4484					(unsigned long)sdebug_dev_size_mb *
4485					1048576;
4486
4487				fake_storep = vmalloc(sz);
4488				if (NULL == fake_storep) {
4489					pr_err("out of memory, 9\n");
4490					return -ENOMEM;
4491				}
4492				memset(fake_storep, 0, sz);
4493			}
4494			sdebug_fake_rw = n;
4495		}
4496		return count;
4497	}
4498	return -EINVAL;
4499}
4500static DRIVER_ATTR_RW(fake_rw);
4501
4502static ssize_t no_lun_0_show(struct device_driver *ddp, char *buf)
4503{
4504	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_no_lun_0);
4505}
4506static ssize_t no_lun_0_store(struct device_driver *ddp, const char *buf,
4507			      size_t count)
4508{
4509        int n;
4510
4511	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4512		sdebug_no_lun_0 = n;
4513		return count;
4514	}
4515	return -EINVAL;
4516}
4517static DRIVER_ATTR_RW(no_lun_0);
4518
4519static ssize_t num_tgts_show(struct device_driver *ddp, char *buf)
4520{
4521	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_num_tgts);
4522}
4523static ssize_t num_tgts_store(struct device_driver *ddp, const char *buf,
4524			      size_t count)
4525{
4526        int n;
4527
4528	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4529		sdebug_num_tgts = n;
4530		sdebug_max_tgts_luns();
4531		return count;
4532	}
4533	return -EINVAL;
4534}
4535static DRIVER_ATTR_RW(num_tgts);
4536
4537static ssize_t dev_size_mb_show(struct device_driver *ddp, char *buf)
4538{
4539	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_dev_size_mb);
4540}
4541static DRIVER_ATTR_RO(dev_size_mb);
4542
4543static ssize_t num_parts_show(struct device_driver *ddp, char *buf)
4544{
4545	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_num_parts);
4546}
4547static DRIVER_ATTR_RO(num_parts);
4548
4549static ssize_t every_nth_show(struct device_driver *ddp, char *buf)
4550{
4551	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_every_nth);
4552}
4553static ssize_t every_nth_store(struct device_driver *ddp, const char *buf,
4554			       size_t count)
4555{
4556        int nth;
4557
4558	if ((count > 0) && (1 == sscanf(buf, "%d", &nth))) {
4559		sdebug_every_nth = nth;
4560		if (nth && !sdebug_statistics) {
4561			pr_info("every_nth needs statistics=1, set it\n");
4562			sdebug_statistics = true;
4563		}
4564		tweak_cmnd_count();
4565		return count;
4566	}
4567	return -EINVAL;
4568}
4569static DRIVER_ATTR_RW(every_nth);
4570
4571static ssize_t max_luns_show(struct device_driver *ddp, char *buf)
4572{
4573	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_max_luns);
4574}
4575static ssize_t max_luns_store(struct device_driver *ddp, const char *buf,
4576			      size_t count)
4577{
4578        int n;
4579	bool changed;
4580
4581	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4582		if (n > 256) {
4583			pr_warn("max_luns can be no more than 256\n");
4584			return -EINVAL;
4585		}
4586		changed = (sdebug_max_luns != n);
4587		sdebug_max_luns = n;
4588		sdebug_max_tgts_luns();
4589		if (changed && (sdebug_scsi_level >= 5)) {	/* >= SPC-3 */
4590			struct sdebug_host_info *sdhp;
4591			struct sdebug_dev_info *dp;
4592
4593			spin_lock(&sdebug_host_list_lock);
4594			list_for_each_entry(sdhp, &sdebug_host_list,
4595					    host_list) {
4596				list_for_each_entry(dp, &sdhp->dev_info_list,
4597						    dev_list) {
4598					set_bit(SDEBUG_UA_LUNS_CHANGED,
4599						dp->uas_bm);
4600				}
4601			}
4602			spin_unlock(&sdebug_host_list_lock);
4603		}
4604		return count;
4605	}
4606	return -EINVAL;
4607}
4608static DRIVER_ATTR_RW(max_luns);
4609
4610static ssize_t max_queue_show(struct device_driver *ddp, char *buf)
4611{
4612	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_max_queue);
4613}
4614/* N.B. max_queue can be changed while there are queued commands. In flight
4615 * commands beyond the new max_queue will be completed. */
4616static ssize_t max_queue_store(struct device_driver *ddp, const char *buf,
4617			       size_t count)
4618{
4619	int j, n, k, a;
4620	struct sdebug_queue *sqp;
4621
4622	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n > 0) &&
4623	    (n <= SDEBUG_CANQUEUE)) {
4624		block_unblock_all_queues(true);
4625		k = 0;
4626		for (j = 0, sqp = sdebug_q_arr; j < submit_queues;
4627		     ++j, ++sqp) {
4628			a = find_last_bit(sqp->in_use_bm, SDEBUG_CANQUEUE);
4629			if (a > k)
4630				k = a;
4631		}
4632		sdebug_max_queue = n;
4633		if (k == SDEBUG_CANQUEUE)
4634			atomic_set(&retired_max_queue, 0);
4635		else if (k >= n)
4636			atomic_set(&retired_max_queue, k + 1);
4637		else
4638			atomic_set(&retired_max_queue, 0);
4639		block_unblock_all_queues(false);
4640		return count;
4641	}
4642	return -EINVAL;
4643}
4644static DRIVER_ATTR_RW(max_queue);
4645
4646static ssize_t no_uld_show(struct device_driver *ddp, char *buf)
4647{
4648	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_no_uld);
4649}
4650static DRIVER_ATTR_RO(no_uld);
4651
4652static ssize_t scsi_level_show(struct device_driver *ddp, char *buf)
4653{
4654	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_scsi_level);
4655}
4656static DRIVER_ATTR_RO(scsi_level);
4657
4658static ssize_t virtual_gb_show(struct device_driver *ddp, char *buf)
4659{
4660	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_virtual_gb);
4661}
4662static ssize_t virtual_gb_store(struct device_driver *ddp, const char *buf,
4663				size_t count)
4664{
4665        int n;
4666	bool changed;
4667
4668	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4669		changed = (sdebug_virtual_gb != n);
4670		sdebug_virtual_gb = n;
4671		sdebug_capacity = get_sdebug_capacity();
4672		if (changed) {
4673			struct sdebug_host_info *sdhp;
4674			struct sdebug_dev_info *dp;
4675
4676			spin_lock(&sdebug_host_list_lock);
4677			list_for_each_entry(sdhp, &sdebug_host_list,
4678					    host_list) {
4679				list_for_each_entry(dp, &sdhp->dev_info_list,
4680						    dev_list) {
4681					set_bit(SDEBUG_UA_CAPACITY_CHANGED,
4682						dp->uas_bm);
4683				}
4684			}
4685			spin_unlock(&sdebug_host_list_lock);
4686		}
4687		return count;
4688	}
4689	return -EINVAL;
4690}
4691static DRIVER_ATTR_RW(virtual_gb);
4692
4693static ssize_t add_host_show(struct device_driver *ddp, char *buf)
4694{
4695	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_add_host);
4696}
4697
4698static int sdebug_add_adapter(void);
4699static void sdebug_remove_adapter(void);
4700
4701static ssize_t add_host_store(struct device_driver *ddp, const char *buf,
4702			      size_t count)
4703{
4704	int delta_hosts;
4705
4706	if (sscanf(buf, "%d", &delta_hosts) != 1)
4707		return -EINVAL;
4708	if (delta_hosts > 0) {
4709		do {
4710			sdebug_add_adapter();
4711		} while (--delta_hosts);
4712	} else if (delta_hosts < 0) {
4713		do {
4714			sdebug_remove_adapter();
4715		} while (++delta_hosts);
4716	}
4717	return count;
4718}
4719static DRIVER_ATTR_RW(add_host);
4720
4721static ssize_t vpd_use_hostno_show(struct device_driver *ddp, char *buf)
4722{
4723	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_vpd_use_hostno);
4724}
4725static ssize_t vpd_use_hostno_store(struct device_driver *ddp, const char *buf,
4726				    size_t count)
4727{
4728	int n;
4729
4730	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4731		sdebug_vpd_use_hostno = n;
4732		return count;
4733	}
4734	return -EINVAL;
4735}
4736static DRIVER_ATTR_RW(vpd_use_hostno);
4737
4738static ssize_t statistics_show(struct device_driver *ddp, char *buf)
4739{
4740	return scnprintf(buf, PAGE_SIZE, "%d\n", (int)sdebug_statistics);
4741}
4742static ssize_t statistics_store(struct device_driver *ddp, const char *buf,
4743				size_t count)
4744{
4745	int n;
4746
4747	if ((count > 0) && (sscanf(buf, "%d", &n) == 1) && (n >= 0)) {
4748		if (n > 0)
4749			sdebug_statistics = true;
4750		else {
4751			clear_queue_stats();
4752			sdebug_statistics = false;
4753		}
4754		return count;
4755	}
4756	return -EINVAL;
4757}
4758static DRIVER_ATTR_RW(statistics);
4759
4760static ssize_t sector_size_show(struct device_driver *ddp, char *buf)
4761{
4762	return scnprintf(buf, PAGE_SIZE, "%u\n", sdebug_sector_size);
4763}
4764static DRIVER_ATTR_RO(sector_size);
4765
4766static ssize_t submit_queues_show(struct device_driver *ddp, char *buf)
4767{
4768	return scnprintf(buf, PAGE_SIZE, "%d\n", submit_queues);
4769}
4770static DRIVER_ATTR_RO(submit_queues);
4771
4772static ssize_t dix_show(struct device_driver *ddp, char *buf)
4773{
4774	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_dix);
4775}
4776static DRIVER_ATTR_RO(dix);
4777
4778static ssize_t dif_show(struct device_driver *ddp, char *buf)
4779{
4780	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_dif);
4781}
4782static DRIVER_ATTR_RO(dif);
4783
4784static ssize_t guard_show(struct device_driver *ddp, char *buf)
4785{
4786	return scnprintf(buf, PAGE_SIZE, "%u\n", sdebug_guard);
4787}
4788static DRIVER_ATTR_RO(guard);
4789
4790static ssize_t ato_show(struct device_driver *ddp, char *buf)
4791{
4792	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_ato);
4793}
4794static DRIVER_ATTR_RO(ato);
4795
4796static ssize_t map_show(struct device_driver *ddp, char *buf)
4797{
4798	ssize_t count;
4799
4800	if (!scsi_debug_lbp())
4801		return scnprintf(buf, PAGE_SIZE, "0-%u\n",
4802				 sdebug_store_sectors);
4803
4804	count = scnprintf(buf, PAGE_SIZE - 1, "%*pbl",
4805			  (int)map_size, map_storep);
4806	buf[count++] = '\n';
4807	buf[count] = '\0';
4808
4809	return count;
4810}
4811static DRIVER_ATTR_RO(map);
4812
4813static ssize_t removable_show(struct device_driver *ddp, char *buf)
4814{
4815	return scnprintf(buf, PAGE_SIZE, "%d\n", sdebug_removable ? 1 : 0);
4816}
4817static ssize_t removable_store(struct device_driver *ddp, const char *buf,
4818			       size_t count)
4819{
4820	int n;
4821
4822	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4823		sdebug_removable = (n > 0);
4824		return count;
4825	}
4826	return -EINVAL;
4827}
4828static DRIVER_ATTR_RW(removable);
4829
4830static ssize_t host_lock_show(struct device_driver *ddp, char *buf)
4831{
4832	return scnprintf(buf, PAGE_SIZE, "%d\n", !!sdebug_host_lock);
4833}
4834/* N.B. sdebug_host_lock does nothing, kept for backward compatibility */
4835static ssize_t host_lock_store(struct device_driver *ddp, const char *buf,
4836			       size_t count)
4837{
4838	int n;
4839
4840	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4841		sdebug_host_lock = (n > 0);
4842		return count;
4843	}
4844	return -EINVAL;
4845}
4846static DRIVER_ATTR_RW(host_lock);
4847
4848static ssize_t strict_show(struct device_driver *ddp, char *buf)
4849{
4850	return scnprintf(buf, PAGE_SIZE, "%d\n", !!sdebug_strict);
4851}
4852static ssize_t strict_store(struct device_driver *ddp, const char *buf,
4853			    size_t count)
4854{
4855	int n;
4856
4857	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4858		sdebug_strict = (n > 0);
4859		return count;
4860	}
4861	return -EINVAL;
4862}
4863static DRIVER_ATTR_RW(strict);
4864
4865static ssize_t uuid_ctl_show(struct device_driver *ddp, char *buf)
4866{
4867	return scnprintf(buf, PAGE_SIZE, "%d\n", !!sdebug_uuid_ctl);
4868}
4869static DRIVER_ATTR_RO(uuid_ctl);
4870
4871
4872/* Note: The following array creates attribute files in the
4873   /sys/bus/pseudo/drivers/scsi_debug directory. The advantage of these
4874   files (over those found in the /sys/module/scsi_debug/parameters
4875   directory) is that auxiliary actions can be triggered when an attribute
4876   is changed. For example see: sdebug_add_host_store() above.
4877 */
4878
4879static struct attribute *sdebug_drv_attrs[] = {
4880	&driver_attr_delay.attr,
4881	&driver_attr_opts.attr,
4882	&driver_attr_ptype.attr,
4883	&driver_attr_dsense.attr,
4884	&driver_attr_fake_rw.attr,
4885	&driver_attr_no_lun_0.attr,
4886	&driver_attr_num_tgts.attr,
4887	&driver_attr_dev_size_mb.attr,
4888	&driver_attr_num_parts.attr,
4889	&driver_attr_every_nth.attr,
4890	&driver_attr_max_luns.attr,
4891	&driver_attr_max_queue.attr,
4892	&driver_attr_no_uld.attr,
4893	&driver_attr_scsi_level.attr,
4894	&driver_attr_virtual_gb.attr,
4895	&driver_attr_add_host.attr,
4896	&driver_attr_vpd_use_hostno.attr,
4897	&driver_attr_sector_size.attr,
4898	&driver_attr_statistics.attr,
4899	&driver_attr_submit_queues.attr,
4900	&driver_attr_dix.attr,
4901	&driver_attr_dif.attr,
4902	&driver_attr_guard.attr,
4903	&driver_attr_ato.attr,
4904	&driver_attr_map.attr,
4905	&driver_attr_removable.attr,
4906	&driver_attr_host_lock.attr,
4907	&driver_attr_ndelay.attr,
4908	&driver_attr_strict.attr,
4909	&driver_attr_uuid_ctl.attr,
4910	NULL,
4911};
4912ATTRIBUTE_GROUPS(sdebug_drv);
4913
4914static struct device *pseudo_primary;
4915
4916static int __init scsi_debug_init(void)
4917{
4918	unsigned long sz;
4919	int host_to_add;
4920	int k;
4921	int ret;
4922
4923	atomic_set(&retired_max_queue, 0);
4924
4925	if (sdebug_ndelay >= 1000 * 1000 * 1000) {
4926		pr_warn("ndelay must be less than 1 second, ignored\n");
4927		sdebug_ndelay = 0;
4928	} else if (sdebug_ndelay > 0)
4929		sdebug_jdelay = JDELAY_OVERRIDDEN;
4930
4931	switch (sdebug_sector_size) {
4932	case  512:
4933	case 1024:
4934	case 2048:
4935	case 4096:
4936		break;
4937	default:
4938		pr_err("invalid sector_size %d\n", sdebug_sector_size);
 
4939		return -EINVAL;
4940	}
4941
4942	switch (sdebug_dif) {
4943	case T10_PI_TYPE0_PROTECTION:
4944		break;
4945	case T10_PI_TYPE1_PROTECTION:
4946	case T10_PI_TYPE2_PROTECTION:
4947	case T10_PI_TYPE3_PROTECTION:
4948		have_dif_prot = true;
4949		break;
4950
4951	default:
4952		pr_err("dif must be 0, 1, 2 or 3\n");
4953		return -EINVAL;
4954	}
4955
4956	if (sdebug_guard > 1) {
4957		pr_err("guard must be 0 or 1\n");
4958		return -EINVAL;
4959	}
4960
4961	if (sdebug_ato > 1) {
4962		pr_err("ato must be 0 or 1\n");
4963		return -EINVAL;
4964	}
4965
4966	if (sdebug_physblk_exp > 15) {
4967		pr_err("invalid physblk_exp %u\n", sdebug_physblk_exp);
4968		return -EINVAL;
4969	}
4970	if (sdebug_max_luns > 256) {
4971		pr_warn("max_luns can be no more than 256, use default\n");
4972		sdebug_max_luns = DEF_MAX_LUNS;
4973	}
4974
4975	if (sdebug_lowest_aligned > 0x3fff) {
4976		pr_err("lowest_aligned too big: %u\n", sdebug_lowest_aligned);
 
4977		return -EINVAL;
4978	}
4979
4980	if (submit_queues < 1) {
4981		pr_err("submit_queues must be 1 or more\n");
 
4982		return -EINVAL;
4983	}
4984	sdebug_q_arr = kcalloc(submit_queues, sizeof(struct sdebug_queue),
4985			       GFP_KERNEL);
4986	if (sdebug_q_arr == NULL)
4987		return -ENOMEM;
4988	for (k = 0; k < submit_queues; ++k)
4989		spin_lock_init(&sdebug_q_arr[k].qc_lock);
4990
4991	if (sdebug_dev_size_mb < 1)
4992		sdebug_dev_size_mb = 1;  /* force minimum 1 MB ramdisk */
4993	sz = (unsigned long)sdebug_dev_size_mb * 1048576;
4994	sdebug_store_sectors = sz / sdebug_sector_size;
4995	sdebug_capacity = get_sdebug_capacity();
4996
4997	/* play around with geometry, don't waste too much on track 0 */
4998	sdebug_heads = 8;
4999	sdebug_sectors_per = 32;
5000	if (sdebug_dev_size_mb >= 256)
 
 
5001		sdebug_heads = 64;
5002	else if (sdebug_dev_size_mb >= 16)
5003		sdebug_heads = 32;
5004	sdebug_cylinders_per = (unsigned long)sdebug_capacity /
5005			       (sdebug_sectors_per * sdebug_heads);
5006	if (sdebug_cylinders_per >= 1024) {
5007		/* other LLDs do this; implies >= 1GB ram disk ... */
5008		sdebug_heads = 255;
5009		sdebug_sectors_per = 63;
5010		sdebug_cylinders_per = (unsigned long)sdebug_capacity /
5011			       (sdebug_sectors_per * sdebug_heads);
5012	}
5013
5014	if (sdebug_fake_rw == 0) {
5015		fake_storep = vmalloc(sz);
5016		if (NULL == fake_storep) {
5017			pr_err("out of memory, 1\n");
5018			ret = -ENOMEM;
5019			goto free_q_arr;
5020		}
5021		memset(fake_storep, 0, sz);
5022		if (sdebug_num_parts > 0)
5023			sdebug_build_parts(fake_storep, sz);
5024	}
 
 
 
5025
5026	if (sdebug_dix) {
5027		int dif_size;
5028
5029		dif_size = sdebug_store_sectors * sizeof(struct t10_pi_tuple);
5030		dif_storep = vmalloc(dif_size);
5031
5032		pr_err("dif_storep %u bytes @ %p\n", dif_size, dif_storep);
 
5033
5034		if (dif_storep == NULL) {
5035			pr_err("out of mem. (DIX)\n");
5036			ret = -ENOMEM;
5037			goto free_vm;
5038		}
5039
5040		memset(dif_storep, 0xff, dif_size);
5041	}
5042
5043	/* Logical Block Provisioning */
5044	if (scsi_debug_lbp()) {
5045		sdebug_unmap_max_blocks =
5046			clamp(sdebug_unmap_max_blocks, 0U, 0xffffffffU);
5047
5048		sdebug_unmap_max_desc =
5049			clamp(sdebug_unmap_max_desc, 0U, 256U);
5050
5051		sdebug_unmap_granularity =
5052			clamp(sdebug_unmap_granularity, 1U, 0xffffffffU);
5053
5054		if (sdebug_unmap_alignment &&
5055		    sdebug_unmap_granularity <=
5056		    sdebug_unmap_alignment) {
5057			pr_err("ERR: unmap_granularity <= unmap_alignment\n");
5058			ret = -EINVAL;
5059			goto free_vm;
 
5060		}
5061
5062		map_size = lba_to_map_index(sdebug_store_sectors - 1) + 1;
5063		map_storep = vmalloc(BITS_TO_LONGS(map_size) * sizeof(long));
5064
5065		pr_info("%lu provisioning blocks\n", map_size);
 
5066
5067		if (map_storep == NULL) {
5068			pr_err("out of mem. (MAP)\n");
5069			ret = -ENOMEM;
5070			goto free_vm;
5071		}
5072
5073		bitmap_zero(map_storep, map_size);
5074
5075		/* Map first 1KB for partition table */
5076		if (sdebug_num_parts)
5077			map_region(0, 2);
5078	}
5079
5080	pseudo_primary = root_device_register("pseudo_0");
5081	if (IS_ERR(pseudo_primary)) {
5082		pr_warn("root_device_register() error\n");
5083		ret = PTR_ERR(pseudo_primary);
5084		goto free_vm;
5085	}
5086	ret = bus_register(&pseudo_lld_bus);
5087	if (ret < 0) {
5088		pr_warn("bus_register error: %d\n", ret);
 
5089		goto dev_unreg;
5090	}
5091	ret = driver_register(&sdebug_driverfs_driver);
5092	if (ret < 0) {
5093		pr_warn("driver_register error: %d\n", ret);
 
5094		goto bus_unreg;
5095	}
5096
5097	host_to_add = sdebug_add_host;
5098	sdebug_add_host = 0;
 
 
5099
5100        for (k = 0; k < host_to_add; k++) {
5101                if (sdebug_add_adapter()) {
5102			pr_err("sdebug_add_adapter failed k=%d\n", k);
 
5103                        break;
5104                }
5105        }
5106
5107	if (sdebug_verbose)
5108		pr_info("built %d host(s)\n", sdebug_add_host);
5109
 
5110	return 0;
5111
5112bus_unreg:
5113	bus_unregister(&pseudo_lld_bus);
5114dev_unreg:
5115	root_device_unregister(pseudo_primary);
5116free_vm:
5117	vfree(map_storep);
5118	vfree(dif_storep);
 
 
5119	vfree(fake_storep);
5120free_q_arr:
5121	kfree(sdebug_q_arr);
5122	return ret;
5123}
5124
5125static void __exit scsi_debug_exit(void)
5126{
5127	int k = sdebug_add_host;
5128
5129	stop_all_queued();
5130	free_all_queued();
5131	for (; k; k--)
5132		sdebug_remove_adapter();
5133	driver_unregister(&sdebug_driverfs_driver);
5134	bus_unregister(&pseudo_lld_bus);
5135	root_device_unregister(pseudo_primary);
5136
5137	vfree(map_storep);
5138	vfree(dif_storep);
 
5139	vfree(fake_storep);
5140	kfree(sdebug_q_arr);
5141}
5142
5143device_initcall(scsi_debug_init);
5144module_exit(scsi_debug_exit);
5145
5146static void sdebug_release_adapter(struct device * dev)
5147{
5148        struct sdebug_host_info *sdbg_host;
5149
5150	sdbg_host = to_sdebug_host(dev);
5151        kfree(sdbg_host);
5152}
5153
5154static int sdebug_add_adapter(void)
5155{
5156	int k, devs_per_host;
5157        int error = 0;
5158        struct sdebug_host_info *sdbg_host;
5159	struct sdebug_dev_info *sdbg_devinfo, *tmp;
5160
5161        sdbg_host = kzalloc(sizeof(*sdbg_host),GFP_KERNEL);
5162        if (NULL == sdbg_host) {
5163		pr_err("out of memory at line %d\n", __LINE__);
 
5164                return -ENOMEM;
5165        }
5166
5167        INIT_LIST_HEAD(&sdbg_host->dev_info_list);
5168
5169	devs_per_host = sdebug_num_tgts * sdebug_max_luns;
5170        for (k = 0; k < devs_per_host; k++) {
5171		sdbg_devinfo = sdebug_device_create(sdbg_host, GFP_KERNEL);
5172		if (!sdbg_devinfo) {
5173			pr_err("out of memory at line %d\n", __LINE__);
 
5174                        error = -ENOMEM;
5175			goto clean;
5176                }
5177        }
5178
5179        spin_lock(&sdebug_host_list_lock);
5180        list_add_tail(&sdbg_host->host_list, &sdebug_host_list);
5181        spin_unlock(&sdebug_host_list_lock);
5182
5183        sdbg_host->dev.bus = &pseudo_lld_bus;
5184        sdbg_host->dev.parent = pseudo_primary;
5185        sdbg_host->dev.release = &sdebug_release_adapter;
5186	dev_set_name(&sdbg_host->dev, "adapter%d", sdebug_add_host);
5187
5188        error = device_register(&sdbg_host->dev);
5189
5190        if (error)
5191		goto clean;
5192
5193	++sdebug_add_host;
5194        return error;
5195
5196clean:
5197	list_for_each_entry_safe(sdbg_devinfo, tmp, &sdbg_host->dev_info_list,
5198				 dev_list) {
5199		list_del(&sdbg_devinfo->dev_list);
5200		kfree(sdbg_devinfo);
5201	}
5202
5203	kfree(sdbg_host);
5204        return error;
5205}
5206
5207static void sdebug_remove_adapter(void)
5208{
5209        struct sdebug_host_info * sdbg_host = NULL;
5210
5211        spin_lock(&sdebug_host_list_lock);
5212        if (!list_empty(&sdebug_host_list)) {
5213                sdbg_host = list_entry(sdebug_host_list.prev,
5214                                       struct sdebug_host_info, host_list);
5215		list_del(&sdbg_host->host_list);
5216	}
5217        spin_unlock(&sdebug_host_list_lock);
5218
5219	if (!sdbg_host)
5220		return;
5221
5222	device_unregister(&sdbg_host->dev);
5223	--sdebug_add_host;
5224}
5225
5226static int sdebug_change_qdepth(struct scsi_device *sdev, int qdepth)
 
5227{
5228	int num_in_q = 0;
5229	struct sdebug_dev_info *devip;
5230
5231	block_unblock_all_queues(true);
5232	devip = (struct sdebug_dev_info *)sdev->hostdata;
5233	if (NULL == devip) {
5234		block_unblock_all_queues(false);
5235		return	-ENODEV;
5236	}
5237	num_in_q = atomic_read(&devip->num_in_q);
5238
5239	if (qdepth < 1)
5240		qdepth = 1;
5241	/* allow to exceed max host qc_arr elements for testing */
5242	if (qdepth > SDEBUG_CANQUEUE + 10)
5243		qdepth = SDEBUG_CANQUEUE + 10;
5244	scsi_change_queue_depth(sdev, qdepth);
5245
5246	if (SDEBUG_OPT_Q_NOISE & sdebug_opts) {
5247		sdev_printk(KERN_INFO, sdev, "%s: qdepth=%d, num_in_q=%d\n",
5248			    __func__, qdepth, num_in_q);
5249	}
5250	block_unblock_all_queues(false);
5251	return sdev->queue_depth;
5252}
5253
5254static bool fake_timeout(struct scsi_cmnd *scp)
5255{
5256	if (0 == (atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth))) {
5257		if (sdebug_every_nth < -1)
5258			sdebug_every_nth = -1;
5259		if (SDEBUG_OPT_TIMEOUT & sdebug_opts)
5260			return true; /* ignore command causing timeout */
5261		else if (SDEBUG_OPT_MAC_TIMEOUT & sdebug_opts &&
5262			 scsi_medium_access_command(scp))
5263			return true; /* time out reads and writes */
5264	}
5265	return false;
5266}
5267
5268static int scsi_debug_queuecommand(struct Scsi_Host *shost,
5269				   struct scsi_cmnd *scp)
5270{
5271	u8 sdeb_i;
5272	struct scsi_device *sdp = scp->device;
5273	const struct opcode_info_t *oip;
5274	const struct opcode_info_t *r_oip;
5275	struct sdebug_dev_info *devip;
5276	u8 *cmd = scp->cmnd;
5277	int (*r_pfp)(struct scsi_cmnd *, struct sdebug_dev_info *);
5278	int k, na;
5279	int errsts = 0;
5280	u32 flags;
5281	u16 sa;
5282	u8 opcode = cmd[0];
5283	bool has_wlun_rl;
5284
5285	scsi_set_resid(scp, 0);
5286	if (sdebug_statistics)
5287		atomic_inc(&sdebug_cmnd_count);
5288	if (unlikely(sdebug_verbose &&
5289		     !(SDEBUG_OPT_NO_CDB_NOISE & sdebug_opts))) {
5290		char b[120];
5291		int n, len, sb;
5292
5293		len = scp->cmd_len;
5294		sb = (int)sizeof(b);
5295		if (len > 32)
5296			strcpy(b, "too long, over 32 bytes");
5297		else {
5298			for (k = 0, n = 0; k < len && n < sb; ++k)
5299				n += scnprintf(b + n, sb - n, "%02x ",
5300					       (u32)cmd[k]);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5301		}
5302		if (sdebug_mq_active)
5303			sdev_printk(KERN_INFO, sdp, "%s: tag=%u, cmd %s\n",
5304				    my_name, blk_mq_unique_tag(scp->request),
5305				    b);
5306		else
5307			sdev_printk(KERN_INFO, sdp, "%s: cmd %s\n", my_name,
5308				    b);
5309	}
5310	has_wlun_rl = (sdp->lun == SCSI_W_LUN_REPORT_LUNS);
5311	if (unlikely((sdp->lun >= sdebug_max_luns) && !has_wlun_rl))
5312		goto err_out;
5313
5314	sdeb_i = opcode_ind_arr[opcode];	/* fully mapped */
5315	oip = &opcode_info_arr[sdeb_i];		/* safe if table consistent */
5316	devip = (struct sdebug_dev_info *)sdp->hostdata;
5317	if (unlikely(!devip)) {
5318		devip = find_build_dev_info(sdp);
5319		if (NULL == devip)
5320			goto err_out;
5321	}
5322	na = oip->num_attached;
5323	r_pfp = oip->pfp;
5324	if (na) {	/* multiple commands with this opcode */
5325		r_oip = oip;
5326		if (FF_SA & r_oip->flags) {
5327			if (F_SA_LOW & oip->flags)
5328				sa = 0x1f & cmd[1];
5329			else
5330				sa = get_unaligned_be16(cmd + 8);
5331			for (k = 0; k <= na; oip = r_oip->arrp + k++) {
5332				if (opcode == oip->opcode && sa == oip->sa)
5333					break;
5334			}
5335		} else {   /* since no service action only check opcode */
5336			for (k = 0; k <= na; oip = r_oip->arrp + k++) {
5337				if (opcode == oip->opcode)
5338					break;
5339			}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5340		}
5341		if (k > na) {
5342			if (F_SA_LOW & r_oip->flags)
5343				mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 4);
5344			else if (F_SA_HIGH & r_oip->flags)
5345				mk_sense_invalid_fld(scp, SDEB_IN_CDB, 8, 7);
5346			else
5347				mk_sense_invalid_opcode(scp);
5348			goto check_cond;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5349		}
5350	}	/* else (when na==0) we assume the oip is a match */
5351	flags = oip->flags;
5352	if (unlikely(F_INV_OP & flags)) {
5353		mk_sense_invalid_opcode(scp);
5354		goto check_cond;
5355	}
5356	if (unlikely(has_wlun_rl && !(F_RL_WLUN_OK & flags))) {
5357		if (sdebug_verbose)
5358			sdev_printk(KERN_INFO, sdp, "%s: Opcode 0x%x not%s\n",
5359				    my_name, opcode, " supported for wlun");
5360		mk_sense_invalid_opcode(scp);
5361		goto check_cond;
5362	}
5363	if (unlikely(sdebug_strict)) {	/* check cdb against mask */
5364		u8 rem;
5365		int j;
5366
5367		for (k = 1; k < oip->len_mask[0] && k < 16; ++k) {
5368			rem = ~oip->len_mask[k] & cmd[k];
5369			if (rem) {
5370				for (j = 7; j >= 0; --j, rem <<= 1) {
5371					if (0x80 & rem)
5372						break;
5373				}
5374				mk_sense_invalid_fld(scp, SDEB_IN_CDB, k, j);
5375				goto check_cond;
 
 
 
 
5376			}
5377		}
5378	}
5379	if (unlikely(!(F_SKIP_UA & flags) &&
5380		     find_first_bit(devip->uas_bm,
5381				    SDEBUG_NUM_UAS) != SDEBUG_NUM_UAS)) {
5382		errsts = make_ua(scp, devip);
 
 
 
 
 
 
5383		if (errsts)
5384			goto check_cond;
5385	}
5386	if (unlikely((F_M_ACCESS & flags) && atomic_read(&devip->stopped))) {
5387		mk_sense_buffer(scp, NOT_READY, LOGICAL_UNIT_NOT_READY, 0x2);
5388		if (sdebug_verbose)
5389			sdev_printk(KERN_INFO, sdp, "%s reports: Not ready: "
5390				    "%s\n", my_name, "initializing command "
5391				    "required");
5392		errsts = check_condition_result;
5393		goto fini;
5394	}
5395	if (sdebug_fake_rw && (F_FAKE_RW & flags))
5396		goto fini;
5397	if (unlikely(sdebug_every_nth)) {
5398		if (fake_timeout(scp))
5399			return 0;	/* ignore command: make trouble */
5400	}
5401	if (likely(oip->pfp))
5402		errsts = oip->pfp(scp, devip);	/* calls a resp_* function */
5403	else if (r_pfp)	/* if leaf function ptr NULL, try the root's */
5404		errsts = r_pfp(scp, devip);
5405
5406fini:
5407	return schedule_resp(scp, devip, errsts,
5408			     ((F_DELAY_OVERR & flags) ? 0 : sdebug_jdelay));
5409check_cond:
5410	return schedule_resp(scp, devip, check_condition_result, 0);
5411err_out:
5412	return schedule_resp(scp, NULL, DID_NO_CONNECT << 16, 0);
5413}
5414
 
 
5415static struct scsi_host_template sdebug_driver_template = {
5416	.show_info =		scsi_debug_show_info,
5417	.write_info =		scsi_debug_write_info,
5418	.proc_name =		sdebug_proc_name,
5419	.name =			"SCSI DEBUG",
5420	.info =			scsi_debug_info,
5421	.slave_alloc =		scsi_debug_slave_alloc,
5422	.slave_configure =	scsi_debug_slave_configure,
5423	.slave_destroy =	scsi_debug_slave_destroy,
5424	.ioctl =		scsi_debug_ioctl,
5425	.queuecommand =		scsi_debug_queuecommand,
5426	.change_queue_depth =	sdebug_change_qdepth,
5427	.eh_abort_handler =	scsi_debug_abort,
 
5428	.eh_device_reset_handler = scsi_debug_device_reset,
5429	.eh_target_reset_handler = scsi_debug_target_reset,
5430	.eh_bus_reset_handler = scsi_debug_bus_reset,
5431	.eh_host_reset_handler = scsi_debug_host_reset,
5432	.can_queue =		SDEBUG_CANQUEUE,
 
5433	.this_id =		7,
5434	.sg_tablesize =		SG_MAX_SEGMENTS,
5435	.cmd_per_lun =		DEF_CMD_PER_LUN,
5436	.max_sectors =		-1U,
5437	.use_clustering = 	DISABLE_CLUSTERING,
5438	.module =		THIS_MODULE,
5439	.track_queue_depth =	1,
5440};
5441
5442static int sdebug_driver_probe(struct device * dev)
5443{
5444	int error = 0;
5445	struct sdebug_host_info *sdbg_host;
5446	struct Scsi_Host *hpnt;
5447	int hprot;
5448
5449	sdbg_host = to_sdebug_host(dev);
5450
5451	sdebug_driver_template.can_queue = sdebug_max_queue;
5452	if (sdebug_clustering)
5453		sdebug_driver_template.use_clustering = ENABLE_CLUSTERING;
5454	hpnt = scsi_host_alloc(&sdebug_driver_template, sizeof(sdbg_host));
5455	if (NULL == hpnt) {
5456		pr_err("scsi_host_alloc failed\n");
5457		error = -ENODEV;
5458		return error;
5459	}
5460	if (submit_queues > nr_cpu_ids) {
5461		pr_warn("%s: trim submit_queues (was %d) to nr_cpu_ids=%d\n",
5462			my_name, submit_queues, nr_cpu_ids);
5463		submit_queues = nr_cpu_ids;
5464	}
5465	/* Decide whether to tell scsi subsystem that we want mq */
5466	/* Following should give the same answer for each host */
5467	sdebug_mq_active = shost_use_blk_mq(hpnt) && (submit_queues > 1);
5468	if (sdebug_mq_active)
5469		hpnt->nr_hw_queues = submit_queues;
5470
5471        sdbg_host->shost = hpnt;
5472	*((struct sdebug_host_info **)hpnt->hostdata) = sdbg_host;
5473	if ((hpnt->this_id >= 0) && (sdebug_num_tgts > hpnt->this_id))
5474		hpnt->max_id = sdebug_num_tgts + 1;
5475	else
5476		hpnt->max_id = sdebug_num_tgts;
5477	/* = sdebug_max_luns; */
5478	hpnt->max_lun = SCSI_W_LUN_REPORT_LUNS + 1;
5479
5480	hprot = 0;
5481
5482	switch (sdebug_dif) {
5483
5484	case T10_PI_TYPE1_PROTECTION:
5485		hprot = SHOST_DIF_TYPE1_PROTECTION;
5486		if (sdebug_dix)
5487			hprot |= SHOST_DIX_TYPE1_PROTECTION;
5488		break;
5489
5490	case T10_PI_TYPE2_PROTECTION:
5491		hprot = SHOST_DIF_TYPE2_PROTECTION;
5492		if (sdebug_dix)
5493			hprot |= SHOST_DIX_TYPE2_PROTECTION;
5494		break;
5495
5496	case T10_PI_TYPE3_PROTECTION:
5497		hprot = SHOST_DIF_TYPE3_PROTECTION;
5498		if (sdebug_dix)
5499			hprot |= SHOST_DIX_TYPE3_PROTECTION;
5500		break;
5501
5502	default:
5503		if (sdebug_dix)
5504			hprot |= SHOST_DIX_TYPE0_PROTECTION;
5505		break;
5506	}
5507
5508	scsi_host_set_prot(hpnt, hprot);
5509
5510	if (have_dif_prot || sdebug_dix)
5511		pr_info("host protection%s%s%s%s%s%s%s\n",
5512			(hprot & SHOST_DIF_TYPE1_PROTECTION) ? " DIF1" : "",
5513			(hprot & SHOST_DIF_TYPE2_PROTECTION) ? " DIF2" : "",
5514			(hprot & SHOST_DIF_TYPE3_PROTECTION) ? " DIF3" : "",
5515			(hprot & SHOST_DIX_TYPE0_PROTECTION) ? " DIX0" : "",
5516			(hprot & SHOST_DIX_TYPE1_PROTECTION) ? " DIX1" : "",
5517			(hprot & SHOST_DIX_TYPE2_PROTECTION) ? " DIX2" : "",
5518			(hprot & SHOST_DIX_TYPE3_PROTECTION) ? " DIX3" : "");
5519
5520	if (sdebug_guard == 1)
5521		scsi_host_set_guard(hpnt, SHOST_DIX_GUARD_IP);
5522	else
5523		scsi_host_set_guard(hpnt, SHOST_DIX_GUARD_CRC);
5524
5525	sdebug_verbose = !!(SDEBUG_OPT_NOISE & sdebug_opts);
5526	sdebug_any_injecting_opt = !!(SDEBUG_OPT_ALL_INJECTING & sdebug_opts);
5527	if (sdebug_every_nth)	/* need stats counters for every_nth */
5528		sdebug_statistics = true;
5529        error = scsi_add_host(hpnt, &sdbg_host->dev);
5530        if (error) {
5531		pr_err("scsi_add_host failed\n");
5532                error = -ENODEV;
5533		scsi_host_put(hpnt);
5534        } else
5535		scsi_scan_host(hpnt);
5536
5537	return error;
 
5538}
5539
5540static int sdebug_driver_remove(struct device * dev)
5541{
5542        struct sdebug_host_info *sdbg_host;
5543	struct sdebug_dev_info *sdbg_devinfo, *tmp;
5544
5545	sdbg_host = to_sdebug_host(dev);
5546
5547	if (!sdbg_host) {
5548		pr_err("Unable to locate host info\n");
 
5549		return -ENODEV;
5550	}
5551
5552        scsi_remove_host(sdbg_host->shost);
5553
5554	list_for_each_entry_safe(sdbg_devinfo, tmp, &sdbg_host->dev_info_list,
5555				 dev_list) {
5556                list_del(&sdbg_devinfo->dev_list);
5557                kfree(sdbg_devinfo);
5558        }
5559
5560        scsi_host_put(sdbg_host->shost);
5561        return 0;
5562}
5563
5564static int pseudo_lld_bus_match(struct device *dev,
5565				struct device_driver *dev_driver)
5566{
5567	return 1;
5568}
5569
5570static struct bus_type pseudo_lld_bus = {
5571	.name = "pseudo",
5572	.match = pseudo_lld_bus_match,
5573	.probe = sdebug_driver_probe,
5574	.remove = sdebug_driver_remove,
5575	.drv_groups = sdebug_drv_groups,
5576};