Linux Audio

Check our new training course

Loading...
v6.2
  1/* SPDX-License-Identifier: MIT */
  2/******************************************************************************
  3 * xen_netif.h
  4 *
  5 * Unified network-device I/O interface for Xen guest OSes.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  6 *
  7 * Copyright (c) 2003-2004, Keir Fraser
  8 */
  9
 10#ifndef __XEN_PUBLIC_IO_XEN_NETIF_H__
 11#define __XEN_PUBLIC_IO_XEN_NETIF_H__
 12
 13#include "ring.h"
 14#include "../grant_table.h"
 15
 16/*
 17 * Older implementation of Xen network frontend / backend has an
 18 * implicit dependency on the MAX_SKB_FRAGS as the maximum number of
 19 * ring slots a skb can use. Netfront / netback may not work as
 20 * expected when frontend and backend have different MAX_SKB_FRAGS.
 21 *
 22 * A better approach is to add mechanism for netfront / netback to
 23 * negotiate this value. However we cannot fix all possible
 24 * frontends, so we need to define a value which states the minimum
 25 * slots backend must support.
 26 *
 27 * The minimum value derives from older Linux kernel's MAX_SKB_FRAGS
 28 * (18), which is proved to work with most frontends. Any new backend
 29 * which doesn't negotiate with frontend should expect frontend to
 30 * send a valid packet using slots up to this value.
 31 */
 32#define XEN_NETIF_NR_SLOTS_MIN 18
 33
 34/*
 35 * Notifications after enqueuing any type of message should be conditional on
 36 * the appropriate req_event or rsp_event field in the shared ring.
 37 * If the client sends notification for rx requests then it should specify
 38 * feature 'feature-rx-notify' via xenbus. Otherwise the backend will assume
 39 * that it cannot safely queue packets (as it may not be kicked to send them).
 40 */
 41
 42/*
 43 * "feature-split-event-channels" is introduced to separate guest TX
 44 * and RX notification. Backend either doesn't support this feature or
 45 * advertises it via xenstore as 0 (disabled) or 1 (enabled).
 46 *
 47 * To make use of this feature, frontend should allocate two event
 48 * channels for TX and RX, advertise them to backend as
 49 * "event-channel-tx" and "event-channel-rx" respectively. If frontend
 50 * doesn't want to use this feature, it just writes "event-channel"
 51 * node as before.
 52 */
 53
 54/*
 55 * Multiple transmit and receive queues:
 56 * If supported, the backend will write the key "multi-queue-max-queues" to
 57 * the directory for that vif, and set its value to the maximum supported
 58 * number of queues.
 59 * Frontends that are aware of this feature and wish to use it can write the
 60 * key "multi-queue-num-queues", set to the number they wish to use, which
 61 * must be greater than zero, and no more than the value reported by the backend
 62 * in "multi-queue-max-queues".
 63 *
 64 * Queues replicate the shared rings and event channels.
 65 * "feature-split-event-channels" may optionally be used when using
 66 * multiple queues, but is not mandatory.
 67 *
 68 * Each queue consists of one shared ring pair, i.e. there must be the same
 69 * number of tx and rx rings.
 70 *
 71 * For frontends requesting just one queue, the usual event-channel and
 72 * ring-ref keys are written as before, simplifying the backend processing
 73 * to avoid distinguishing between a frontend that doesn't understand the
 74 * multi-queue feature, and one that does, but requested only one queue.
 75 *
 76 * Frontends requesting two or more queues must not write the toplevel
 77 * event-channel (or event-channel-{tx,rx}) and {tx,rx}-ring-ref keys,
 78 * instead writing those keys under sub-keys having the name "queue-N" where
 79 * N is the integer ID of the queue for which those keys belong. Queues
 80 * are indexed from zero. For example, a frontend with two queues and split
 81 * event channels must write the following set of queue-related keys:
 82 *
 83 * /local/domain/1/device/vif/0/multi-queue-num-queues = "2"
 84 * /local/domain/1/device/vif/0/queue-0 = ""
 85 * /local/domain/1/device/vif/0/queue-0/tx-ring-ref = "<ring-ref-tx0>"
 86 * /local/domain/1/device/vif/0/queue-0/rx-ring-ref = "<ring-ref-rx0>"
 87 * /local/domain/1/device/vif/0/queue-0/event-channel-tx = "<evtchn-tx0>"
 88 * /local/domain/1/device/vif/0/queue-0/event-channel-rx = "<evtchn-rx0>"
 89 * /local/domain/1/device/vif/0/queue-1 = ""
 90 * /local/domain/1/device/vif/0/queue-1/tx-ring-ref = "<ring-ref-tx1>"
 91 * /local/domain/1/device/vif/0/queue-1/rx-ring-ref = "<ring-ref-rx1"
 92 * /local/domain/1/device/vif/0/queue-1/event-channel-tx = "<evtchn-tx1>"
 93 * /local/domain/1/device/vif/0/queue-1/event-channel-rx = "<evtchn-rx1>"
 94 *
 95 * If there is any inconsistency in the XenStore data, the backend may
 96 * choose not to connect any queues, instead treating the request as an
 97 * error. This includes scenarios where more (or fewer) queues were
 98 * requested than the frontend provided details for.
 99 *
100 * Mapping of packets to queues is considered to be a function of the
101 * transmitting system (backend or frontend) and is not negotiated
102 * between the two. Guests are free to transmit packets on any queue
103 * they choose, provided it has been set up correctly. Guests must be
104 * prepared to receive packets on any queue they have requested be set up.
105 */
106
107/*
108 * "feature-no-csum-offload" should be used to turn IPv4 TCP/UDP checksum
109 * offload off or on. If it is missing then the feature is assumed to be on.
110 * "feature-ipv6-csum-offload" should be used to turn IPv6 TCP/UDP checksum
111 * offload on or off. If it is missing then the feature is assumed to be off.
112 */
113
114/*
115 * "feature-gso-tcpv4" and "feature-gso-tcpv6" advertise the capability to
116 * handle large TCP packets (in IPv4 or IPv6 form respectively). Neither
117 * frontends nor backends are assumed to be capable unless the flags are
118 * present.
119 */
120
121/*
122 * "feature-multicast-control" and "feature-dynamic-multicast-control"
123 * advertise the capability to filter ethernet multicast packets in the
124 * backend. If the frontend wishes to take advantage of this feature then
125 * it may set "request-multicast-control". If the backend only advertises
126 * "feature-multicast-control" then "request-multicast-control" must be set
127 * before the frontend moves into the connected state. The backend will
128 * sample the value on this state transition and any subsequent change in
129 * value will have no effect. However, if the backend also advertises
130 * "feature-dynamic-multicast-control" then "request-multicast-control"
131 * may be set by the frontend at any time. In this case, the backend will
132 * watch the value and re-sample on watch events.
133 *
134 * If the sampled value of "request-multicast-control" is set then the
135 * backend transmit side should no longer flood multicast packets to the
136 * frontend, it should instead drop any multicast packet that does not
137 * match in a filter list.
138 * The list is amended by the frontend by sending dummy transmit requests
139 * containing XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL} extra-info fragments as
140 * specified below.
141 * Note that the filter list may be amended even if the sampled value of
142 * "request-multicast-control" is not set, however the filter should only
143 * be applied if it is set.
144 */
145
146/*
147 * "xdp-headroom" is used to request that extra space is added
148 * for XDP processing.  The value is measured in bytes and passed by
149 * the frontend to be consistent between both ends.
150 * If the value is greater than zero that means that
151 * an RX response is going to be passed to an XDP program for processing.
152 * XEN_NETIF_MAX_XDP_HEADROOM defines the maximum headroom offset in bytes
153 *
154 * "feature-xdp-headroom" is set to "1" by the netback side like other features
155 * so a guest can check if an XDP program can be processed.
156 */
157#define XEN_NETIF_MAX_XDP_HEADROOM 0x7FFF
158
159/*
160 * Control ring
161 * ============
162 *
163 * Some features, such as hashing (detailed below), require a
164 * significant amount of out-of-band data to be passed from frontend to
165 * backend. Use of xenstore is not suitable for large quantities of data
166 * because of quota limitations and so a dedicated 'control ring' is used.
167 * The ability of the backend to use a control ring is advertised by
168 * setting:
169 *
170 * /local/domain/X/backend/<domid>/<vif>/feature-ctrl-ring = "1"
171 *
172 * The frontend provides a control ring to the backend by setting:
173 *
174 * /local/domain/<domid>/device/vif/<vif>/ctrl-ring-ref = <gref>
175 * /local/domain/<domid>/device/vif/<vif>/event-channel-ctrl = <port>
176 *
177 * where <gref> is the grant reference of the shared page used to
178 * implement the control ring and <port> is an event channel to be used
179 * as a mailbox interrupt. These keys must be set before the frontend
180 * moves into the connected state.
181 *
182 * The control ring uses a fixed request/response message size and is
183 * balanced (i.e. one request to one response), so operationally it is much
184 * the same as a transmit or receive ring.
185 * Note that there is no requirement that responses are issued in the same
186 * order as requests.
187 */
188
189/*
190 * Hash types
191 * ==========
192 *
193 * For the purposes of the definitions below, 'Packet[]' is an array of
194 * octets containing an IP packet without options, 'Array[X..Y]' means a
195 * sub-array of 'Array' containing bytes X thru Y inclusive, and '+' is
196 * used to indicate concatenation of arrays.
197 */
198
199/*
200 * A hash calculated over an IP version 4 header as follows:
201 *
202 * Buffer[0..8] = Packet[12..15] (source address) +
203 *                Packet[16..19] (destination address)
204 *
205 * Result = Hash(Buffer, 8)
206 */
207#define _XEN_NETIF_CTRL_HASH_TYPE_IPV4 0
208#define XEN_NETIF_CTRL_HASH_TYPE_IPV4 \
209	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4)
210
211/*
212 * A hash calculated over an IP version 4 header and TCP header as
213 * follows:
214 *
215 * Buffer[0..12] = Packet[12..15] (source address) +
216 *                 Packet[16..19] (destination address) +
217 *                 Packet[20..21] (source port) +
218 *                 Packet[22..23] (destination port)
219 *
220 * Result = Hash(Buffer, 12)
221 */
222#define _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP 1
223#define XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP \
224	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP)
225
226/*
227 * A hash calculated over an IP version 6 header as follows:
228 *
229 * Buffer[0..32] = Packet[8..23]  (source address ) +
230 *                 Packet[24..39] (destination address)
231 *
232 * Result = Hash(Buffer, 32)
233 */
234#define _XEN_NETIF_CTRL_HASH_TYPE_IPV6 2
235#define XEN_NETIF_CTRL_HASH_TYPE_IPV6 \
236	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6)
237
238/*
239 * A hash calculated over an IP version 6 header and TCP header as
240 * follows:
241 *
242 * Buffer[0..36] = Packet[8..23]  (source address) +
243 *                 Packet[24..39] (destination address) +
244 *                 Packet[40..41] (source port) +
245 *                 Packet[42..43] (destination port)
246 *
247 * Result = Hash(Buffer, 36)
248 */
249#define _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP 3
250#define XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP \
251	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)
252
253/*
254 * Hash algorithms
255 * ===============
256 */
257
258#define XEN_NETIF_CTRL_HASH_ALGORITHM_NONE 0
259
260/*
261 * Toeplitz hash:
262 */
263
264#define XEN_NETIF_CTRL_HASH_ALGORITHM_TOEPLITZ 1
265
266/*
267 * This algorithm uses a 'key' as well as the data buffer itself.
268 * (Buffer[] and Key[] are treated as shift-registers where the MSB of
269 * Buffer/Key[0] is considered 'left-most' and the LSB of Buffer/Key[N-1]
270 * is the 'right-most').
271 *
272 * Value = 0
273 * For number of bits in Buffer[]
274 *    If (left-most bit of Buffer[] is 1)
275 *        Value ^= left-most 32 bits of Key[]
276 *    Key[] << 1
277 *    Buffer[] << 1
278 *
279 * The code below is provided for convenience where an operating system
280 * does not already provide an implementation.
281 */
282#ifdef XEN_NETIF_DEFINE_TOEPLITZ
283static uint32_t xen_netif_toeplitz_hash(const uint8_t *key,
284					unsigned int keylen,
285					const uint8_t *buf, unsigned int buflen)
286{
287	unsigned int keyi, bufi;
288	uint64_t prefix = 0;
289	uint64_t hash = 0;
290
291	/* Pre-load prefix with the first 8 bytes of the key */
292	for (keyi = 0; keyi < 8; keyi++) {
293		prefix <<= 8;
294		prefix |= (keyi < keylen) ? key[keyi] : 0;
295	}
296
297	for (bufi = 0; bufi < buflen; bufi++) {
298		uint8_t byte = buf[bufi];
299		unsigned int bit;
300
301		for (bit = 0; bit < 8; bit++) {
302			if (byte & 0x80)
303				hash ^= prefix;
304			prefix <<= 1;
305			byte <<= 1;
306		}
307
308		/*
309		 * 'prefix' has now been left-shifted by 8, so
310		 * OR in the next byte.
311		 */
312		prefix |= (keyi < keylen) ? key[keyi] : 0;
313		keyi++;
314	}
315
316	/* The valid part of the hash is in the upper 32 bits. */
317	return hash >> 32;
318}
319#endif				/* XEN_NETIF_DEFINE_TOEPLITZ */
320
321/*
322 * Control requests (struct xen_netif_ctrl_request)
323 * ================================================
324 *
325 * All requests have the following format:
326 *
327 *    0     1     2     3     4     5     6     7  octet
328 * +-----+-----+-----+-----+-----+-----+-----+-----+
329 * |    id     |   type    |         data[0]       |
330 * +-----+-----+-----+-----+-----+-----+-----+-----+
331 * |         data[1]       |         data[2]       |
332 * +-----+-----+-----+-----+-----------------------+
333 *
334 * id: the request identifier, echoed in response.
335 * type: the type of request (see below)
336 * data[]: any data associated with the request (determined by type)
337 */
338
339struct xen_netif_ctrl_request {
340	uint16_t id;
341	uint16_t type;
342
343#define XEN_NETIF_CTRL_TYPE_INVALID               0
344#define XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS        1
345#define XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS        2
346#define XEN_NETIF_CTRL_TYPE_SET_HASH_KEY          3
347#define XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE 4
348#define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE 5
349#define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING      6
350#define XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM    7
351
352	uint32_t data[3];
353};
354
355/*
356 * Control responses (struct xen_netif_ctrl_response)
357 * ==================================================
358 *
359 * All responses have the following format:
360 *
361 *    0     1     2     3     4     5     6     7  octet
362 * +-----+-----+-----+-----+-----+-----+-----+-----+
363 * |    id     |   type    |         status        |
364 * +-----+-----+-----+-----+-----+-----+-----+-----+
365 * |         data          |
366 * +-----+-----+-----+-----+
367 *
368 * id: the corresponding request identifier
369 * type: the type of the corresponding request
370 * status: the status of request processing
371 * data: any data associated with the response (determined by type and
372 *       status)
373 */
374
375struct xen_netif_ctrl_response {
376	uint16_t id;
377	uint16_t type;
378	uint32_t status;
379
380#define XEN_NETIF_CTRL_STATUS_SUCCESS           0
381#define XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     1
382#define XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER 2
383#define XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW   3
384
385	uint32_t data;
386};
387
388/*
389 * Control messages
390 * ================
391 *
392 * XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
393 * --------------------------------------
394 *
395 * This is sent by the frontend to set the desired hash algorithm.
396 *
397 * Request:
398 *
399 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
400 *  data[0] = a XEN_NETIF_CTRL_HASH_ALGORITHM_* value
401 *  data[1] = 0
402 *  data[2] = 0
403 *
404 * Response:
405 *
406 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
407 *                                                     supported
408 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - The algorithm is not
409 *                                                     supported
410 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
411 *
412 * NOTE: Setting data[0] to XEN_NETIF_CTRL_HASH_ALGORITHM_NONE disables
413 *       hashing and the backend is free to choose how it steers packets
414 *       to queues (which is the default behaviour).
415 *
416 * XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
417 * ----------------------------------
418 *
419 * This is sent by the frontend to query the types of hash supported by
420 * the backend.
421 *
422 * Request:
423 *
424 *  type    = XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
425 *  data[0] = 0
426 *  data[1] = 0
427 *  data[2] = 0
428 *
429 * Response:
430 *
431 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
432 *           XEN_NETIF_CTRL_STATUS_SUCCESS       - Operation successful
433 *  data   = supported hash types (if operation was successful)
434 *
435 * NOTE: A valid hash algorithm must be selected before this operation can
436 *       succeed.
437 *
438 * XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
439 * ----------------------------------
440 *
441 * This is sent by the frontend to set the types of hash that the backend
442 * should calculate. (See above for hash type definitions).
443 * Note that the 'maximal' type of hash should always be chosen. For
444 * example, if the frontend sets both IPV4 and IPV4_TCP hash types then
445 * the latter hash type should be calculated for any TCP packet and the
446 * former only calculated for non-TCP packets.
447 *
448 * Request:
449 *
450 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
451 *  data[0] = bitwise OR of XEN_NETIF_CTRL_HASH_TYPE_* values
452 *  data[1] = 0
453 *  data[2] = 0
454 *
455 * Response:
456 *
457 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
458 *                                                     supported
459 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - One or more flag
460 *                                                     value is invalid or
461 *                                                     unsupported
462 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
463 *  data   = 0
464 *
465 * NOTE: A valid hash algorithm must be selected before this operation can
466 *       succeed.
467 *       Also, setting data[0] to zero disables hashing and the backend
468 *       is free to choose how it steers packets to queues.
469 *
470 * XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
471 * --------------------------------
472 *
473 * This is sent by the frontend to set the key of the hash if the algorithm
474 * requires it. (See hash algorithms above).
475 *
476 * Request:
477 *
478 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
479 *  data[0] = grant reference of page containing the key (assumed to
480 *            start at beginning of grant)
481 *  data[1] = size of key in octets
482 *  data[2] = 0
483 *
484 * Response:
485 *
486 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
487 *                                                     supported
488 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Key size is invalid
489 *           XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW   - Key size is larger
490 *                                                     than the backend
491 *                                                     supports
492 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
493 *  data   = 0
494 *
495 * NOTE: Any key octets not specified are assumed to be zero (the key
496 *       is assumed to be empty by default) and specifying a new key
497 *       invalidates any previous key, hence specifying a key size of
498 *       zero will clear the key (which ensures that the calculated hash
499 *       will always be zero).
500 *       The maximum size of key is algorithm and backend specific, but
501 *       is also limited by the single grant reference.
502 *       The grant reference may be read-only and must remain valid until
503 *       the response has been processed.
504 *
505 * XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
506 * -----------------------------------------
507 *
508 * This is sent by the frontend to query the maximum size of mapping
509 * table supported by the backend. The size is specified in terms of
510 * table entries.
511 *
512 * Request:
513 *
514 *  type    = XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
515 *  data[0] = 0
516 *  data[1] = 0
517 *  data[2] = 0
518 *
519 * Response:
520 *
521 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
522 *           XEN_NETIF_CTRL_STATUS_SUCCESS       - Operation successful
523 *  data   = maximum number of entries allowed in the mapping table
524 *           (if operation was successful) or zero if a mapping table is
525 *           not supported (i.e. hash mapping is done only by modular
526 *           arithmetic).
527 *
528 * XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
529 * -------------------------------------
530 *
531 * This is sent by the frontend to set the actual size of the mapping
532 * table to be used by the backend. The size is specified in terms of
533 * table entries.
534 * Any previous table is invalidated by this message and any new table
535 * is assumed to be zero filled.
536 *
537 * Request:
538 *
539 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
540 *  data[0] = number of entries in mapping table
541 *  data[1] = 0
542 *  data[2] = 0
543 *
544 * Response:
545 *
546 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
547 *                                                     supported
548 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size is invalid
549 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
550 *  data   = 0
551 *
552 * NOTE: Setting data[0] to 0 means that hash mapping should be done
553 *       using modular arithmetic.
554 *
555 * XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
556 * ------------------------------------
557 *
558 * This is sent by the frontend to set the content of the table mapping
559 * hash value to queue number. The backend should calculate the hash from
560 * the packet header, use it as an index into the table (modulo the size
561 * of the table) and then steer the packet to the queue number found at
562 * that index.
563 *
564 * Request:
565 *
566 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
567 *  data[0] = grant reference of page containing the mapping (sub-)table
568 *            (assumed to start at beginning of grant)
569 *  data[1] = size of (sub-)table in entries
570 *  data[2] = offset, in entries, of sub-table within overall table
571 *
572 * Response:
573 *
574 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
575 *                                                     supported
576 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size or content
577 *                                                     is invalid
578 *           XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW   - Table size is larger
579 *                                                     than the backend
580 *                                                     supports
581 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
582 *  data   = 0
583 *
584 * NOTE: The overall table has the following format:
585 *
586 *          0     1     2     3     4     5     6     7  octet
587 *       +-----+-----+-----+-----+-----+-----+-----+-----+
588 *       |       mapping[0]      |       mapping[1]      |
589 *       +-----+-----+-----+-----+-----+-----+-----+-----+
590 *       |                       .                       |
591 *       |                       .                       |
592 *       |                       .                       |
593 *       +-----+-----+-----+-----+-----+-----+-----+-----+
594 *       |      mapping[N-2]     |      mapping[N-1]     |
595 *       +-----+-----+-----+-----+-----+-----+-----+-----+
596 *
597 *       where N is specified by a XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
598 *       message and each  mapping must specifies a queue between 0 and
599 *       "multi-queue-num-queues" (see above).
600 *       The backend may support a mapping table larger than can be
601 *       mapped by a single grant reference. Thus sub-tables within a
602 *       larger table can be individually set by sending multiple messages
603 *       with differing offset values. Specifying a new sub-table does not
604 *       invalidate any table data outside that range.
605 *       The grant reference may be read-only and must remain valid until
606 *       the response has been processed.
607 */
608
609DEFINE_RING_TYPES(xen_netif_ctrl,
610		  struct xen_netif_ctrl_request,
611		  struct xen_netif_ctrl_response);
612
613/*
614 * Guest transmit
615 * ==============
616 *
617 * This is the 'wire' format for transmit (frontend -> backend) packets:
618 *
619 *  Fragment 1: xen_netif_tx_request_t  - flags = XEN_NETTXF_*
620 *                                    size = total packet size
621 * [Extra 1: xen_netif_extra_info_t]    - (only if fragment 1 flags include
622 *                                     XEN_NETTXF_extra_info)
623 *  ...
624 * [Extra N: xen_netif_extra_info_t]    - (only if extra N-1 flags include
625 *                                     XEN_NETIF_EXTRA_MORE)
626 *  ...
627 *  Fragment N: xen_netif_tx_request_t  - (only if fragment N-1 flags include
628 *                                     XEN_NETTXF_more_data - flags on preceding
629 *                                     extras are not relevant here)
630 *                                    flags = 0
631 *                                    size = fragment size
632 *
633 * NOTE:
634 *
635 * This format slightly is different from that used for receive
636 * (backend -> frontend) packets. Specifically, in a multi-fragment
637 * packet the actual size of fragment 1 can only be determined by
638 * subtracting the sizes of fragments 2..N from the total packet size.
639 *
640 * Ring slot size is 12 octets, however not all request/response
641 * structs use the full size.
642 *
643 * tx request data (xen_netif_tx_request_t)
644 * ------------------------------------
645 *
646 *    0     1     2     3     4     5     6     7  octet
647 * +-----+-----+-----+-----+-----+-----+-----+-----+
648 * | grant ref             | offset    | flags     |
649 * +-----+-----+-----+-----+-----+-----+-----+-----+
650 * | id        | size      |
651 * +-----+-----+-----+-----+
652 *
653 * grant ref: Reference to buffer page.
654 * offset: Offset within buffer page.
655 * flags: XEN_NETTXF_*.
656 * id: request identifier, echoed in response.
657 * size: packet size in bytes.
658 *
659 * tx response (xen_netif_tx_response_t)
660 * ---------------------------------
661 *
662 *    0     1     2     3     4     5     6     7  octet
663 * +-----+-----+-----+-----+-----+-----+-----+-----+
664 * | id        | status    | unused                |
665 * +-----+-----+-----+-----+-----+-----+-----+-----+
666 * | unused                |
667 * +-----+-----+-----+-----+
668 *
669 * id: reflects id in transmit request
670 * status: XEN_NETIF_RSP_*
671 *
672 * Guest receive
673 * =============
674 *
675 * This is the 'wire' format for receive (backend -> frontend) packets:
676 *
677 *  Fragment 1: xen_netif_rx_request_t  - flags = XEN_NETRXF_*
678 *                                    size = fragment size
679 * [Extra 1: xen_netif_extra_info_t]    - (only if fragment 1 flags include
680 *                                     XEN_NETRXF_extra_info)
681 *  ...
682 * [Extra N: xen_netif_extra_info_t]    - (only if extra N-1 flags include
683 *                                     XEN_NETIF_EXTRA_MORE)
684 *  ...
685 *  Fragment N: xen_netif_rx_request_t  - (only if fragment N-1 flags include
686 *                                     XEN_NETRXF_more_data - flags on preceding
687 *                                     extras are not relevant here)
688 *                                    flags = 0
689 *                                    size = fragment size
690 *
691 * NOTE:
692 *
693 * This format slightly is different from that used for transmit
694 * (frontend -> backend) packets. Specifically, in a multi-fragment
695 * packet the size of the packet can only be determined by summing the
696 * sizes of fragments 1..N.
697 *
698 * Ring slot size is 8 octets.
699 *
700 * rx request (xen_netif_rx_request_t)
701 * -------------------------------
702 *
703 *    0     1     2     3     4     5     6     7  octet
704 * +-----+-----+-----+-----+-----+-----+-----+-----+
705 * | id        | pad       | gref                  |
706 * +-----+-----+-----+-----+-----+-----+-----+-----+
707 *
708 * id: request identifier, echoed in response.
709 * gref: reference to incoming granted frame.
710 *
711 * rx response (xen_netif_rx_response_t)
712 * ---------------------------------
713 *
714 *    0     1     2     3     4     5     6     7  octet
715 * +-----+-----+-----+-----+-----+-----+-----+-----+
716 * | id        | offset    | flags     | status    |
717 * +-----+-----+-----+-----+-----+-----+-----+-----+
718 *
719 * id: reflects id in receive request
720 * offset: offset in page of start of received packet
721 * flags: XEN_NETRXF_*
722 * status: -ve: XEN_NETIF_RSP_*; +ve: Rx'ed pkt size.
723 *
724 * NOTE: Historically, to support GSO on the frontend receive side, Linux
725 *       netfront does not make use of the rx response id (because, as
726 *       described below, extra info structures overlay the id field).
727 *       Instead it assumes that responses always appear in the same ring
728 *       slot as their corresponding request. Thus, to maintain
729 *       compatibility, backends must make sure this is the case.
730 *
731 * Extra Info
732 * ==========
733 *
734 * Can be present if initial request or response has NET{T,R}XF_extra_info,
735 * or previous extra request has XEN_NETIF_EXTRA_MORE.
736 *
737 * The struct therefore needs to fit into either a tx or rx slot and
738 * is therefore limited to 8 octets.
739 *
740 * NOTE: Because extra info data overlays the usual request/response
741 *       structures, there is no id information in the opposite direction.
742 *       So, if an extra info overlays an rx response the frontend can
743 *       assume that it is in the same ring slot as the request that was
744 *       consumed to make the slot available, and the backend must ensure
745 *       this assumption is true.
746 *
747 * extra info (xen_netif_extra_info_t)
748 * -------------------------------
749 *
750 * General format:
751 *
752 *    0     1     2     3     4     5     6     7  octet
753 * +-----+-----+-----+-----+-----+-----+-----+-----+
754 * |type |flags| type specific data                |
755 * +-----+-----+-----+-----+-----+-----+-----+-----+
756 * | padding for tx        |
757 * +-----+-----+-----+-----+
758 *
759 * type: XEN_NETIF_EXTRA_TYPE_*
760 * flags: XEN_NETIF_EXTRA_FLAG_*
761 * padding for tx: present only in the tx case due to 8 octet limit
762 *                 from rx case. Not shown in type specific entries
763 *                 below.
764 *
765 * XEN_NETIF_EXTRA_TYPE_GSO:
766 *
767 *    0     1     2     3     4     5     6     7  octet
768 * +-----+-----+-----+-----+-----+-----+-----+-----+
769 * |type |flags| size      |type | pad | features  |
770 * +-----+-----+-----+-----+-----+-----+-----+-----+
771 *
772 * type: Must be XEN_NETIF_EXTRA_TYPE_GSO
773 * flags: XEN_NETIF_EXTRA_FLAG_*
774 * size: Maximum payload size of each segment. For example,
775 *       for TCP this is just the path MSS.
776 * type: XEN_NETIF_GSO_TYPE_*: This determines the protocol of
777 *       the packet and any extra features required to segment the
778 *       packet properly.
779 * features: EN_XEN_NETIF_GSO_FEAT_*: This specifies any extra GSO
780 *           features required to process this packet, such as ECN
781 *           support for TCPv4.
782 *
783 * XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}:
784 *
785 *    0     1     2     3     4     5     6     7  octet
786 * +-----+-----+-----+-----+-----+-----+-----+-----+
787 * |type |flags| addr                              |
788 * +-----+-----+-----+-----+-----+-----+-----+-----+
789 *
790 * type: Must be XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}
791 * flags: XEN_NETIF_EXTRA_FLAG_*
792 * addr: address to add/remove
793 *
794 * XEN_NETIF_EXTRA_TYPE_HASH:
795 *
796 * A backend that supports teoplitz hashing is assumed to accept
797 * this type of extra info in transmit packets.
798 * A frontend that enables hashing is assumed to accept
799 * this type of extra info in receive packets.
800 *
801 *    0     1     2     3     4     5     6     7  octet
802 * +-----+-----+-----+-----+-----+-----+-----+-----+
803 * |type |flags|htype| alg |LSB ---- value ---- MSB|
804 * +-----+-----+-----+-----+-----+-----+-----+-----+
805 *
806 * type: Must be XEN_NETIF_EXTRA_TYPE_HASH
807 * flags: XEN_NETIF_EXTRA_FLAG_*
808 * htype: Hash type (one of _XEN_NETIF_CTRL_HASH_TYPE_* - see above)
809 * alg: The algorithm used to calculate the hash (one of
810 *      XEN_NETIF_CTRL_HASH_TYPE_ALGORITHM_* - see above)
811 * value: Hash value
812 */
813
814/* Protocol checksum field is blank in the packet (hardware offload)? */
815#define _XEN_NETTXF_csum_blank     (0)
816#define  XEN_NETTXF_csum_blank     (1U<<_XEN_NETTXF_csum_blank)
817
818/* Packet data has been validated against protocol checksum. */
819#define _XEN_NETTXF_data_validated (1)
820#define  XEN_NETTXF_data_validated (1U<<_XEN_NETTXF_data_validated)
821
822/* Packet continues in the next request descriptor. */
823#define _XEN_NETTXF_more_data      (2)
824#define  XEN_NETTXF_more_data      (1U<<_XEN_NETTXF_more_data)
825
826/* Packet to be followed by extra descriptor(s). */
827#define _XEN_NETTXF_extra_info     (3)
828#define  XEN_NETTXF_extra_info     (1U<<_XEN_NETTXF_extra_info)
829
830#define XEN_NETIF_MAX_TX_SIZE 0xFFFF
831struct xen_netif_tx_request {
832	grant_ref_t gref;
833	uint16_t offset;
834	uint16_t flags;
835	uint16_t id;
836	uint16_t size;
837};
838
839/* Types of xen_netif_extra_info descriptors. */
840#define XEN_NETIF_EXTRA_TYPE_NONE      (0)	/* Never used - invalid */
841#define XEN_NETIF_EXTRA_TYPE_GSO       (1)	/* u.gso */
842#define XEN_NETIF_EXTRA_TYPE_MCAST_ADD (2)	/* u.mcast */
843#define XEN_NETIF_EXTRA_TYPE_MCAST_DEL (3)	/* u.mcast */
844#define XEN_NETIF_EXTRA_TYPE_HASH      (4)	/* u.hash */
845#define XEN_NETIF_EXTRA_TYPE_XDP       (5)	/* u.xdp */
846#define XEN_NETIF_EXTRA_TYPE_MAX       (6)
847
848/* xen_netif_extra_info_t flags. */
849#define _XEN_NETIF_EXTRA_FLAG_MORE (0)
850#define XEN_NETIF_EXTRA_FLAG_MORE  (1U<<_XEN_NETIF_EXTRA_FLAG_MORE)
851
852/* GSO types */
853#define XEN_NETIF_GSO_TYPE_NONE         (0)
854#define XEN_NETIF_GSO_TYPE_TCPV4        (1)
855#define XEN_NETIF_GSO_TYPE_TCPV6        (2)
856
857/*
858 * This structure needs to fit within both xen_netif_tx_request_t and
859 * xen_netif_rx_response_t for compatibility.
860 */
861struct xen_netif_extra_info {
862	uint8_t type;
863	uint8_t flags;
864	union {
865		struct {
866			uint16_t size;
867			uint8_t type;
868			uint8_t pad;
869			uint16_t features;
870		} gso;
871		struct {
872			uint8_t addr[6];
873		} mcast;
874		struct {
875			uint8_t type;
876			uint8_t algorithm;
877			uint8_t value[4];
878		} hash;
879		struct {
880			uint16_t headroom;
881			uint16_t pad[2];
882		} xdp;
883		uint16_t pad[3];
884	} u;
885};
886
887struct xen_netif_tx_response {
888	uint16_t id;
889	int16_t status;
890};
891
892struct xen_netif_rx_request {
893	uint16_t id;		/* Echoed in response message.        */
894	uint16_t pad;
895	grant_ref_t gref;
896};
897
898/* Packet data has been validated against protocol checksum. */
899#define _XEN_NETRXF_data_validated (0)
900#define  XEN_NETRXF_data_validated (1U<<_XEN_NETRXF_data_validated)
901
902/* Protocol checksum field is blank in the packet (hardware offload)? */
903#define _XEN_NETRXF_csum_blank     (1)
904#define  XEN_NETRXF_csum_blank     (1U<<_XEN_NETRXF_csum_blank)
905
906/* Packet continues in the next request descriptor. */
907#define _XEN_NETRXF_more_data      (2)
908#define  XEN_NETRXF_more_data      (1U<<_XEN_NETRXF_more_data)
909
910/* Packet to be followed by extra descriptor(s). */
911#define _XEN_NETRXF_extra_info     (3)
912#define  XEN_NETRXF_extra_info     (1U<<_XEN_NETRXF_extra_info)
913
914/* Packet has GSO prefix. Deprecated but included for compatibility */
915#define _XEN_NETRXF_gso_prefix     (4)
916#define  XEN_NETRXF_gso_prefix     (1U<<_XEN_NETRXF_gso_prefix)
917
918struct xen_netif_rx_response {
919	uint16_t id;
920	uint16_t offset;
921	uint16_t flags;
922	int16_t status;
923};
924
925/*
926 * Generate xen_netif ring structures and types.
927 */
928
929DEFINE_RING_TYPES(xen_netif_tx, struct xen_netif_tx_request,
930		  struct xen_netif_tx_response);
931DEFINE_RING_TYPES(xen_netif_rx, struct xen_netif_rx_request,
932		  struct xen_netif_rx_response);
933
934#define XEN_NETIF_RSP_DROPPED         -2
935#define XEN_NETIF_RSP_ERROR           -1
936#define XEN_NETIF_RSP_OKAY             0
937/* No response: used for auxiliary requests (e.g., xen_netif_extra_info_t). */
938#define XEN_NETIF_RSP_NULL             1
939
940#endif
v5.14.15
 
  1/******************************************************************************
  2 * xen_netif.h
  3 *
  4 * Unified network-device I/O interface for Xen guest OSes.
  5 *
  6 * Permission is hereby granted, free of charge, to any person obtaining a copy
  7 * of this software and associated documentation files (the "Software"), to
  8 * deal in the Software without restriction, including without limitation the
  9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
 10 * sell copies of the Software, and to permit persons to whom the Software is
 11 * furnished to do so, subject to the following conditions:
 12 *
 13 * The above copyright notice and this permission notice shall be included in
 14 * all copies or substantial portions of the Software.
 15 *
 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
 19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 21 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
 22 * DEALINGS IN THE SOFTWARE.
 23 *
 24 * Copyright (c) 2003-2004, Keir Fraser
 25 */
 26
 27#ifndef __XEN_PUBLIC_IO_XEN_NETIF_H__
 28#define __XEN_PUBLIC_IO_XEN_NETIF_H__
 29
 30#include "ring.h"
 31#include "../grant_table.h"
 32
 33/*
 34 * Older implementation of Xen network frontend / backend has an
 35 * implicit dependency on the MAX_SKB_FRAGS as the maximum number of
 36 * ring slots a skb can use. Netfront / netback may not work as
 37 * expected when frontend and backend have different MAX_SKB_FRAGS.
 38 *
 39 * A better approach is to add mechanism for netfront / netback to
 40 * negotiate this value. However we cannot fix all possible
 41 * frontends, so we need to define a value which states the minimum
 42 * slots backend must support.
 43 *
 44 * The minimum value derives from older Linux kernel's MAX_SKB_FRAGS
 45 * (18), which is proved to work with most frontends. Any new backend
 46 * which doesn't negotiate with frontend should expect frontend to
 47 * send a valid packet using slots up to this value.
 48 */
 49#define XEN_NETIF_NR_SLOTS_MIN 18
 50
 51/*
 52 * Notifications after enqueuing any type of message should be conditional on
 53 * the appropriate req_event or rsp_event field in the shared ring.
 54 * If the client sends notification for rx requests then it should specify
 55 * feature 'feature-rx-notify' via xenbus. Otherwise the backend will assume
 56 * that it cannot safely queue packets (as it may not be kicked to send them).
 57 */
 58
 59/*
 60 * "feature-split-event-channels" is introduced to separate guest TX
 61 * and RX notification. Backend either doesn't support this feature or
 62 * advertises it via xenstore as 0 (disabled) or 1 (enabled).
 63 *
 64 * To make use of this feature, frontend should allocate two event
 65 * channels for TX and RX, advertise them to backend as
 66 * "event-channel-tx" and "event-channel-rx" respectively. If frontend
 67 * doesn't want to use this feature, it just writes "event-channel"
 68 * node as before.
 69 */
 70
 71/*
 72 * Multiple transmit and receive queues:
 73 * If supported, the backend will write the key "multi-queue-max-queues" to
 74 * the directory for that vif, and set its value to the maximum supported
 75 * number of queues.
 76 * Frontends that are aware of this feature and wish to use it can write the
 77 * key "multi-queue-num-queues", set to the number they wish to use, which
 78 * must be greater than zero, and no more than the value reported by the backend
 79 * in "multi-queue-max-queues".
 80 *
 81 * Queues replicate the shared rings and event channels.
 82 * "feature-split-event-channels" may optionally be used when using
 83 * multiple queues, but is not mandatory.
 84 *
 85 * Each queue consists of one shared ring pair, i.e. there must be the same
 86 * number of tx and rx rings.
 87 *
 88 * For frontends requesting just one queue, the usual event-channel and
 89 * ring-ref keys are written as before, simplifying the backend processing
 90 * to avoid distinguishing between a frontend that doesn't understand the
 91 * multi-queue feature, and one that does, but requested only one queue.
 92 *
 93 * Frontends requesting two or more queues must not write the toplevel
 94 * event-channel (or event-channel-{tx,rx}) and {tx,rx}-ring-ref keys,
 95 * instead writing those keys under sub-keys having the name "queue-N" where
 96 * N is the integer ID of the queue for which those keys belong. Queues
 97 * are indexed from zero. For example, a frontend with two queues and split
 98 * event channels must write the following set of queue-related keys:
 99 *
100 * /local/domain/1/device/vif/0/multi-queue-num-queues = "2"
101 * /local/domain/1/device/vif/0/queue-0 = ""
102 * /local/domain/1/device/vif/0/queue-0/tx-ring-ref = "<ring-ref-tx0>"
103 * /local/domain/1/device/vif/0/queue-0/rx-ring-ref = "<ring-ref-rx0>"
104 * /local/domain/1/device/vif/0/queue-0/event-channel-tx = "<evtchn-tx0>"
105 * /local/domain/1/device/vif/0/queue-0/event-channel-rx = "<evtchn-rx0>"
106 * /local/domain/1/device/vif/0/queue-1 = ""
107 * /local/domain/1/device/vif/0/queue-1/tx-ring-ref = "<ring-ref-tx1>"
108 * /local/domain/1/device/vif/0/queue-1/rx-ring-ref = "<ring-ref-rx1"
109 * /local/domain/1/device/vif/0/queue-1/event-channel-tx = "<evtchn-tx1>"
110 * /local/domain/1/device/vif/0/queue-1/event-channel-rx = "<evtchn-rx1>"
111 *
112 * If there is any inconsistency in the XenStore data, the backend may
113 * choose not to connect any queues, instead treating the request as an
114 * error. This includes scenarios where more (or fewer) queues were
115 * requested than the frontend provided details for.
116 *
117 * Mapping of packets to queues is considered to be a function of the
118 * transmitting system (backend or frontend) and is not negotiated
119 * between the two. Guests are free to transmit packets on any queue
120 * they choose, provided it has been set up correctly. Guests must be
121 * prepared to receive packets on any queue they have requested be set up.
122 */
123
124/*
125 * "feature-no-csum-offload" should be used to turn IPv4 TCP/UDP checksum
126 * offload off or on. If it is missing then the feature is assumed to be on.
127 * "feature-ipv6-csum-offload" should be used to turn IPv6 TCP/UDP checksum
128 * offload on or off. If it is missing then the feature is assumed to be off.
129 */
130
131/*
132 * "feature-gso-tcpv4" and "feature-gso-tcpv6" advertise the capability to
133 * handle large TCP packets (in IPv4 or IPv6 form respectively). Neither
134 * frontends nor backends are assumed to be capable unless the flags are
135 * present.
136 */
137
138/*
139 * "feature-multicast-control" and "feature-dynamic-multicast-control"
140 * advertise the capability to filter ethernet multicast packets in the
141 * backend. If the frontend wishes to take advantage of this feature then
142 * it may set "request-multicast-control". If the backend only advertises
143 * "feature-multicast-control" then "request-multicast-control" must be set
144 * before the frontend moves into the connected state. The backend will
145 * sample the value on this state transition and any subsequent change in
146 * value will have no effect. However, if the backend also advertises
147 * "feature-dynamic-multicast-control" then "request-multicast-control"
148 * may be set by the frontend at any time. In this case, the backend will
149 * watch the value and re-sample on watch events.
150 *
151 * If the sampled value of "request-multicast-control" is set then the
152 * backend transmit side should no longer flood multicast packets to the
153 * frontend, it should instead drop any multicast packet that does not
154 * match in a filter list.
155 * The list is amended by the frontend by sending dummy transmit requests
156 * containing XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL} extra-info fragments as
157 * specified below.
158 * Note that the filter list may be amended even if the sampled value of
159 * "request-multicast-control" is not set, however the filter should only
160 * be applied if it is set.
161 */
162
163/*
164 * "xdp-headroom" is used to request that extra space is added
165 * for XDP processing.  The value is measured in bytes and passed by
166 * the frontend to be consistent between both ends.
167 * If the value is greater than zero that means that
168 * an RX response is going to be passed to an XDP program for processing.
169 * XEN_NETIF_MAX_XDP_HEADROOM defines the maximum headroom offset in bytes
170 *
171 * "feature-xdp-headroom" is set to "1" by the netback side like other features
172 * so a guest can check if an XDP program can be processed.
173 */
174#define XEN_NETIF_MAX_XDP_HEADROOM 0x7FFF
175
176/*
177 * Control ring
178 * ============
179 *
180 * Some features, such as hashing (detailed below), require a
181 * significant amount of out-of-band data to be passed from frontend to
182 * backend. Use of xenstore is not suitable for large quantities of data
183 * because of quota limitations and so a dedicated 'control ring' is used.
184 * The ability of the backend to use a control ring is advertised by
185 * setting:
186 *
187 * /local/domain/X/backend/<domid>/<vif>/feature-ctrl-ring = "1"
188 *
189 * The frontend provides a control ring to the backend by setting:
190 *
191 * /local/domain/<domid>/device/vif/<vif>/ctrl-ring-ref = <gref>
192 * /local/domain/<domid>/device/vif/<vif>/event-channel-ctrl = <port>
193 *
194 * where <gref> is the grant reference of the shared page used to
195 * implement the control ring and <port> is an event channel to be used
196 * as a mailbox interrupt. These keys must be set before the frontend
197 * moves into the connected state.
198 *
199 * The control ring uses a fixed request/response message size and is
200 * balanced (i.e. one request to one response), so operationally it is much
201 * the same as a transmit or receive ring.
202 * Note that there is no requirement that responses are issued in the same
203 * order as requests.
204 */
205
206/*
207 * Hash types
208 * ==========
209 *
210 * For the purposes of the definitions below, 'Packet[]' is an array of
211 * octets containing an IP packet without options, 'Array[X..Y]' means a
212 * sub-array of 'Array' containing bytes X thru Y inclusive, and '+' is
213 * used to indicate concatenation of arrays.
214 */
215
216/*
217 * A hash calculated over an IP version 4 header as follows:
218 *
219 * Buffer[0..8] = Packet[12..15] (source address) +
220 *                Packet[16..19] (destination address)
221 *
222 * Result = Hash(Buffer, 8)
223 */
224#define _XEN_NETIF_CTRL_HASH_TYPE_IPV4 0
225#define XEN_NETIF_CTRL_HASH_TYPE_IPV4 \
226	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4)
227
228/*
229 * A hash calculated over an IP version 4 header and TCP header as
230 * follows:
231 *
232 * Buffer[0..12] = Packet[12..15] (source address) +
233 *                 Packet[16..19] (destination address) +
234 *                 Packet[20..21] (source port) +
235 *                 Packet[22..23] (destination port)
236 *
237 * Result = Hash(Buffer, 12)
238 */
239#define _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP 1
240#define XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP \
241	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP)
242
243/*
244 * A hash calculated over an IP version 6 header as follows:
245 *
246 * Buffer[0..32] = Packet[8..23]  (source address ) +
247 *                 Packet[24..39] (destination address)
248 *
249 * Result = Hash(Buffer, 32)
250 */
251#define _XEN_NETIF_CTRL_HASH_TYPE_IPV6 2
252#define XEN_NETIF_CTRL_HASH_TYPE_IPV6 \
253	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6)
254
255/*
256 * A hash calculated over an IP version 6 header and TCP header as
257 * follows:
258 *
259 * Buffer[0..36] = Packet[8..23]  (source address) +
260 *                 Packet[24..39] (destination address) +
261 *                 Packet[40..41] (source port) +
262 *                 Packet[42..43] (destination port)
263 *
264 * Result = Hash(Buffer, 36)
265 */
266#define _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP 3
267#define XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP \
268	(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)
269
270/*
271 * Hash algorithms
272 * ===============
273 */
274
275#define XEN_NETIF_CTRL_HASH_ALGORITHM_NONE 0
276
277/*
278 * Toeplitz hash:
279 */
280
281#define XEN_NETIF_CTRL_HASH_ALGORITHM_TOEPLITZ 1
282
283/*
284 * This algorithm uses a 'key' as well as the data buffer itself.
285 * (Buffer[] and Key[] are treated as shift-registers where the MSB of
286 * Buffer/Key[0] is considered 'left-most' and the LSB of Buffer/Key[N-1]
287 * is the 'right-most').
288 *
289 * Value = 0
290 * For number of bits in Buffer[]
291 *    If (left-most bit of Buffer[] is 1)
292 *        Value ^= left-most 32 bits of Key[]
293 *    Key[] << 1
294 *    Buffer[] << 1
295 *
296 * The code below is provided for convenience where an operating system
297 * does not already provide an implementation.
298 */
299#ifdef XEN_NETIF_DEFINE_TOEPLITZ
300static uint32_t xen_netif_toeplitz_hash(const uint8_t *key,
301					unsigned int keylen,
302					const uint8_t *buf, unsigned int buflen)
303{
304	unsigned int keyi, bufi;
305	uint64_t prefix = 0;
306	uint64_t hash = 0;
307
308	/* Pre-load prefix with the first 8 bytes of the key */
309	for (keyi = 0; keyi < 8; keyi++) {
310		prefix <<= 8;
311		prefix |= (keyi < keylen) ? key[keyi] : 0;
312	}
313
314	for (bufi = 0; bufi < buflen; bufi++) {
315		uint8_t byte = buf[bufi];
316		unsigned int bit;
317
318		for (bit = 0; bit < 8; bit++) {
319			if (byte & 0x80)
320				hash ^= prefix;
321			prefix <<= 1;
322			byte <<= 1;
323		}
324
325		/*
326		 * 'prefix' has now been left-shifted by 8, so
327		 * OR in the next byte.
328		 */
329		prefix |= (keyi < keylen) ? key[keyi] : 0;
330		keyi++;
331	}
332
333	/* The valid part of the hash is in the upper 32 bits. */
334	return hash >> 32;
335}
336#endif				/* XEN_NETIF_DEFINE_TOEPLITZ */
337
338/*
339 * Control requests (struct xen_netif_ctrl_request)
340 * ================================================
341 *
342 * All requests have the following format:
343 *
344 *    0     1     2     3     4     5     6     7  octet
345 * +-----+-----+-----+-----+-----+-----+-----+-----+
346 * |    id     |   type    |         data[0]       |
347 * +-----+-----+-----+-----+-----+-----+-----+-----+
348 * |         data[1]       |         data[2]       |
349 * +-----+-----+-----+-----+-----------------------+
350 *
351 * id: the request identifier, echoed in response.
352 * type: the type of request (see below)
353 * data[]: any data associated with the request (determined by type)
354 */
355
356struct xen_netif_ctrl_request {
357	uint16_t id;
358	uint16_t type;
359
360#define XEN_NETIF_CTRL_TYPE_INVALID               0
361#define XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS        1
362#define XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS        2
363#define XEN_NETIF_CTRL_TYPE_SET_HASH_KEY          3
364#define XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE 4
365#define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE 5
366#define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING      6
367#define XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM    7
368
369	uint32_t data[3];
370};
371
372/*
373 * Control responses (struct xen_netif_ctrl_response)
374 * ==================================================
375 *
376 * All responses have the following format:
377 *
378 *    0     1     2     3     4     5     6     7  octet
379 * +-----+-----+-----+-----+-----+-----+-----+-----+
380 * |    id     |   type    |         status        |
381 * +-----+-----+-----+-----+-----+-----+-----+-----+
382 * |         data          |
383 * +-----+-----+-----+-----+
384 *
385 * id: the corresponding request identifier
386 * type: the type of the corresponding request
387 * status: the status of request processing
388 * data: any data associated with the response (determined by type and
389 *       status)
390 */
391
392struct xen_netif_ctrl_response {
393	uint16_t id;
394	uint16_t type;
395	uint32_t status;
396
397#define XEN_NETIF_CTRL_STATUS_SUCCESS           0
398#define XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     1
399#define XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER 2
400#define XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW   3
401
402	uint32_t data;
403};
404
405/*
406 * Control messages
407 * ================
408 *
409 * XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
410 * --------------------------------------
411 *
412 * This is sent by the frontend to set the desired hash algorithm.
413 *
414 * Request:
415 *
416 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
417 *  data[0] = a XEN_NETIF_CTRL_HASH_ALGORITHM_* value
418 *  data[1] = 0
419 *  data[2] = 0
420 *
421 * Response:
422 *
423 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
424 *                                                     supported
425 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - The algorithm is not
426 *                                                     supported
427 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
428 *
429 * NOTE: Setting data[0] to XEN_NETIF_CTRL_HASH_ALGORITHM_NONE disables
430 *       hashing and the backend is free to choose how it steers packets
431 *       to queues (which is the default behaviour).
432 *
433 * XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
434 * ----------------------------------
435 *
436 * This is sent by the frontend to query the types of hash supported by
437 * the backend.
438 *
439 * Request:
440 *
441 *  type    = XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
442 *  data[0] = 0
443 *  data[1] = 0
444 *  data[2] = 0
445 *
446 * Response:
447 *
448 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
449 *           XEN_NETIF_CTRL_STATUS_SUCCESS       - Operation successful
450 *  data   = supported hash types (if operation was successful)
451 *
452 * NOTE: A valid hash algorithm must be selected before this operation can
453 *       succeed.
454 *
455 * XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
456 * ----------------------------------
457 *
458 * This is sent by the frontend to set the types of hash that the backend
459 * should calculate. (See above for hash type definitions).
460 * Note that the 'maximal' type of hash should always be chosen. For
461 * example, if the frontend sets both IPV4 and IPV4_TCP hash types then
462 * the latter hash type should be calculated for any TCP packet and the
463 * former only calculated for non-TCP packets.
464 *
465 * Request:
466 *
467 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
468 *  data[0] = bitwise OR of XEN_NETIF_CTRL_HASH_TYPE_* values
469 *  data[1] = 0
470 *  data[2] = 0
471 *
472 * Response:
473 *
474 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
475 *                                                     supported
476 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - One or more flag
477 *                                                     value is invalid or
478 *                                                     unsupported
479 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
480 *  data   = 0
481 *
482 * NOTE: A valid hash algorithm must be selected before this operation can
483 *       succeed.
484 *       Also, setting data[0] to zero disables hashing and the backend
485 *       is free to choose how it steers packets to queues.
486 *
487 * XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
488 * --------------------------------
489 *
490 * This is sent by the frontend to set the key of the hash if the algorithm
491 * requires it. (See hash algorithms above).
492 *
493 * Request:
494 *
495 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
496 *  data[0] = grant reference of page containing the key (assumed to
497 *            start at beginning of grant)
498 *  data[1] = size of key in octets
499 *  data[2] = 0
500 *
501 * Response:
502 *
503 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
504 *                                                     supported
505 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Key size is invalid
506 *           XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW   - Key size is larger
507 *                                                     than the backend
508 *                                                     supports
509 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
510 *  data   = 0
511 *
512 * NOTE: Any key octets not specified are assumed to be zero (the key
513 *       is assumed to be empty by default) and specifying a new key
514 *       invalidates any previous key, hence specifying a key size of
515 *       zero will clear the key (which ensures that the calculated hash
516 *       will always be zero).
517 *       The maximum size of key is algorithm and backend specific, but
518 *       is also limited by the single grant reference.
519 *       The grant reference may be read-only and must remain valid until
520 *       the response has been processed.
521 *
522 * XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
523 * -----------------------------------------
524 *
525 * This is sent by the frontend to query the maximum size of mapping
526 * table supported by the backend. The size is specified in terms of
527 * table entries.
528 *
529 * Request:
530 *
531 *  type    = XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
532 *  data[0] = 0
533 *  data[1] = 0
534 *  data[2] = 0
535 *
536 * Response:
537 *
538 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
539 *           XEN_NETIF_CTRL_STATUS_SUCCESS       - Operation successful
540 *  data   = maximum number of entries allowed in the mapping table
541 *           (if operation was successful) or zero if a mapping table is
542 *           not supported (i.e. hash mapping is done only by modular
543 *           arithmetic).
544 *
545 * XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
546 * -------------------------------------
547 *
548 * This is sent by the frontend to set the actual size of the mapping
549 * table to be used by the backend. The size is specified in terms of
550 * table entries.
551 * Any previous table is invalidated by this message and any new table
552 * is assumed to be zero filled.
553 *
554 * Request:
555 *
556 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
557 *  data[0] = number of entries in mapping table
558 *  data[1] = 0
559 *  data[2] = 0
560 *
561 * Response:
562 *
563 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
564 *                                                     supported
565 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size is invalid
566 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
567 *  data   = 0
568 *
569 * NOTE: Setting data[0] to 0 means that hash mapping should be done
570 *       using modular arithmetic.
571 *
572 * XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
573 * ------------------------------------
574 *
575 * This is sent by the frontend to set the content of the table mapping
576 * hash value to queue number. The backend should calculate the hash from
577 * the packet header, use it as an index into the table (modulo the size
578 * of the table) and then steer the packet to the queue number found at
579 * that index.
580 *
581 * Request:
582 *
583 *  type    = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
584 *  data[0] = grant reference of page containing the mapping (sub-)table
585 *            (assumed to start at beginning of grant)
586 *  data[1] = size of (sub-)table in entries
587 *  data[2] = offset, in entries, of sub-table within overall table
588 *
589 * Response:
590 *
591 *  status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED     - Operation not
592 *                                                     supported
593 *           XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size or content
594 *                                                     is invalid
595 *           XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW   - Table size is larger
596 *                                                     than the backend
597 *                                                     supports
598 *           XEN_NETIF_CTRL_STATUS_SUCCESS           - Operation successful
599 *  data   = 0
600 *
601 * NOTE: The overall table has the following format:
602 *
603 *          0     1     2     3     4     5     6     7  octet
604 *       +-----+-----+-----+-----+-----+-----+-----+-----+
605 *       |       mapping[0]      |       mapping[1]      |
606 *       +-----+-----+-----+-----+-----+-----+-----+-----+
607 *       |                       .                       |
608 *       |                       .                       |
609 *       |                       .                       |
610 *       +-----+-----+-----+-----+-----+-----+-----+-----+
611 *       |      mapping[N-2]     |      mapping[N-1]     |
612 *       +-----+-----+-----+-----+-----+-----+-----+-----+
613 *
614 *       where N is specified by a XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
615 *       message and each  mapping must specifies a queue between 0 and
616 *       "multi-queue-num-queues" (see above).
617 *       The backend may support a mapping table larger than can be
618 *       mapped by a single grant reference. Thus sub-tables within a
619 *       larger table can be individually set by sending multiple messages
620 *       with differing offset values. Specifying a new sub-table does not
621 *       invalidate any table data outside that range.
622 *       The grant reference may be read-only and must remain valid until
623 *       the response has been processed.
624 */
625
626DEFINE_RING_TYPES(xen_netif_ctrl,
627		  struct xen_netif_ctrl_request,
628		  struct xen_netif_ctrl_response);
629
630/*
631 * Guest transmit
632 * ==============
633 *
634 * This is the 'wire' format for transmit (frontend -> backend) packets:
635 *
636 *  Fragment 1: xen_netif_tx_request_t  - flags = XEN_NETTXF_*
637 *                                    size = total packet size
638 * [Extra 1: xen_netif_extra_info_t]    - (only if fragment 1 flags include
639 *                                     XEN_NETTXF_extra_info)
640 *  ...
641 * [Extra N: xen_netif_extra_info_t]    - (only if extra N-1 flags include
642 *                                     XEN_NETIF_EXTRA_MORE)
643 *  ...
644 *  Fragment N: xen_netif_tx_request_t  - (only if fragment N-1 flags include
645 *                                     XEN_NETTXF_more_data - flags on preceding
646 *                                     extras are not relevant here)
647 *                                    flags = 0
648 *                                    size = fragment size
649 *
650 * NOTE:
651 *
652 * This format slightly is different from that used for receive
653 * (backend -> frontend) packets. Specifically, in a multi-fragment
654 * packet the actual size of fragment 1 can only be determined by
655 * subtracting the sizes of fragments 2..N from the total packet size.
656 *
657 * Ring slot size is 12 octets, however not all request/response
658 * structs use the full size.
659 *
660 * tx request data (xen_netif_tx_request_t)
661 * ------------------------------------
662 *
663 *    0     1     2     3     4     5     6     7  octet
664 * +-----+-----+-----+-----+-----+-----+-----+-----+
665 * | grant ref             | offset    | flags     |
666 * +-----+-----+-----+-----+-----+-----+-----+-----+
667 * | id        | size      |
668 * +-----+-----+-----+-----+
669 *
670 * grant ref: Reference to buffer page.
671 * offset: Offset within buffer page.
672 * flags: XEN_NETTXF_*.
673 * id: request identifier, echoed in response.
674 * size: packet size in bytes.
675 *
676 * tx response (xen_netif_tx_response_t)
677 * ---------------------------------
678 *
679 *    0     1     2     3     4     5     6     7  octet
680 * +-----+-----+-----+-----+-----+-----+-----+-----+
681 * | id        | status    | unused                |
682 * +-----+-----+-----+-----+-----+-----+-----+-----+
683 * | unused                |
684 * +-----+-----+-----+-----+
685 *
686 * id: reflects id in transmit request
687 * status: XEN_NETIF_RSP_*
688 *
689 * Guest receive
690 * =============
691 *
692 * This is the 'wire' format for receive (backend -> frontend) packets:
693 *
694 *  Fragment 1: xen_netif_rx_request_t  - flags = XEN_NETRXF_*
695 *                                    size = fragment size
696 * [Extra 1: xen_netif_extra_info_t]    - (only if fragment 1 flags include
697 *                                     XEN_NETRXF_extra_info)
698 *  ...
699 * [Extra N: xen_netif_extra_info_t]    - (only if extra N-1 flags include
700 *                                     XEN_NETIF_EXTRA_MORE)
701 *  ...
702 *  Fragment N: xen_netif_rx_request_t  - (only if fragment N-1 flags include
703 *                                     XEN_NETRXF_more_data - flags on preceding
704 *                                     extras are not relevant here)
705 *                                    flags = 0
706 *                                    size = fragment size
707 *
708 * NOTE:
709 *
710 * This format slightly is different from that used for transmit
711 * (frontend -> backend) packets. Specifically, in a multi-fragment
712 * packet the size of the packet can only be determined by summing the
713 * sizes of fragments 1..N.
714 *
715 * Ring slot size is 8 octets.
716 *
717 * rx request (xen_netif_rx_request_t)
718 * -------------------------------
719 *
720 *    0     1     2     3     4     5     6     7  octet
721 * +-----+-----+-----+-----+-----+-----+-----+-----+
722 * | id        | pad       | gref                  |
723 * +-----+-----+-----+-----+-----+-----+-----+-----+
724 *
725 * id: request identifier, echoed in response.
726 * gref: reference to incoming granted frame.
727 *
728 * rx response (xen_netif_rx_response_t)
729 * ---------------------------------
730 *
731 *    0     1     2     3     4     5     6     7  octet
732 * +-----+-----+-----+-----+-----+-----+-----+-----+
733 * | id        | offset    | flags     | status    |
734 * +-----+-----+-----+-----+-----+-----+-----+-----+
735 *
736 * id: reflects id in receive request
737 * offset: offset in page of start of received packet
738 * flags: XEN_NETRXF_*
739 * status: -ve: XEN_NETIF_RSP_*; +ve: Rx'ed pkt size.
740 *
741 * NOTE: Historically, to support GSO on the frontend receive side, Linux
742 *       netfront does not make use of the rx response id (because, as
743 *       described below, extra info structures overlay the id field).
744 *       Instead it assumes that responses always appear in the same ring
745 *       slot as their corresponding request. Thus, to maintain
746 *       compatibility, backends must make sure this is the case.
747 *
748 * Extra Info
749 * ==========
750 *
751 * Can be present if initial request or response has NET{T,R}XF_extra_info,
752 * or previous extra request has XEN_NETIF_EXTRA_MORE.
753 *
754 * The struct therefore needs to fit into either a tx or rx slot and
755 * is therefore limited to 8 octets.
756 *
757 * NOTE: Because extra info data overlays the usual request/response
758 *       structures, there is no id information in the opposite direction.
759 *       So, if an extra info overlays an rx response the frontend can
760 *       assume that it is in the same ring slot as the request that was
761 *       consumed to make the slot available, and the backend must ensure
762 *       this assumption is true.
763 *
764 * extra info (xen_netif_extra_info_t)
765 * -------------------------------
766 *
767 * General format:
768 *
769 *    0     1     2     3     4     5     6     7  octet
770 * +-----+-----+-----+-----+-----+-----+-----+-----+
771 * |type |flags| type specific data                |
772 * +-----+-----+-----+-----+-----+-----+-----+-----+
773 * | padding for tx        |
774 * +-----+-----+-----+-----+
775 *
776 * type: XEN_NETIF_EXTRA_TYPE_*
777 * flags: XEN_NETIF_EXTRA_FLAG_*
778 * padding for tx: present only in the tx case due to 8 octet limit
779 *                 from rx case. Not shown in type specific entries
780 *                 below.
781 *
782 * XEN_NETIF_EXTRA_TYPE_GSO:
783 *
784 *    0     1     2     3     4     5     6     7  octet
785 * +-----+-----+-----+-----+-----+-----+-----+-----+
786 * |type |flags| size      |type | pad | features  |
787 * +-----+-----+-----+-----+-----+-----+-----+-----+
788 *
789 * type: Must be XEN_NETIF_EXTRA_TYPE_GSO
790 * flags: XEN_NETIF_EXTRA_FLAG_*
791 * size: Maximum payload size of each segment. For example,
792 *       for TCP this is just the path MSS.
793 * type: XEN_NETIF_GSO_TYPE_*: This determines the protocol of
794 *       the packet and any extra features required to segment the
795 *       packet properly.
796 * features: EN_XEN_NETIF_GSO_FEAT_*: This specifies any extra GSO
797 *           features required to process this packet, such as ECN
798 *           support for TCPv4.
799 *
800 * XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}:
801 *
802 *    0     1     2     3     4     5     6     7  octet
803 * +-----+-----+-----+-----+-----+-----+-----+-----+
804 * |type |flags| addr                              |
805 * +-----+-----+-----+-----+-----+-----+-----+-----+
806 *
807 * type: Must be XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}
808 * flags: XEN_NETIF_EXTRA_FLAG_*
809 * addr: address to add/remove
810 *
811 * XEN_NETIF_EXTRA_TYPE_HASH:
812 *
813 * A backend that supports teoplitz hashing is assumed to accept
814 * this type of extra info in transmit packets.
815 * A frontend that enables hashing is assumed to accept
816 * this type of extra info in receive packets.
817 *
818 *    0     1     2     3     4     5     6     7  octet
819 * +-----+-----+-----+-----+-----+-----+-----+-----+
820 * |type |flags|htype| alg |LSB ---- value ---- MSB|
821 * +-----+-----+-----+-----+-----+-----+-----+-----+
822 *
823 * type: Must be XEN_NETIF_EXTRA_TYPE_HASH
824 * flags: XEN_NETIF_EXTRA_FLAG_*
825 * htype: Hash type (one of _XEN_NETIF_CTRL_HASH_TYPE_* - see above)
826 * alg: The algorithm used to calculate the hash (one of
827 *      XEN_NETIF_CTRL_HASH_TYPE_ALGORITHM_* - see above)
828 * value: Hash value
829 */
830
831/* Protocol checksum field is blank in the packet (hardware offload)? */
832#define _XEN_NETTXF_csum_blank     (0)
833#define  XEN_NETTXF_csum_blank     (1U<<_XEN_NETTXF_csum_blank)
834
835/* Packet data has been validated against protocol checksum. */
836#define _XEN_NETTXF_data_validated (1)
837#define  XEN_NETTXF_data_validated (1U<<_XEN_NETTXF_data_validated)
838
839/* Packet continues in the next request descriptor. */
840#define _XEN_NETTXF_more_data      (2)
841#define  XEN_NETTXF_more_data      (1U<<_XEN_NETTXF_more_data)
842
843/* Packet to be followed by extra descriptor(s). */
844#define _XEN_NETTXF_extra_info     (3)
845#define  XEN_NETTXF_extra_info     (1U<<_XEN_NETTXF_extra_info)
846
847#define XEN_NETIF_MAX_TX_SIZE 0xFFFF
848struct xen_netif_tx_request {
849	grant_ref_t gref;
850	uint16_t offset;
851	uint16_t flags;
852	uint16_t id;
853	uint16_t size;
854};
855
856/* Types of xen_netif_extra_info descriptors. */
857#define XEN_NETIF_EXTRA_TYPE_NONE      (0)	/* Never used - invalid */
858#define XEN_NETIF_EXTRA_TYPE_GSO       (1)	/* u.gso */
859#define XEN_NETIF_EXTRA_TYPE_MCAST_ADD (2)	/* u.mcast */
860#define XEN_NETIF_EXTRA_TYPE_MCAST_DEL (3)	/* u.mcast */
861#define XEN_NETIF_EXTRA_TYPE_HASH      (4)	/* u.hash */
862#define XEN_NETIF_EXTRA_TYPE_XDP       (5)	/* u.xdp */
863#define XEN_NETIF_EXTRA_TYPE_MAX       (6)
864
865/* xen_netif_extra_info_t flags. */
866#define _XEN_NETIF_EXTRA_FLAG_MORE (0)
867#define XEN_NETIF_EXTRA_FLAG_MORE  (1U<<_XEN_NETIF_EXTRA_FLAG_MORE)
868
869/* GSO types */
870#define XEN_NETIF_GSO_TYPE_NONE         (0)
871#define XEN_NETIF_GSO_TYPE_TCPV4        (1)
872#define XEN_NETIF_GSO_TYPE_TCPV6        (2)
873
874/*
875 * This structure needs to fit within both xen_netif_tx_request_t and
876 * xen_netif_rx_response_t for compatibility.
877 */
878struct xen_netif_extra_info {
879	uint8_t type;
880	uint8_t flags;
881	union {
882		struct {
883			uint16_t size;
884			uint8_t type;
885			uint8_t pad;
886			uint16_t features;
887		} gso;
888		struct {
889			uint8_t addr[6];
890		} mcast;
891		struct {
892			uint8_t type;
893			uint8_t algorithm;
894			uint8_t value[4];
895		} hash;
896		struct {
897			uint16_t headroom;
898			uint16_t pad[2];
899		} xdp;
900		uint16_t pad[3];
901	} u;
902};
903
904struct xen_netif_tx_response {
905	uint16_t id;
906	int16_t status;
907};
908
909struct xen_netif_rx_request {
910	uint16_t id;		/* Echoed in response message.        */
911	uint16_t pad;
912	grant_ref_t gref;
913};
914
915/* Packet data has been validated against protocol checksum. */
916#define _XEN_NETRXF_data_validated (0)
917#define  XEN_NETRXF_data_validated (1U<<_XEN_NETRXF_data_validated)
918
919/* Protocol checksum field is blank in the packet (hardware offload)? */
920#define _XEN_NETRXF_csum_blank     (1)
921#define  XEN_NETRXF_csum_blank     (1U<<_XEN_NETRXF_csum_blank)
922
923/* Packet continues in the next request descriptor. */
924#define _XEN_NETRXF_more_data      (2)
925#define  XEN_NETRXF_more_data      (1U<<_XEN_NETRXF_more_data)
926
927/* Packet to be followed by extra descriptor(s). */
928#define _XEN_NETRXF_extra_info     (3)
929#define  XEN_NETRXF_extra_info     (1U<<_XEN_NETRXF_extra_info)
930
931/* Packet has GSO prefix. Deprecated but included for compatibility */
932#define _XEN_NETRXF_gso_prefix     (4)
933#define  XEN_NETRXF_gso_prefix     (1U<<_XEN_NETRXF_gso_prefix)
934
935struct xen_netif_rx_response {
936	uint16_t id;
937	uint16_t offset;
938	uint16_t flags;
939	int16_t status;
940};
941
942/*
943 * Generate xen_netif ring structures and types.
944 */
945
946DEFINE_RING_TYPES(xen_netif_tx, struct xen_netif_tx_request,
947		  struct xen_netif_tx_response);
948DEFINE_RING_TYPES(xen_netif_rx, struct xen_netif_rx_request,
949		  struct xen_netif_rx_response);
950
951#define XEN_NETIF_RSP_DROPPED         -2
952#define XEN_NETIF_RSP_ERROR           -1
953#define XEN_NETIF_RSP_OKAY             0
954/* No response: used for auxiliary requests (e.g., xen_netif_extra_info_t). */
955#define XEN_NETIF_RSP_NULL             1
956
957#endif