Loading...
1/*
2 *
3 * Intel Management Engine Interface (Intel MEI) Linux driver
4 * Copyright (c) 2003-2012, Intel Corporation.
5 *
6 * This program is free software; you can redistribute it and/or modify it
7 * under the terms and conditions of the GNU General Public License,
8 * version 2, as published by the Free Software Foundation.
9 *
10 * This program is distributed in the hope it will be useful, but WITHOUT
11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
13 * more details.
14 *
15 */
16#include <linux/module.h>
17#include <linux/moduleparam.h>
18#include <linux/kernel.h>
19#include <linux/device.h>
20#include <linux/slab.h>
21#include <linux/fs.h>
22#include <linux/errno.h>
23#include <linux/types.h>
24#include <linux/fcntl.h>
25#include <linux/poll.h>
26#include <linux/init.h>
27#include <linux/ioctl.h>
28#include <linux/cdev.h>
29#include <linux/sched/signal.h>
30#include <linux/uuid.h>
31#include <linux/compat.h>
32#include <linux/jiffies.h>
33#include <linux/interrupt.h>
34
35#include <linux/mei.h>
36
37#include "mei_dev.h"
38#include "client.h"
39
40/**
41 * mei_open - the open function
42 *
43 * @inode: pointer to inode structure
44 * @file: pointer to file structure
45 *
46 * Return: 0 on success, <0 on error
47 */
48static int mei_open(struct inode *inode, struct file *file)
49{
50 struct mei_device *dev;
51 struct mei_cl *cl;
52
53 int err;
54
55 dev = container_of(inode->i_cdev, struct mei_device, cdev);
56 if (!dev)
57 return -ENODEV;
58
59 mutex_lock(&dev->device_lock);
60
61 if (dev->dev_state != MEI_DEV_ENABLED) {
62 dev_dbg(dev->dev, "dev_state != MEI_ENABLED dev_state = %s\n",
63 mei_dev_state_str(dev->dev_state));
64 err = -ENODEV;
65 goto err_unlock;
66 }
67
68 cl = mei_cl_alloc_linked(dev);
69 if (IS_ERR(cl)) {
70 err = PTR_ERR(cl);
71 goto err_unlock;
72 }
73
74 cl->fp = file;
75 file->private_data = cl;
76
77 mutex_unlock(&dev->device_lock);
78
79 return nonseekable_open(inode, file);
80
81err_unlock:
82 mutex_unlock(&dev->device_lock);
83 return err;
84}
85
86/**
87 * mei_release - the release function
88 *
89 * @inode: pointer to inode structure
90 * @file: pointer to file structure
91 *
92 * Return: 0 on success, <0 on error
93 */
94static int mei_release(struct inode *inode, struct file *file)
95{
96 struct mei_cl *cl = file->private_data;
97 struct mei_device *dev;
98 int rets;
99
100 if (WARN_ON(!cl || !cl->dev))
101 return -ENODEV;
102
103 dev = cl->dev;
104
105 mutex_lock(&dev->device_lock);
106
107 rets = mei_cl_disconnect(cl);
108
109 mei_cl_flush_queues(cl, file);
110 cl_dbg(dev, cl, "removing\n");
111
112 mei_cl_unlink(cl);
113
114 file->private_data = NULL;
115
116 kfree(cl);
117
118 mutex_unlock(&dev->device_lock);
119 return rets;
120}
121
122
123/**
124 * mei_read - the read function.
125 *
126 * @file: pointer to file structure
127 * @ubuf: pointer to user buffer
128 * @length: buffer length
129 * @offset: data offset in buffer
130 *
131 * Return: >=0 data length on success , <0 on error
132 */
133static ssize_t mei_read(struct file *file, char __user *ubuf,
134 size_t length, loff_t *offset)
135{
136 struct mei_cl *cl = file->private_data;
137 struct mei_device *dev;
138 struct mei_cl_cb *cb = NULL;
139 bool nonblock = !!(file->f_flags & O_NONBLOCK);
140 int rets;
141
142 if (WARN_ON(!cl || !cl->dev))
143 return -ENODEV;
144
145 dev = cl->dev;
146
147
148 mutex_lock(&dev->device_lock);
149 if (dev->dev_state != MEI_DEV_ENABLED) {
150 rets = -ENODEV;
151 goto out;
152 }
153
154 if (length == 0) {
155 rets = 0;
156 goto out;
157 }
158
159 if (ubuf == NULL) {
160 rets = -EMSGSIZE;
161 goto out;
162 }
163
164 cb = mei_cl_read_cb(cl, file);
165 if (cb)
166 goto copy_buffer;
167
168 if (*offset > 0)
169 *offset = 0;
170
171 rets = mei_cl_read_start(cl, length, file);
172 if (rets && rets != -EBUSY) {
173 cl_dbg(dev, cl, "mei start read failure status = %d\n", rets);
174 goto out;
175 }
176
177 if (nonblock) {
178 rets = -EAGAIN;
179 goto out;
180 }
181
182 mutex_unlock(&dev->device_lock);
183 if (wait_event_interruptible(cl->rx_wait,
184 !list_empty(&cl->rd_completed) ||
185 !mei_cl_is_connected(cl))) {
186 if (signal_pending(current))
187 return -EINTR;
188 return -ERESTARTSYS;
189 }
190 mutex_lock(&dev->device_lock);
191
192 if (!mei_cl_is_connected(cl)) {
193 rets = -ENODEV;
194 goto out;
195 }
196
197 cb = mei_cl_read_cb(cl, file);
198 if (!cb) {
199 rets = 0;
200 goto out;
201 }
202
203copy_buffer:
204 /* now copy the data to user space */
205 if (cb->status) {
206 rets = cb->status;
207 cl_dbg(dev, cl, "read operation failed %d\n", rets);
208 goto free;
209 }
210
211 cl_dbg(dev, cl, "buf.size = %zu buf.idx = %zu offset = %lld\n",
212 cb->buf.size, cb->buf_idx, *offset);
213 if (*offset >= cb->buf_idx) {
214 rets = 0;
215 goto free;
216 }
217
218 /* length is being truncated to PAGE_SIZE,
219 * however buf_idx may point beyond that */
220 length = min_t(size_t, length, cb->buf_idx - *offset);
221
222 if (copy_to_user(ubuf, cb->buf.data + *offset, length)) {
223 dev_dbg(dev->dev, "failed to copy data to userland\n");
224 rets = -EFAULT;
225 goto free;
226 }
227
228 rets = length;
229 *offset += length;
230 /* not all data was read, keep the cb */
231 if (*offset < cb->buf_idx)
232 goto out;
233
234free:
235 mei_io_cb_free(cb);
236 *offset = 0;
237
238out:
239 cl_dbg(dev, cl, "end mei read rets = %d\n", rets);
240 mutex_unlock(&dev->device_lock);
241 return rets;
242}
243/**
244 * mei_write - the write function.
245 *
246 * @file: pointer to file structure
247 * @ubuf: pointer to user buffer
248 * @length: buffer length
249 * @offset: data offset in buffer
250 *
251 * Return: >=0 data length on success , <0 on error
252 */
253static ssize_t mei_write(struct file *file, const char __user *ubuf,
254 size_t length, loff_t *offset)
255{
256 struct mei_cl *cl = file->private_data;
257 struct mei_cl_cb *cb;
258 struct mei_device *dev;
259 int rets;
260
261 if (WARN_ON(!cl || !cl->dev))
262 return -ENODEV;
263
264 dev = cl->dev;
265
266 mutex_lock(&dev->device_lock);
267
268 if (dev->dev_state != MEI_DEV_ENABLED) {
269 rets = -ENODEV;
270 goto out;
271 }
272
273 if (!mei_cl_is_connected(cl)) {
274 cl_err(dev, cl, "is not connected");
275 rets = -ENODEV;
276 goto out;
277 }
278
279 if (!mei_me_cl_is_active(cl->me_cl)) {
280 rets = -ENOTTY;
281 goto out;
282 }
283
284 if (length > mei_cl_mtu(cl)) {
285 rets = -EFBIG;
286 goto out;
287 }
288
289 if (length == 0) {
290 rets = 0;
291 goto out;
292 }
293
294 while (cl->tx_cb_queued >= dev->tx_queue_limit) {
295 if (file->f_flags & O_NONBLOCK) {
296 rets = -EAGAIN;
297 goto out;
298 }
299 mutex_unlock(&dev->device_lock);
300 rets = wait_event_interruptible(cl->tx_wait,
301 cl->writing_state == MEI_WRITE_COMPLETE ||
302 (!mei_cl_is_connected(cl)));
303 mutex_lock(&dev->device_lock);
304 if (rets) {
305 if (signal_pending(current))
306 rets = -EINTR;
307 goto out;
308 }
309 if (!mei_cl_is_connected(cl)) {
310 rets = -ENODEV;
311 goto out;
312 }
313 }
314
315 *offset = 0;
316 cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, file);
317 if (!cb) {
318 rets = -ENOMEM;
319 goto out;
320 }
321
322 rets = copy_from_user(cb->buf.data, ubuf, length);
323 if (rets) {
324 dev_dbg(dev->dev, "failed to copy data from userland\n");
325 rets = -EFAULT;
326 mei_io_cb_free(cb);
327 goto out;
328 }
329
330 rets = mei_cl_write(cl, cb);
331out:
332 mutex_unlock(&dev->device_lock);
333 return rets;
334}
335
336/**
337 * mei_ioctl_connect_client - the connect to fw client IOCTL function
338 *
339 * @file: private data of the file object
340 * @data: IOCTL connect data, input and output parameters
341 *
342 * Locking: called under "dev->device_lock" lock
343 *
344 * Return: 0 on success, <0 on failure.
345 */
346static int mei_ioctl_connect_client(struct file *file,
347 struct mei_connect_client_data *data)
348{
349 struct mei_device *dev;
350 struct mei_client *client;
351 struct mei_me_client *me_cl;
352 struct mei_cl *cl;
353 int rets;
354
355 cl = file->private_data;
356 dev = cl->dev;
357
358 if (dev->dev_state != MEI_DEV_ENABLED)
359 return -ENODEV;
360
361 if (cl->state != MEI_FILE_INITIALIZING &&
362 cl->state != MEI_FILE_DISCONNECTED)
363 return -EBUSY;
364
365 /* find ME client we're trying to connect to */
366 me_cl = mei_me_cl_by_uuid(dev, &data->in_client_uuid);
367 if (!me_cl) {
368 dev_dbg(dev->dev, "Cannot connect to FW Client UUID = %pUl\n",
369 &data->in_client_uuid);
370 rets = -ENOTTY;
371 goto end;
372 }
373
374 if (me_cl->props.fixed_address) {
375 bool forbidden = dev->override_fixed_address ?
376 !dev->allow_fixed_address : !dev->hbm_f_fa_supported;
377 if (forbidden) {
378 dev_dbg(dev->dev, "Connection forbidden to FW Client UUID = %pUl\n",
379 &data->in_client_uuid);
380 rets = -ENOTTY;
381 goto end;
382 }
383 }
384
385 dev_dbg(dev->dev, "Connect to FW Client ID = %d\n",
386 me_cl->client_id);
387 dev_dbg(dev->dev, "FW Client - Protocol Version = %d\n",
388 me_cl->props.protocol_version);
389 dev_dbg(dev->dev, "FW Client - Max Msg Len = %d\n",
390 me_cl->props.max_msg_length);
391
392 /* prepare the output buffer */
393 client = &data->out_client_properties;
394 client->max_msg_length = me_cl->props.max_msg_length;
395 client->protocol_version = me_cl->props.protocol_version;
396 dev_dbg(dev->dev, "Can connect?\n");
397
398 rets = mei_cl_connect(cl, me_cl, file);
399
400end:
401 mei_me_cl_put(me_cl);
402 return rets;
403}
404
405/**
406 * mei_ioctl_client_notify_request -
407 * propagate event notification request to client
408 *
409 * @file: pointer to file structure
410 * @request: 0 - disable, 1 - enable
411 *
412 * Return: 0 on success , <0 on error
413 */
414static int mei_ioctl_client_notify_request(const struct file *file, u32 request)
415{
416 struct mei_cl *cl = file->private_data;
417
418 if (request != MEI_HBM_NOTIFICATION_START &&
419 request != MEI_HBM_NOTIFICATION_STOP)
420 return -EINVAL;
421
422 return mei_cl_notify_request(cl, file, (u8)request);
423}
424
425/**
426 * mei_ioctl_client_notify_get - wait for notification request
427 *
428 * @file: pointer to file structure
429 * @notify_get: 0 - disable, 1 - enable
430 *
431 * Return: 0 on success , <0 on error
432 */
433static int mei_ioctl_client_notify_get(const struct file *file, u32 *notify_get)
434{
435 struct mei_cl *cl = file->private_data;
436 bool notify_ev;
437 bool block = (file->f_flags & O_NONBLOCK) == 0;
438 int rets;
439
440 rets = mei_cl_notify_get(cl, block, ¬ify_ev);
441 if (rets)
442 return rets;
443
444 *notify_get = notify_ev ? 1 : 0;
445 return 0;
446}
447
448/**
449 * mei_ioctl - the IOCTL function
450 *
451 * @file: pointer to file structure
452 * @cmd: ioctl command
453 * @data: pointer to mei message structure
454 *
455 * Return: 0 on success , <0 on error
456 */
457static long mei_ioctl(struct file *file, unsigned int cmd, unsigned long data)
458{
459 struct mei_device *dev;
460 struct mei_cl *cl = file->private_data;
461 struct mei_connect_client_data connect_data;
462 u32 notify_get, notify_req;
463 int rets;
464
465
466 if (WARN_ON(!cl || !cl->dev))
467 return -ENODEV;
468
469 dev = cl->dev;
470
471 dev_dbg(dev->dev, "IOCTL cmd = 0x%x", cmd);
472
473 mutex_lock(&dev->device_lock);
474 if (dev->dev_state != MEI_DEV_ENABLED) {
475 rets = -ENODEV;
476 goto out;
477 }
478
479 switch (cmd) {
480 case IOCTL_MEI_CONNECT_CLIENT:
481 dev_dbg(dev->dev, ": IOCTL_MEI_CONNECT_CLIENT.\n");
482 if (copy_from_user(&connect_data, (char __user *)data,
483 sizeof(struct mei_connect_client_data))) {
484 dev_dbg(dev->dev, "failed to copy data from userland\n");
485 rets = -EFAULT;
486 goto out;
487 }
488
489 rets = mei_ioctl_connect_client(file, &connect_data);
490 if (rets)
491 goto out;
492
493 /* if all is ok, copying the data back to user. */
494 if (copy_to_user((char __user *)data, &connect_data,
495 sizeof(struct mei_connect_client_data))) {
496 dev_dbg(dev->dev, "failed to copy data to userland\n");
497 rets = -EFAULT;
498 goto out;
499 }
500
501 break;
502
503 case IOCTL_MEI_NOTIFY_SET:
504 dev_dbg(dev->dev, ": IOCTL_MEI_NOTIFY_SET.\n");
505 if (copy_from_user(¬ify_req,
506 (char __user *)data, sizeof(notify_req))) {
507 dev_dbg(dev->dev, "failed to copy data from userland\n");
508 rets = -EFAULT;
509 goto out;
510 }
511 rets = mei_ioctl_client_notify_request(file, notify_req);
512 break;
513
514 case IOCTL_MEI_NOTIFY_GET:
515 dev_dbg(dev->dev, ": IOCTL_MEI_NOTIFY_GET.\n");
516 rets = mei_ioctl_client_notify_get(file, ¬ify_get);
517 if (rets)
518 goto out;
519
520 dev_dbg(dev->dev, "copy connect data to user\n");
521 if (copy_to_user((char __user *)data,
522 ¬ify_get, sizeof(notify_get))) {
523 dev_dbg(dev->dev, "failed to copy data to userland\n");
524 rets = -EFAULT;
525 goto out;
526
527 }
528 break;
529
530 default:
531 rets = -ENOIOCTLCMD;
532 }
533
534out:
535 mutex_unlock(&dev->device_lock);
536 return rets;
537}
538
539/**
540 * mei_compat_ioctl - the compat IOCTL function
541 *
542 * @file: pointer to file structure
543 * @cmd: ioctl command
544 * @data: pointer to mei message structure
545 *
546 * Return: 0 on success , <0 on error
547 */
548#ifdef CONFIG_COMPAT
549static long mei_compat_ioctl(struct file *file,
550 unsigned int cmd, unsigned long data)
551{
552 return mei_ioctl(file, cmd, (unsigned long)compat_ptr(data));
553}
554#endif
555
556
557/**
558 * mei_poll - the poll function
559 *
560 * @file: pointer to file structure
561 * @wait: pointer to poll_table structure
562 *
563 * Return: poll mask
564 */
565static __poll_t mei_poll(struct file *file, poll_table *wait)
566{
567 __poll_t req_events = poll_requested_events(wait);
568 struct mei_cl *cl = file->private_data;
569 struct mei_device *dev;
570 __poll_t mask = 0;
571 bool notify_en;
572
573 if (WARN_ON(!cl || !cl->dev))
574 return EPOLLERR;
575
576 dev = cl->dev;
577
578 mutex_lock(&dev->device_lock);
579
580 notify_en = cl->notify_en && (req_events & EPOLLPRI);
581
582 if (dev->dev_state != MEI_DEV_ENABLED ||
583 !mei_cl_is_connected(cl)) {
584 mask = EPOLLERR;
585 goto out;
586 }
587
588 if (notify_en) {
589 poll_wait(file, &cl->ev_wait, wait);
590 if (cl->notify_ev)
591 mask |= EPOLLPRI;
592 }
593
594 if (req_events & (EPOLLIN | EPOLLRDNORM)) {
595 poll_wait(file, &cl->rx_wait, wait);
596
597 if (!list_empty(&cl->rd_completed))
598 mask |= EPOLLIN | EPOLLRDNORM;
599 else
600 mei_cl_read_start(cl, mei_cl_mtu(cl), file);
601 }
602
603 if (req_events & (POLLOUT | POLLWRNORM)) {
604 poll_wait(file, &cl->tx_wait, wait);
605 if (cl->tx_cb_queued < dev->tx_queue_limit)
606 mask |= POLLOUT | POLLWRNORM;
607 }
608
609out:
610 mutex_unlock(&dev->device_lock);
611 return mask;
612}
613
614/**
615 * mei_cl_is_write_queued - check if the client has pending writes.
616 *
617 * @cl: writing host client
618 *
619 * Return: true if client is writing, false otherwise.
620 */
621static bool mei_cl_is_write_queued(struct mei_cl *cl)
622{
623 struct mei_device *dev = cl->dev;
624 struct mei_cl_cb *cb;
625
626 list_for_each_entry(cb, &dev->write_list, list)
627 if (cb->cl == cl)
628 return true;
629 list_for_each_entry(cb, &dev->write_waiting_list, list)
630 if (cb->cl == cl)
631 return true;
632 return false;
633}
634
635/**
636 * mei_fsync - the fsync handler
637 *
638 * @fp: pointer to file structure
639 * @start: unused
640 * @end: unused
641 * @datasync: unused
642 *
643 * Return: 0 on success, -ENODEV if client is not connected
644 */
645static int mei_fsync(struct file *fp, loff_t start, loff_t end, int datasync)
646{
647 struct mei_cl *cl = fp->private_data;
648 struct mei_device *dev;
649 int rets;
650
651 if (WARN_ON(!cl || !cl->dev))
652 return -ENODEV;
653
654 dev = cl->dev;
655
656 mutex_lock(&dev->device_lock);
657
658 if (dev->dev_state != MEI_DEV_ENABLED || !mei_cl_is_connected(cl)) {
659 rets = -ENODEV;
660 goto out;
661 }
662
663 while (mei_cl_is_write_queued(cl)) {
664 mutex_unlock(&dev->device_lock);
665 rets = wait_event_interruptible(cl->tx_wait,
666 cl->writing_state == MEI_WRITE_COMPLETE ||
667 !mei_cl_is_connected(cl));
668 mutex_lock(&dev->device_lock);
669 if (rets) {
670 if (signal_pending(current))
671 rets = -EINTR;
672 goto out;
673 }
674 if (!mei_cl_is_connected(cl)) {
675 rets = -ENODEV;
676 goto out;
677 }
678 }
679 rets = 0;
680out:
681 mutex_unlock(&dev->device_lock);
682 return rets;
683}
684
685/**
686 * mei_fasync - asynchronous io support
687 *
688 * @fd: file descriptor
689 * @file: pointer to file structure
690 * @band: band bitmap
691 *
692 * Return: negative on error,
693 * 0 if it did no changes,
694 * and positive a process was added or deleted
695 */
696static int mei_fasync(int fd, struct file *file, int band)
697{
698
699 struct mei_cl *cl = file->private_data;
700
701 if (!mei_cl_is_connected(cl))
702 return -ENODEV;
703
704 return fasync_helper(fd, file, band, &cl->ev_async);
705}
706
707/**
708 * fw_status_show - mei device fw_status attribute show method
709 *
710 * @device: device pointer
711 * @attr: attribute pointer
712 * @buf: char out buffer
713 *
714 * Return: number of the bytes printed into buf or error
715 */
716static ssize_t fw_status_show(struct device *device,
717 struct device_attribute *attr, char *buf)
718{
719 struct mei_device *dev = dev_get_drvdata(device);
720 struct mei_fw_status fw_status;
721 int err, i;
722 ssize_t cnt = 0;
723
724 mutex_lock(&dev->device_lock);
725 err = mei_fw_status(dev, &fw_status);
726 mutex_unlock(&dev->device_lock);
727 if (err) {
728 dev_err(device, "read fw_status error = %d\n", err);
729 return err;
730 }
731
732 for (i = 0; i < fw_status.count; i++)
733 cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "%08X\n",
734 fw_status.status[i]);
735 return cnt;
736}
737static DEVICE_ATTR_RO(fw_status);
738
739/**
740 * hbm_ver_show - display HBM protocol version negotiated with FW
741 *
742 * @device: device pointer
743 * @attr: attribute pointer
744 * @buf: char out buffer
745 *
746 * Return: number of the bytes printed into buf or error
747 */
748static ssize_t hbm_ver_show(struct device *device,
749 struct device_attribute *attr, char *buf)
750{
751 struct mei_device *dev = dev_get_drvdata(device);
752 struct hbm_version ver;
753
754 mutex_lock(&dev->device_lock);
755 ver = dev->version;
756 mutex_unlock(&dev->device_lock);
757
758 return sprintf(buf, "%u.%u\n", ver.major_version, ver.minor_version);
759}
760static DEVICE_ATTR_RO(hbm_ver);
761
762/**
763 * hbm_ver_drv_show - display HBM protocol version advertised by driver
764 *
765 * @device: device pointer
766 * @attr: attribute pointer
767 * @buf: char out buffer
768 *
769 * Return: number of the bytes printed into buf or error
770 */
771static ssize_t hbm_ver_drv_show(struct device *device,
772 struct device_attribute *attr, char *buf)
773{
774 return sprintf(buf, "%u.%u\n", HBM_MAJOR_VERSION, HBM_MINOR_VERSION);
775}
776static DEVICE_ATTR_RO(hbm_ver_drv);
777
778static ssize_t tx_queue_limit_show(struct device *device,
779 struct device_attribute *attr, char *buf)
780{
781 struct mei_device *dev = dev_get_drvdata(device);
782 u8 size = 0;
783
784 mutex_lock(&dev->device_lock);
785 size = dev->tx_queue_limit;
786 mutex_unlock(&dev->device_lock);
787
788 return snprintf(buf, PAGE_SIZE, "%u\n", size);
789}
790
791static ssize_t tx_queue_limit_store(struct device *device,
792 struct device_attribute *attr,
793 const char *buf, size_t count)
794{
795 struct mei_device *dev = dev_get_drvdata(device);
796 u8 limit;
797 unsigned int inp;
798 int err;
799
800 err = kstrtouint(buf, 10, &inp);
801 if (err)
802 return err;
803 if (inp > MEI_TX_QUEUE_LIMIT_MAX || inp < MEI_TX_QUEUE_LIMIT_MIN)
804 return -EINVAL;
805 limit = inp;
806
807 mutex_lock(&dev->device_lock);
808 dev->tx_queue_limit = limit;
809 mutex_unlock(&dev->device_lock);
810
811 return count;
812}
813static DEVICE_ATTR_RW(tx_queue_limit);
814
815static struct attribute *mei_attrs[] = {
816 &dev_attr_fw_status.attr,
817 &dev_attr_hbm_ver.attr,
818 &dev_attr_hbm_ver_drv.attr,
819 &dev_attr_tx_queue_limit.attr,
820 NULL
821};
822ATTRIBUTE_GROUPS(mei);
823
824/*
825 * file operations structure will be used for mei char device.
826 */
827static const struct file_operations mei_fops = {
828 .owner = THIS_MODULE,
829 .read = mei_read,
830 .unlocked_ioctl = mei_ioctl,
831#ifdef CONFIG_COMPAT
832 .compat_ioctl = mei_compat_ioctl,
833#endif
834 .open = mei_open,
835 .release = mei_release,
836 .write = mei_write,
837 .poll = mei_poll,
838 .fsync = mei_fsync,
839 .fasync = mei_fasync,
840 .llseek = no_llseek
841};
842
843static struct class *mei_class;
844static dev_t mei_devt;
845#define MEI_MAX_DEVS MINORMASK
846static DEFINE_MUTEX(mei_minor_lock);
847static DEFINE_IDR(mei_idr);
848
849/**
850 * mei_minor_get - obtain next free device minor number
851 *
852 * @dev: device pointer
853 *
854 * Return: allocated minor, or -ENOSPC if no free minor left
855 */
856static int mei_minor_get(struct mei_device *dev)
857{
858 int ret;
859
860 mutex_lock(&mei_minor_lock);
861 ret = idr_alloc(&mei_idr, dev, 0, MEI_MAX_DEVS, GFP_KERNEL);
862 if (ret >= 0)
863 dev->minor = ret;
864 else if (ret == -ENOSPC)
865 dev_err(dev->dev, "too many mei devices\n");
866
867 mutex_unlock(&mei_minor_lock);
868 return ret;
869}
870
871/**
872 * mei_minor_free - mark device minor number as free
873 *
874 * @dev: device pointer
875 */
876static void mei_minor_free(struct mei_device *dev)
877{
878 mutex_lock(&mei_minor_lock);
879 idr_remove(&mei_idr, dev->minor);
880 mutex_unlock(&mei_minor_lock);
881}
882
883int mei_register(struct mei_device *dev, struct device *parent)
884{
885 struct device *clsdev; /* class device */
886 int ret, devno;
887
888 ret = mei_minor_get(dev);
889 if (ret < 0)
890 return ret;
891
892 /* Fill in the data structures */
893 devno = MKDEV(MAJOR(mei_devt), dev->minor);
894 cdev_init(&dev->cdev, &mei_fops);
895 dev->cdev.owner = parent->driver->owner;
896
897 /* Add the device */
898 ret = cdev_add(&dev->cdev, devno, 1);
899 if (ret) {
900 dev_err(parent, "unable to add device %d:%d\n",
901 MAJOR(mei_devt), dev->minor);
902 goto err_dev_add;
903 }
904
905 clsdev = device_create_with_groups(mei_class, parent, devno,
906 dev, mei_groups,
907 "mei%d", dev->minor);
908
909 if (IS_ERR(clsdev)) {
910 dev_err(parent, "unable to create device %d:%d\n",
911 MAJOR(mei_devt), dev->minor);
912 ret = PTR_ERR(clsdev);
913 goto err_dev_create;
914 }
915
916 ret = mei_dbgfs_register(dev, dev_name(clsdev));
917 if (ret) {
918 dev_err(clsdev, "cannot register debugfs ret = %d\n", ret);
919 goto err_dev_dbgfs;
920 }
921
922 return 0;
923
924err_dev_dbgfs:
925 device_destroy(mei_class, devno);
926err_dev_create:
927 cdev_del(&dev->cdev);
928err_dev_add:
929 mei_minor_free(dev);
930 return ret;
931}
932EXPORT_SYMBOL_GPL(mei_register);
933
934void mei_deregister(struct mei_device *dev)
935{
936 int devno;
937
938 devno = dev->cdev.dev;
939 cdev_del(&dev->cdev);
940
941 mei_dbgfs_deregister(dev);
942
943 device_destroy(mei_class, devno);
944
945 mei_minor_free(dev);
946}
947EXPORT_SYMBOL_GPL(mei_deregister);
948
949static int __init mei_init(void)
950{
951 int ret;
952
953 mei_class = class_create(THIS_MODULE, "mei");
954 if (IS_ERR(mei_class)) {
955 pr_err("couldn't create class\n");
956 ret = PTR_ERR(mei_class);
957 goto err;
958 }
959
960 ret = alloc_chrdev_region(&mei_devt, 0, MEI_MAX_DEVS, "mei");
961 if (ret < 0) {
962 pr_err("unable to allocate char dev region\n");
963 goto err_class;
964 }
965
966 ret = mei_cl_bus_init();
967 if (ret < 0) {
968 pr_err("unable to initialize bus\n");
969 goto err_chrdev;
970 }
971
972 return 0;
973
974err_chrdev:
975 unregister_chrdev_region(mei_devt, MEI_MAX_DEVS);
976err_class:
977 class_destroy(mei_class);
978err:
979 return ret;
980}
981
982static void __exit mei_exit(void)
983{
984 unregister_chrdev_region(mei_devt, MEI_MAX_DEVS);
985 class_destroy(mei_class);
986 mei_cl_bus_exit();
987}
988
989module_init(mei_init);
990module_exit(mei_exit);
991
992MODULE_AUTHOR("Intel Corporation");
993MODULE_DESCRIPTION("Intel(R) Management Engine Interface");
994MODULE_LICENSE("GPL v2");
995
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (c) 2003-2022, Intel Corporation. All rights reserved.
4 * Intel Management Engine Interface (Intel MEI) Linux driver
5 */
6
7#include <linux/module.h>
8#include <linux/moduleparam.h>
9#include <linux/kernel.h>
10#include <linux/device.h>
11#include <linux/slab.h>
12#include <linux/fs.h>
13#include <linux/errno.h>
14#include <linux/types.h>
15#include <linux/fcntl.h>
16#include <linux/poll.h>
17#include <linux/init.h>
18#include <linux/ioctl.h>
19#include <linux/cdev.h>
20#include <linux/sched/signal.h>
21#include <linux/compat.h>
22#include <linux/jiffies.h>
23#include <linux/interrupt.h>
24
25#include <linux/mei.h>
26
27#include "mei_dev.h"
28#include "client.h"
29
30static const struct class mei_class = {
31 .name = "mei",
32};
33
34static dev_t mei_devt;
35#define MEI_MAX_DEVS MINORMASK
36static DEFINE_MUTEX(mei_minor_lock);
37static DEFINE_IDR(mei_idr);
38
39/**
40 * mei_open - the open function
41 *
42 * @inode: pointer to inode structure
43 * @file: pointer to file structure
44 *
45 * Return: 0 on success, <0 on error
46 */
47static int mei_open(struct inode *inode, struct file *file)
48{
49 struct mei_device *dev;
50 struct mei_cl *cl;
51
52 int err;
53
54 dev = container_of(inode->i_cdev, struct mei_device, cdev);
55
56 mutex_lock(&dev->device_lock);
57
58 if (dev->dev_state != MEI_DEV_ENABLED) {
59 dev_dbg(dev->dev, "dev_state != MEI_ENABLED dev_state = %s\n",
60 mei_dev_state_str(dev->dev_state));
61 err = -ENODEV;
62 goto err_unlock;
63 }
64
65 cl = mei_cl_alloc_linked(dev);
66 if (IS_ERR(cl)) {
67 err = PTR_ERR(cl);
68 goto err_unlock;
69 }
70
71 cl->fp = file;
72 file->private_data = cl;
73
74 mutex_unlock(&dev->device_lock);
75
76 return nonseekable_open(inode, file);
77
78err_unlock:
79 mutex_unlock(&dev->device_lock);
80 return err;
81}
82
83/**
84 * mei_cl_vtag_remove_by_fp - remove vtag that corresponds to fp from list
85 *
86 * @cl: host client
87 * @fp: pointer to file structure
88 *
89 */
90static void mei_cl_vtag_remove_by_fp(const struct mei_cl *cl,
91 const struct file *fp)
92{
93 struct mei_cl_vtag *vtag_l, *next;
94
95 list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) {
96 if (vtag_l->fp == fp) {
97 list_del(&vtag_l->list);
98 kfree(vtag_l);
99 return;
100 }
101 }
102}
103
104/**
105 * mei_release - the release function
106 *
107 * @inode: pointer to inode structure
108 * @file: pointer to file structure
109 *
110 * Return: 0 on success, <0 on error
111 */
112static int mei_release(struct inode *inode, struct file *file)
113{
114 struct mei_cl *cl = file->private_data;
115 struct mei_device *dev;
116 int rets;
117
118 if (WARN_ON(!cl || !cl->dev))
119 return -ENODEV;
120
121 dev = cl->dev;
122
123 mutex_lock(&dev->device_lock);
124
125 mei_cl_vtag_remove_by_fp(cl, file);
126
127 if (!list_empty(&cl->vtag_map)) {
128 cl_dbg(dev, cl, "not the last vtag\n");
129 mei_cl_flush_queues(cl, file);
130 rets = 0;
131 goto out;
132 }
133
134 rets = mei_cl_disconnect(cl);
135 /*
136 * Check again: This is necessary since disconnect releases the lock
137 * and another client can connect in the meantime.
138 */
139 if (!list_empty(&cl->vtag_map)) {
140 cl_dbg(dev, cl, "not the last vtag after disconnect\n");
141 mei_cl_flush_queues(cl, file);
142 goto out;
143 }
144
145 mei_cl_flush_queues(cl, NULL);
146 cl_dbg(dev, cl, "removing\n");
147
148 mei_cl_unlink(cl);
149 kfree(cl);
150
151out:
152 file->private_data = NULL;
153
154 mutex_unlock(&dev->device_lock);
155 return rets;
156}
157
158
159/**
160 * mei_read - the read function.
161 *
162 * @file: pointer to file structure
163 * @ubuf: pointer to user buffer
164 * @length: buffer length
165 * @offset: data offset in buffer
166 *
167 * Return: >=0 data length on success , <0 on error
168 */
169static ssize_t mei_read(struct file *file, char __user *ubuf,
170 size_t length, loff_t *offset)
171{
172 struct mei_cl *cl = file->private_data;
173 struct mei_device *dev;
174 struct mei_cl_cb *cb = NULL;
175 bool nonblock = !!(file->f_flags & O_NONBLOCK);
176 ssize_t rets;
177
178 if (WARN_ON(!cl || !cl->dev))
179 return -ENODEV;
180
181 dev = cl->dev;
182
183
184 mutex_lock(&dev->device_lock);
185 if (dev->dev_state != MEI_DEV_ENABLED) {
186 rets = -ENODEV;
187 goto out;
188 }
189
190 if (length == 0) {
191 rets = 0;
192 goto out;
193 }
194
195 if (ubuf == NULL) {
196 rets = -EMSGSIZE;
197 goto out;
198 }
199
200 cb = mei_cl_read_cb(cl, file);
201 if (cb)
202 goto copy_buffer;
203
204 if (*offset > 0)
205 *offset = 0;
206
207 rets = mei_cl_read_start(cl, length, file);
208 if (rets && rets != -EBUSY) {
209 cl_dbg(dev, cl, "mei start read failure status = %zd\n", rets);
210 goto out;
211 }
212
213 if (nonblock) {
214 rets = -EAGAIN;
215 goto out;
216 }
217
218 mutex_unlock(&dev->device_lock);
219 if (wait_event_interruptible(cl->rx_wait,
220 mei_cl_read_cb(cl, file) ||
221 !mei_cl_is_connected(cl))) {
222 if (signal_pending(current))
223 return -EINTR;
224 return -ERESTARTSYS;
225 }
226 mutex_lock(&dev->device_lock);
227
228 if (!mei_cl_is_connected(cl)) {
229 rets = -ENODEV;
230 goto out;
231 }
232
233 cb = mei_cl_read_cb(cl, file);
234 if (!cb) {
235 rets = 0;
236 goto out;
237 }
238
239copy_buffer:
240 /* now copy the data to user space */
241 if (cb->status) {
242 rets = cb->status;
243 cl_dbg(dev, cl, "read operation failed %zd\n", rets);
244 goto free;
245 }
246
247 cl_dbg(dev, cl, "buf.size = %zu buf.idx = %zu offset = %lld\n",
248 cb->buf.size, cb->buf_idx, *offset);
249 if (*offset >= cb->buf_idx) {
250 rets = 0;
251 goto free;
252 }
253
254 /* length is being truncated to PAGE_SIZE,
255 * however buf_idx may point beyond that */
256 length = min_t(size_t, length, cb->buf_idx - *offset);
257
258 if (copy_to_user(ubuf, cb->buf.data + *offset, length)) {
259 dev_dbg(dev->dev, "failed to copy data to userland\n");
260 rets = -EFAULT;
261 goto free;
262 }
263
264 rets = length;
265 *offset += length;
266 /* not all data was read, keep the cb */
267 if (*offset < cb->buf_idx)
268 goto out;
269
270free:
271 mei_cl_del_rd_completed(cl, cb);
272 *offset = 0;
273
274out:
275 cl_dbg(dev, cl, "end mei read rets = %zd\n", rets);
276 mutex_unlock(&dev->device_lock);
277 return rets;
278}
279
280/**
281 * mei_cl_vtag_by_fp - obtain the vtag by file pointer
282 *
283 * @cl: host client
284 * @fp: pointer to file structure
285 *
286 * Return: vtag value on success, otherwise 0
287 */
288static u8 mei_cl_vtag_by_fp(const struct mei_cl *cl, const struct file *fp)
289{
290 struct mei_cl_vtag *cl_vtag;
291
292 if (!fp)
293 return 0;
294
295 list_for_each_entry(cl_vtag, &cl->vtag_map, list)
296 if (cl_vtag->fp == fp)
297 return cl_vtag->vtag;
298 return 0;
299}
300
301/**
302 * mei_write - the write function.
303 *
304 * @file: pointer to file structure
305 * @ubuf: pointer to user buffer
306 * @length: buffer length
307 * @offset: data offset in buffer
308 *
309 * Return: >=0 data length on success , <0 on error
310 */
311static ssize_t mei_write(struct file *file, const char __user *ubuf,
312 size_t length, loff_t *offset)
313{
314 struct mei_cl *cl = file->private_data;
315 struct mei_cl_cb *cb;
316 struct mei_device *dev;
317 ssize_t rets;
318
319 if (WARN_ON(!cl || !cl->dev))
320 return -ENODEV;
321
322 dev = cl->dev;
323
324 mutex_lock(&dev->device_lock);
325
326 if (dev->dev_state != MEI_DEV_ENABLED) {
327 rets = -ENODEV;
328 goto out;
329 }
330
331 if (!mei_cl_is_connected(cl)) {
332 cl_err(dev, cl, "is not connected");
333 rets = -ENODEV;
334 goto out;
335 }
336
337 if (!mei_me_cl_is_active(cl->me_cl)) {
338 rets = -ENOTTY;
339 goto out;
340 }
341
342 if (length > mei_cl_mtu(cl)) {
343 rets = -EFBIG;
344 goto out;
345 }
346
347 if (length == 0) {
348 rets = 0;
349 goto out;
350 }
351
352 while (cl->tx_cb_queued >= dev->tx_queue_limit) {
353 if (file->f_flags & O_NONBLOCK) {
354 rets = -EAGAIN;
355 goto out;
356 }
357 mutex_unlock(&dev->device_lock);
358 rets = wait_event_interruptible(cl->tx_wait,
359 cl->writing_state == MEI_WRITE_COMPLETE ||
360 (!mei_cl_is_connected(cl)));
361 mutex_lock(&dev->device_lock);
362 if (rets) {
363 if (signal_pending(current))
364 rets = -EINTR;
365 goto out;
366 }
367 if (!mei_cl_is_connected(cl)) {
368 rets = -ENODEV;
369 goto out;
370 }
371 }
372
373 cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, file);
374 if (!cb) {
375 rets = -ENOMEM;
376 goto out;
377 }
378 cb->vtag = mei_cl_vtag_by_fp(cl, file);
379
380 rets = copy_from_user(cb->buf.data, ubuf, length);
381 if (rets) {
382 dev_dbg(dev->dev, "failed to copy data from userland\n");
383 rets = -EFAULT;
384 mei_io_cb_free(cb);
385 goto out;
386 }
387
388 rets = mei_cl_write(cl, cb, MAX_SCHEDULE_TIMEOUT);
389out:
390 mutex_unlock(&dev->device_lock);
391 return rets;
392}
393
394/**
395 * mei_ioctl_connect_client - the connect to fw client IOCTL function
396 *
397 * @file: private data of the file object
398 * @in_client_uuid: requested UUID for connection
399 * @client: IOCTL connect data, output parameters
400 *
401 * Locking: called under "dev->device_lock" lock
402 *
403 * Return: 0 on success, <0 on failure.
404 */
405static int mei_ioctl_connect_client(struct file *file,
406 const uuid_le *in_client_uuid,
407 struct mei_client *client)
408{
409 struct mei_device *dev;
410 struct mei_me_client *me_cl;
411 struct mei_cl *cl;
412 int rets;
413
414 cl = file->private_data;
415 dev = cl->dev;
416
417 if (cl->state != MEI_FILE_INITIALIZING &&
418 cl->state != MEI_FILE_DISCONNECTED)
419 return -EBUSY;
420
421 /* find ME client we're trying to connect to */
422 me_cl = mei_me_cl_by_uuid(dev, in_client_uuid);
423 if (!me_cl) {
424 dev_dbg(dev->dev, "Cannot connect to FW Client UUID = %pUl\n",
425 in_client_uuid);
426 rets = -ENOTTY;
427 goto end;
428 }
429
430 if (me_cl->props.fixed_address) {
431 bool forbidden = dev->override_fixed_address ?
432 !dev->allow_fixed_address : !dev->hbm_f_fa_supported;
433 if (forbidden) {
434 dev_dbg(dev->dev, "Connection forbidden to FW Client UUID = %pUl\n",
435 in_client_uuid);
436 rets = -ENOTTY;
437 goto end;
438 }
439 }
440
441 dev_dbg(dev->dev, "Connect to FW Client ID = %d\n",
442 me_cl->client_id);
443 dev_dbg(dev->dev, "FW Client - Protocol Version = %d\n",
444 me_cl->props.protocol_version);
445 dev_dbg(dev->dev, "FW Client - Max Msg Len = %d\n",
446 me_cl->props.max_msg_length);
447
448 /* prepare the output buffer */
449 client->max_msg_length = me_cl->props.max_msg_length;
450 client->protocol_version = me_cl->props.protocol_version;
451 dev_dbg(dev->dev, "Can connect?\n");
452
453 rets = mei_cl_connect(cl, me_cl, file);
454
455end:
456 mei_me_cl_put(me_cl);
457 return rets;
458}
459
460/**
461 * mei_vt_support_check - check if client support vtags
462 *
463 * @dev: mei_device
464 * @uuid: client UUID
465 *
466 * Locking: called under "dev->device_lock" lock
467 *
468 * Return:
469 * 0 - supported
470 * -ENOTTY - no such client
471 * -EOPNOTSUPP - vtags are not supported by client
472 */
473static int mei_vt_support_check(struct mei_device *dev, const uuid_le *uuid)
474{
475 struct mei_me_client *me_cl;
476 int ret;
477
478 if (!dev->hbm_f_vt_supported)
479 return -EOPNOTSUPP;
480
481 me_cl = mei_me_cl_by_uuid(dev, uuid);
482 if (!me_cl) {
483 dev_dbg(dev->dev, "Cannot connect to FW Client UUID = %pUl\n",
484 uuid);
485 return -ENOTTY;
486 }
487 ret = me_cl->props.vt_supported ? 0 : -EOPNOTSUPP;
488 mei_me_cl_put(me_cl);
489
490 return ret;
491}
492
493/**
494 * mei_ioctl_connect_vtag - connect to fw client with vtag IOCTL function
495 *
496 * @file: private data of the file object
497 * @in_client_uuid: requested UUID for connection
498 * @client: IOCTL connect data, output parameters
499 * @vtag: vm tag
500 *
501 * Locking: called under "dev->device_lock" lock
502 *
503 * Return: 0 on success, <0 on failure.
504 */
505static int mei_ioctl_connect_vtag(struct file *file,
506 const uuid_le *in_client_uuid,
507 struct mei_client *client,
508 u8 vtag)
509{
510 struct mei_device *dev;
511 struct mei_cl *cl;
512 struct mei_cl *pos;
513 struct mei_cl_vtag *cl_vtag;
514
515 cl = file->private_data;
516 dev = cl->dev;
517
518 dev_dbg(dev->dev, "FW Client %pUl vtag %d\n", in_client_uuid, vtag);
519
520 switch (cl->state) {
521 case MEI_FILE_DISCONNECTED:
522 if (mei_cl_vtag_by_fp(cl, file) != vtag) {
523 dev_err(dev->dev, "reconnect with different vtag\n");
524 return -EINVAL;
525 }
526 break;
527 case MEI_FILE_INITIALIZING:
528 /* malicious connect from another thread may push vtag */
529 if (!IS_ERR(mei_cl_fp_by_vtag(cl, vtag))) {
530 dev_err(dev->dev, "vtag already filled\n");
531 return -EINVAL;
532 }
533
534 list_for_each_entry(pos, &dev->file_list, link) {
535 if (pos == cl)
536 continue;
537 if (!pos->me_cl)
538 continue;
539
540 /* only search for same UUID */
541 if (uuid_le_cmp(*mei_cl_uuid(pos), *in_client_uuid))
542 continue;
543
544 /* if tag already exist try another fp */
545 if (!IS_ERR(mei_cl_fp_by_vtag(pos, vtag)))
546 continue;
547
548 /* replace cl with acquired one */
549 dev_dbg(dev->dev, "replacing with existing cl\n");
550 mei_cl_unlink(cl);
551 kfree(cl);
552 file->private_data = pos;
553 cl = pos;
554 break;
555 }
556
557 cl_vtag = mei_cl_vtag_alloc(file, vtag);
558 if (IS_ERR(cl_vtag))
559 return -ENOMEM;
560
561 list_add_tail(&cl_vtag->list, &cl->vtag_map);
562 break;
563 default:
564 return -EBUSY;
565 }
566
567 while (cl->state != MEI_FILE_INITIALIZING &&
568 cl->state != MEI_FILE_DISCONNECTED &&
569 cl->state != MEI_FILE_CONNECTED) {
570 mutex_unlock(&dev->device_lock);
571 wait_event_timeout(cl->wait,
572 (cl->state == MEI_FILE_CONNECTED ||
573 cl->state == MEI_FILE_DISCONNECTED ||
574 cl->state == MEI_FILE_DISCONNECT_REQUIRED ||
575 cl->state == MEI_FILE_DISCONNECT_REPLY),
576 dev->timeouts.cl_connect);
577 mutex_lock(&dev->device_lock);
578 }
579
580 if (!mei_cl_is_connected(cl))
581 return mei_ioctl_connect_client(file, in_client_uuid, client);
582
583 client->max_msg_length = cl->me_cl->props.max_msg_length;
584 client->protocol_version = cl->me_cl->props.protocol_version;
585
586 return 0;
587}
588
589/**
590 * mei_ioctl_client_notify_request - propagate event notification
591 * request to client
592 *
593 * @file: pointer to file structure
594 * @request: 0 - disable, 1 - enable
595 *
596 * Return: 0 on success , <0 on error
597 */
598static int mei_ioctl_client_notify_request(const struct file *file, u32 request)
599{
600 struct mei_cl *cl = file->private_data;
601
602 if (request != MEI_HBM_NOTIFICATION_START &&
603 request != MEI_HBM_NOTIFICATION_STOP)
604 return -EINVAL;
605
606 return mei_cl_notify_request(cl, file, (u8)request);
607}
608
609/**
610 * mei_ioctl_client_notify_get - wait for notification request
611 *
612 * @file: pointer to file structure
613 * @notify_get: 0 - disable, 1 - enable
614 *
615 * Return: 0 on success , <0 on error
616 */
617static int mei_ioctl_client_notify_get(const struct file *file, u32 *notify_get)
618{
619 struct mei_cl *cl = file->private_data;
620 bool notify_ev;
621 bool block = (file->f_flags & O_NONBLOCK) == 0;
622 int rets;
623
624 rets = mei_cl_notify_get(cl, block, ¬ify_ev);
625 if (rets)
626 return rets;
627
628 *notify_get = notify_ev ? 1 : 0;
629 return 0;
630}
631
632/**
633 * mei_ioctl - the IOCTL function
634 *
635 * @file: pointer to file structure
636 * @cmd: ioctl command
637 * @data: pointer to mei message structure
638 *
639 * Return: 0 on success , <0 on error
640 */
641static long mei_ioctl(struct file *file, unsigned int cmd, unsigned long data)
642{
643 struct mei_device *dev;
644 struct mei_cl *cl = file->private_data;
645 struct mei_connect_client_data conn;
646 struct mei_connect_client_data_vtag conn_vtag;
647 const uuid_le *cl_uuid;
648 struct mei_client *props;
649 u8 vtag;
650 u32 notify_get, notify_req;
651 int rets;
652
653
654 if (WARN_ON(!cl || !cl->dev))
655 return -ENODEV;
656
657 dev = cl->dev;
658
659 dev_dbg(dev->dev, "IOCTL cmd = 0x%x", cmd);
660
661 mutex_lock(&dev->device_lock);
662 if (dev->dev_state != MEI_DEV_ENABLED) {
663 rets = -ENODEV;
664 goto out;
665 }
666
667 switch (cmd) {
668 case IOCTL_MEI_CONNECT_CLIENT:
669 dev_dbg(dev->dev, ": IOCTL_MEI_CONNECT_CLIENT.\n");
670 if (copy_from_user(&conn, (char __user *)data, sizeof(conn))) {
671 dev_dbg(dev->dev, "failed to copy data from userland\n");
672 rets = -EFAULT;
673 goto out;
674 }
675 cl_uuid = &conn.in_client_uuid;
676 props = &conn.out_client_properties;
677 vtag = 0;
678
679 rets = mei_vt_support_check(dev, cl_uuid);
680 if (rets == -ENOTTY)
681 goto out;
682 if (!rets)
683 rets = mei_ioctl_connect_vtag(file, cl_uuid, props,
684 vtag);
685 else
686 rets = mei_ioctl_connect_client(file, cl_uuid, props);
687 if (rets)
688 goto out;
689
690 /* if all is ok, copying the data back to user. */
691 if (copy_to_user((char __user *)data, &conn, sizeof(conn))) {
692 dev_dbg(dev->dev, "failed to copy data to userland\n");
693 rets = -EFAULT;
694 goto out;
695 }
696
697 break;
698
699 case IOCTL_MEI_CONNECT_CLIENT_VTAG:
700 dev_dbg(dev->dev, "IOCTL_MEI_CONNECT_CLIENT_VTAG\n");
701 if (copy_from_user(&conn_vtag, (char __user *)data,
702 sizeof(conn_vtag))) {
703 dev_dbg(dev->dev, "failed to copy data from userland\n");
704 rets = -EFAULT;
705 goto out;
706 }
707
708 cl_uuid = &conn_vtag.connect.in_client_uuid;
709 props = &conn_vtag.out_client_properties;
710 vtag = conn_vtag.connect.vtag;
711
712 rets = mei_vt_support_check(dev, cl_uuid);
713 if (rets == -EOPNOTSUPP)
714 dev_dbg(dev->dev, "FW Client %pUl does not support vtags\n",
715 cl_uuid);
716 if (rets)
717 goto out;
718
719 if (!vtag) {
720 dev_dbg(dev->dev, "vtag can't be zero\n");
721 rets = -EINVAL;
722 goto out;
723 }
724
725 rets = mei_ioctl_connect_vtag(file, cl_uuid, props, vtag);
726 if (rets)
727 goto out;
728
729 /* if all is ok, copying the data back to user. */
730 if (copy_to_user((char __user *)data, &conn_vtag,
731 sizeof(conn_vtag))) {
732 dev_dbg(dev->dev, "failed to copy data to userland\n");
733 rets = -EFAULT;
734 goto out;
735 }
736
737 break;
738
739 case IOCTL_MEI_NOTIFY_SET:
740 dev_dbg(dev->dev, ": IOCTL_MEI_NOTIFY_SET.\n");
741 if (copy_from_user(¬ify_req,
742 (char __user *)data, sizeof(notify_req))) {
743 dev_dbg(dev->dev, "failed to copy data from userland\n");
744 rets = -EFAULT;
745 goto out;
746 }
747 rets = mei_ioctl_client_notify_request(file, notify_req);
748 break;
749
750 case IOCTL_MEI_NOTIFY_GET:
751 dev_dbg(dev->dev, ": IOCTL_MEI_NOTIFY_GET.\n");
752 rets = mei_ioctl_client_notify_get(file, ¬ify_get);
753 if (rets)
754 goto out;
755
756 dev_dbg(dev->dev, "copy connect data to user\n");
757 if (copy_to_user((char __user *)data,
758 ¬ify_get, sizeof(notify_get))) {
759 dev_dbg(dev->dev, "failed to copy data to userland\n");
760 rets = -EFAULT;
761 goto out;
762
763 }
764 break;
765
766 default:
767 rets = -ENOIOCTLCMD;
768 }
769
770out:
771 mutex_unlock(&dev->device_lock);
772 return rets;
773}
774
775/**
776 * mei_poll - the poll function
777 *
778 * @file: pointer to file structure
779 * @wait: pointer to poll_table structure
780 *
781 * Return: poll mask
782 */
783static __poll_t mei_poll(struct file *file, poll_table *wait)
784{
785 __poll_t req_events = poll_requested_events(wait);
786 struct mei_cl *cl = file->private_data;
787 struct mei_device *dev;
788 __poll_t mask = 0;
789 bool notify_en;
790
791 if (WARN_ON(!cl || !cl->dev))
792 return EPOLLERR;
793
794 dev = cl->dev;
795
796 mutex_lock(&dev->device_lock);
797
798 notify_en = cl->notify_en && (req_events & EPOLLPRI);
799
800 if (dev->dev_state != MEI_DEV_ENABLED ||
801 !mei_cl_is_connected(cl)) {
802 mask = EPOLLERR;
803 goto out;
804 }
805
806 if (notify_en) {
807 poll_wait(file, &cl->ev_wait, wait);
808 if (cl->notify_ev)
809 mask |= EPOLLPRI;
810 }
811
812 if (req_events & (EPOLLIN | EPOLLRDNORM)) {
813 poll_wait(file, &cl->rx_wait, wait);
814
815 if (mei_cl_read_cb(cl, file))
816 mask |= EPOLLIN | EPOLLRDNORM;
817 else
818 mei_cl_read_start(cl, mei_cl_mtu(cl), file);
819 }
820
821 if (req_events & (EPOLLOUT | EPOLLWRNORM)) {
822 poll_wait(file, &cl->tx_wait, wait);
823 if (cl->tx_cb_queued < dev->tx_queue_limit)
824 mask |= EPOLLOUT | EPOLLWRNORM;
825 }
826
827out:
828 mutex_unlock(&dev->device_lock);
829 return mask;
830}
831
832/**
833 * mei_cl_is_write_queued - check if the client has pending writes.
834 *
835 * @cl: writing host client
836 *
837 * Return: true if client is writing, false otherwise.
838 */
839static bool mei_cl_is_write_queued(struct mei_cl *cl)
840{
841 struct mei_device *dev = cl->dev;
842 struct mei_cl_cb *cb;
843
844 list_for_each_entry(cb, &dev->write_list, list)
845 if (cb->cl == cl)
846 return true;
847 list_for_each_entry(cb, &dev->write_waiting_list, list)
848 if (cb->cl == cl)
849 return true;
850 return false;
851}
852
853/**
854 * mei_fsync - the fsync handler
855 *
856 * @fp: pointer to file structure
857 * @start: unused
858 * @end: unused
859 * @datasync: unused
860 *
861 * Return: 0 on success, -ENODEV if client is not connected
862 */
863static int mei_fsync(struct file *fp, loff_t start, loff_t end, int datasync)
864{
865 struct mei_cl *cl = fp->private_data;
866 struct mei_device *dev;
867 int rets;
868
869 if (WARN_ON(!cl || !cl->dev))
870 return -ENODEV;
871
872 dev = cl->dev;
873
874 mutex_lock(&dev->device_lock);
875
876 if (dev->dev_state != MEI_DEV_ENABLED || !mei_cl_is_connected(cl)) {
877 rets = -ENODEV;
878 goto out;
879 }
880
881 while (mei_cl_is_write_queued(cl)) {
882 mutex_unlock(&dev->device_lock);
883 rets = wait_event_interruptible(cl->tx_wait,
884 cl->writing_state == MEI_WRITE_COMPLETE ||
885 !mei_cl_is_connected(cl));
886 mutex_lock(&dev->device_lock);
887 if (rets) {
888 if (signal_pending(current))
889 rets = -EINTR;
890 goto out;
891 }
892 if (!mei_cl_is_connected(cl)) {
893 rets = -ENODEV;
894 goto out;
895 }
896 }
897 rets = 0;
898out:
899 mutex_unlock(&dev->device_lock);
900 return rets;
901}
902
903/**
904 * mei_fasync - asynchronous io support
905 *
906 * @fd: file descriptor
907 * @file: pointer to file structure
908 * @band: band bitmap
909 *
910 * Return: negative on error,
911 * 0 if it did no changes,
912 * and positive a process was added or deleted
913 */
914static int mei_fasync(int fd, struct file *file, int band)
915{
916
917 struct mei_cl *cl = file->private_data;
918
919 if (!mei_cl_is_connected(cl))
920 return -ENODEV;
921
922 return fasync_helper(fd, file, band, &cl->ev_async);
923}
924
925/**
926 * trc_show - mei device trc attribute show method
927 *
928 * @device: device pointer
929 * @attr: attribute pointer
930 * @buf: char out buffer
931 *
932 * Return: number of the bytes printed into buf or error
933 */
934static ssize_t trc_show(struct device *device,
935 struct device_attribute *attr, char *buf)
936{
937 struct mei_device *dev = dev_get_drvdata(device);
938 u32 trc;
939 int ret;
940
941 ret = mei_trc_status(dev, &trc);
942 if (ret)
943 return ret;
944 return sprintf(buf, "%08X\n", trc);
945}
946static DEVICE_ATTR_RO(trc);
947
948/**
949 * fw_status_show - mei device fw_status attribute show method
950 *
951 * @device: device pointer
952 * @attr: attribute pointer
953 * @buf: char out buffer
954 *
955 * Return: number of the bytes printed into buf or error
956 */
957static ssize_t fw_status_show(struct device *device,
958 struct device_attribute *attr, char *buf)
959{
960 struct mei_device *dev = dev_get_drvdata(device);
961 struct mei_fw_status fw_status;
962 int err, i;
963 ssize_t cnt = 0;
964
965 mutex_lock(&dev->device_lock);
966 err = mei_fw_status(dev, &fw_status);
967 mutex_unlock(&dev->device_lock);
968 if (err) {
969 dev_err(device, "read fw_status error = %d\n", err);
970 return err;
971 }
972
973 for (i = 0; i < fw_status.count; i++)
974 cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "%08X\n",
975 fw_status.status[i]);
976 return cnt;
977}
978static DEVICE_ATTR_RO(fw_status);
979
980/**
981 * hbm_ver_show - display HBM protocol version negotiated with FW
982 *
983 * @device: device pointer
984 * @attr: attribute pointer
985 * @buf: char out buffer
986 *
987 * Return: number of the bytes printed into buf or error
988 */
989static ssize_t hbm_ver_show(struct device *device,
990 struct device_attribute *attr, char *buf)
991{
992 struct mei_device *dev = dev_get_drvdata(device);
993 struct hbm_version ver;
994
995 mutex_lock(&dev->device_lock);
996 ver = dev->version;
997 mutex_unlock(&dev->device_lock);
998
999 return sprintf(buf, "%u.%u\n", ver.major_version, ver.minor_version);
1000}
1001static DEVICE_ATTR_RO(hbm_ver);
1002
1003/**
1004 * hbm_ver_drv_show - display HBM protocol version advertised by driver
1005 *
1006 * @device: device pointer
1007 * @attr: attribute pointer
1008 * @buf: char out buffer
1009 *
1010 * Return: number of the bytes printed into buf or error
1011 */
1012static ssize_t hbm_ver_drv_show(struct device *device,
1013 struct device_attribute *attr, char *buf)
1014{
1015 return sprintf(buf, "%u.%u\n", HBM_MAJOR_VERSION, HBM_MINOR_VERSION);
1016}
1017static DEVICE_ATTR_RO(hbm_ver_drv);
1018
1019static ssize_t tx_queue_limit_show(struct device *device,
1020 struct device_attribute *attr, char *buf)
1021{
1022 struct mei_device *dev = dev_get_drvdata(device);
1023 u8 size = 0;
1024
1025 mutex_lock(&dev->device_lock);
1026 size = dev->tx_queue_limit;
1027 mutex_unlock(&dev->device_lock);
1028
1029 return sysfs_emit(buf, "%u\n", size);
1030}
1031
1032static ssize_t tx_queue_limit_store(struct device *device,
1033 struct device_attribute *attr,
1034 const char *buf, size_t count)
1035{
1036 struct mei_device *dev = dev_get_drvdata(device);
1037 u8 limit;
1038 unsigned int inp;
1039 int err;
1040
1041 err = kstrtouint(buf, 10, &inp);
1042 if (err)
1043 return err;
1044 if (inp > MEI_TX_QUEUE_LIMIT_MAX || inp < MEI_TX_QUEUE_LIMIT_MIN)
1045 return -EINVAL;
1046 limit = inp;
1047
1048 mutex_lock(&dev->device_lock);
1049 dev->tx_queue_limit = limit;
1050 mutex_unlock(&dev->device_lock);
1051
1052 return count;
1053}
1054static DEVICE_ATTR_RW(tx_queue_limit);
1055
1056/**
1057 * fw_ver_show - display ME FW version
1058 *
1059 * @device: device pointer
1060 * @attr: attribute pointer
1061 * @buf: char out buffer
1062 *
1063 * Return: number of the bytes printed into buf or error
1064 */
1065static ssize_t fw_ver_show(struct device *device,
1066 struct device_attribute *attr, char *buf)
1067{
1068 struct mei_device *dev = dev_get_drvdata(device);
1069 struct mei_fw_version *ver;
1070 ssize_t cnt = 0;
1071 int i;
1072
1073 ver = dev->fw_ver;
1074
1075 for (i = 0; i < MEI_MAX_FW_VER_BLOCKS; i++)
1076 cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "%u:%u.%u.%u.%u\n",
1077 ver[i].platform, ver[i].major, ver[i].minor,
1078 ver[i].hotfix, ver[i].buildno);
1079 return cnt;
1080}
1081static DEVICE_ATTR_RO(fw_ver);
1082
1083/**
1084 * dev_state_show - display device state
1085 *
1086 * @device: device pointer
1087 * @attr: attribute pointer
1088 * @buf: char out buffer
1089 *
1090 * Return: number of the bytes printed into buf or error
1091 */
1092static ssize_t dev_state_show(struct device *device,
1093 struct device_attribute *attr, char *buf)
1094{
1095 struct mei_device *dev = dev_get_drvdata(device);
1096 enum mei_dev_state dev_state;
1097
1098 mutex_lock(&dev->device_lock);
1099 dev_state = dev->dev_state;
1100 mutex_unlock(&dev->device_lock);
1101
1102 return sprintf(buf, "%s", mei_dev_state_str(dev_state));
1103}
1104static DEVICE_ATTR_RO(dev_state);
1105
1106/**
1107 * mei_set_devstate: set to new device state and notify sysfs file.
1108 *
1109 * @dev: mei_device
1110 * @state: new device state
1111 */
1112void mei_set_devstate(struct mei_device *dev, enum mei_dev_state state)
1113{
1114 struct device *clsdev;
1115
1116 if (dev->dev_state == state)
1117 return;
1118
1119 dev->dev_state = state;
1120
1121 clsdev = class_find_device_by_devt(&mei_class, dev->cdev.dev);
1122 if (clsdev) {
1123 sysfs_notify(&clsdev->kobj, NULL, "dev_state");
1124 put_device(clsdev);
1125 }
1126}
1127
1128/**
1129 * kind_show - display device kind
1130 *
1131 * @device: device pointer
1132 * @attr: attribute pointer
1133 * @buf: char out buffer
1134 *
1135 * Return: number of the bytes printed into buf or error
1136 */
1137static ssize_t kind_show(struct device *device,
1138 struct device_attribute *attr, char *buf)
1139{
1140 struct mei_device *dev = dev_get_drvdata(device);
1141 ssize_t ret;
1142
1143 if (dev->kind)
1144 ret = sprintf(buf, "%s\n", dev->kind);
1145 else
1146 ret = sprintf(buf, "%s\n", "mei");
1147
1148 return ret;
1149}
1150static DEVICE_ATTR_RO(kind);
1151
1152static struct attribute *mei_attrs[] = {
1153 &dev_attr_fw_status.attr,
1154 &dev_attr_hbm_ver.attr,
1155 &dev_attr_hbm_ver_drv.attr,
1156 &dev_attr_tx_queue_limit.attr,
1157 &dev_attr_fw_ver.attr,
1158 &dev_attr_dev_state.attr,
1159 &dev_attr_trc.attr,
1160 &dev_attr_kind.attr,
1161 NULL
1162};
1163ATTRIBUTE_GROUPS(mei);
1164
1165/*
1166 * file operations structure will be used for mei char device.
1167 */
1168static const struct file_operations mei_fops = {
1169 .owner = THIS_MODULE,
1170 .read = mei_read,
1171 .unlocked_ioctl = mei_ioctl,
1172 .compat_ioctl = compat_ptr_ioctl,
1173 .open = mei_open,
1174 .release = mei_release,
1175 .write = mei_write,
1176 .poll = mei_poll,
1177 .fsync = mei_fsync,
1178 .fasync = mei_fasync,
1179 .llseek = no_llseek
1180};
1181
1182/**
1183 * mei_minor_get - obtain next free device minor number
1184 *
1185 * @dev: device pointer
1186 *
1187 * Return: allocated minor, or -ENOSPC if no free minor left
1188 */
1189static int mei_minor_get(struct mei_device *dev)
1190{
1191 int ret;
1192
1193 mutex_lock(&mei_minor_lock);
1194 ret = idr_alloc(&mei_idr, dev, 0, MEI_MAX_DEVS, GFP_KERNEL);
1195 if (ret >= 0)
1196 dev->minor = ret;
1197 else if (ret == -ENOSPC)
1198 dev_err(dev->dev, "too many mei devices\n");
1199
1200 mutex_unlock(&mei_minor_lock);
1201 return ret;
1202}
1203
1204/**
1205 * mei_minor_free - mark device minor number as free
1206 *
1207 * @dev: device pointer
1208 */
1209static void mei_minor_free(struct mei_device *dev)
1210{
1211 mutex_lock(&mei_minor_lock);
1212 idr_remove(&mei_idr, dev->minor);
1213 mutex_unlock(&mei_minor_lock);
1214}
1215
1216int mei_register(struct mei_device *dev, struct device *parent)
1217{
1218 struct device *clsdev; /* class device */
1219 int ret, devno;
1220
1221 ret = mei_minor_get(dev);
1222 if (ret < 0)
1223 return ret;
1224
1225 /* Fill in the data structures */
1226 devno = MKDEV(MAJOR(mei_devt), dev->minor);
1227 cdev_init(&dev->cdev, &mei_fops);
1228 dev->cdev.owner = parent->driver->owner;
1229
1230 /* Add the device */
1231 ret = cdev_add(&dev->cdev, devno, 1);
1232 if (ret) {
1233 dev_err(parent, "unable to add device %d:%d\n",
1234 MAJOR(mei_devt), dev->minor);
1235 goto err_dev_add;
1236 }
1237
1238 clsdev = device_create_with_groups(&mei_class, parent, devno,
1239 dev, mei_groups,
1240 "mei%d", dev->minor);
1241
1242 if (IS_ERR(clsdev)) {
1243 dev_err(parent, "unable to create device %d:%d\n",
1244 MAJOR(mei_devt), dev->minor);
1245 ret = PTR_ERR(clsdev);
1246 goto err_dev_create;
1247 }
1248
1249 mei_dbgfs_register(dev, dev_name(clsdev));
1250
1251 return 0;
1252
1253err_dev_create:
1254 cdev_del(&dev->cdev);
1255err_dev_add:
1256 mei_minor_free(dev);
1257 return ret;
1258}
1259EXPORT_SYMBOL_GPL(mei_register);
1260
1261void mei_deregister(struct mei_device *dev)
1262{
1263 int devno;
1264
1265 devno = dev->cdev.dev;
1266 cdev_del(&dev->cdev);
1267
1268 mei_dbgfs_deregister(dev);
1269
1270 device_destroy(&mei_class, devno);
1271
1272 mei_minor_free(dev);
1273}
1274EXPORT_SYMBOL_GPL(mei_deregister);
1275
1276static int __init mei_init(void)
1277{
1278 int ret;
1279
1280 ret = class_register(&mei_class);
1281 if (ret)
1282 return ret;
1283
1284 ret = alloc_chrdev_region(&mei_devt, 0, MEI_MAX_DEVS, "mei");
1285 if (ret < 0) {
1286 pr_err("unable to allocate char dev region\n");
1287 goto err_class;
1288 }
1289
1290 ret = mei_cl_bus_init();
1291 if (ret < 0) {
1292 pr_err("unable to initialize bus\n");
1293 goto err_chrdev;
1294 }
1295
1296 return 0;
1297
1298err_chrdev:
1299 unregister_chrdev_region(mei_devt, MEI_MAX_DEVS);
1300err_class:
1301 class_unregister(&mei_class);
1302 return ret;
1303}
1304
1305static void __exit mei_exit(void)
1306{
1307 unregister_chrdev_region(mei_devt, MEI_MAX_DEVS);
1308 class_unregister(&mei_class);
1309 mei_cl_bus_exit();
1310}
1311
1312module_init(mei_init);
1313module_exit(mei_exit);
1314
1315MODULE_AUTHOR("Intel Corporation");
1316MODULE_DESCRIPTION("Intel(R) Management Engine Interface");
1317MODULE_LICENSE("GPL v2");
1318