1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Intel On Demand (Software Defined Silicon) driver
4 *
5 * Copyright (c) 2022, Intel Corporation.
6 * All Rights Reserved.
7 *
8 * Author: "David E. Box" <david.e.box@linux.intel.com>
9 */
10
11 #include <linux/auxiliary_bus.h>
12 #include <linux/bits.h>
13 #include <linux/bitfield.h>
14 #include <linux/device.h>
15 #include <linux/iopoll.h>
16 #include <linux/kernel.h>
17 #include <linux/module.h>
18 #include <linux/pci.h>
19 #include <linux/slab.h>
20 #include <linux/sysfs.h>
21 #include <linux/types.h>
22 #include <linux/uaccess.h>
23
24 #include "vsec.h"
25
26 #define ACCESS_TYPE_BARID 2
27 #define ACCESS_TYPE_LOCAL 3
28
29 #define SDSI_MIN_SIZE_DWORDS 276
30 #define SDSI_SIZE_MAILBOX 1024
31 #define SDSI_SIZE_REGS 80
32 #define SDSI_SIZE_CMD sizeof(u64)
33
34 /*
35 * Write messages are currently up to the size of the mailbox
36 * while read messages are up to 4 times the size of the
37 * mailbox, sent in packets
38 */
39 #define SDSI_SIZE_WRITE_MSG SDSI_SIZE_MAILBOX
40 #define SDSI_SIZE_READ_MSG (SDSI_SIZE_MAILBOX * 4)
41
42 #define SDSI_ENABLED_FEATURES_OFFSET 16
43 #define SDSI_FEATURE_SDSI BIT(3)
44 #define SDSI_FEATURE_METERING BIT(26)
45
46 #define SDSI_SOCKET_ID_OFFSET 64
47 #define SDSI_SOCKET_ID GENMASK(3, 0)
48
49 #define SDSI_MBOX_CMD_SUCCESS 0x40
50 #define SDSI_MBOX_CMD_TIMEOUT 0x80
51
52 #define MBOX_TIMEOUT_US 500000
53 #define MBOX_TIMEOUT_ACQUIRE_US 1000
54 #define MBOX_POLLING_PERIOD_US 100
55 #define MBOX_ACQUIRE_NUM_RETRIES 5
56 #define MBOX_ACQUIRE_RETRY_DELAY_MS 500
57 #define MBOX_MAX_PACKETS 4
58
59 #define MBOX_OWNER_NONE 0x00
60 #define MBOX_OWNER_INBAND 0x01
61
62 #define CTRL_RUN_BUSY BIT(0)
63 #define CTRL_READ_WRITE BIT(1)
64 #define CTRL_SOM BIT(2)
65 #define CTRL_EOM BIT(3)
66 #define CTRL_OWNER GENMASK(5, 4)
67 #define CTRL_COMPLETE BIT(6)
68 #define CTRL_READY BIT(7)
69 #define CTRL_STATUS GENMASK(15, 8)
70 #define CTRL_PACKET_SIZE GENMASK(31, 16)
71 #define CTRL_MSG_SIZE GENMASK(63, 48)
72
73 #define DISC_TABLE_SIZE 12
74 #define DT_ACCESS_TYPE GENMASK(3, 0)
75 #define DT_SIZE GENMASK(27, 12)
76 #define DT_TBIR GENMASK(2, 0)
77 #define DT_OFFSET(v) ((v) & GENMASK(31, 3))
78
79 #define SDSI_GUID_V1 0x006DD191
80 #define GUID_V1_CNTRL_SIZE 8
81 #define GUID_V1_REGS_SIZE 72
82 #define SDSI_GUID_V2 0xF210D9EF
83 #define GUID_V2_CNTRL_SIZE 16
84 #define GUID_V2_REGS_SIZE 80
85
86 enum sdsi_command {
87 SDSI_CMD_PROVISION_AKC = 0x0004,
88 SDSI_CMD_PROVISION_CAP = 0x0008,
89 SDSI_CMD_READ_STATE = 0x0010,
90 SDSI_CMD_READ_METER = 0x0014,
91 };
92
93 struct sdsi_mbox_info {
94 u64 *payload;
95 void *buffer;
96 int size;
97 };
98
99 struct disc_table {
100 u32 access_info;
101 u32 guid;
102 u32 offset;
103 };
104
105 struct sdsi_priv {
106 struct mutex mb_lock; /* Mailbox access lock */
107 struct device *dev;
108 void __iomem *control_addr;
109 void __iomem *mbox_addr;
110 void __iomem *regs_addr;
111 int control_size;
112 int maibox_size;
113 int registers_size;
114 u32 guid;
115 u32 features;
116 };
117
118 /* SDSi mailbox operations must be performed using 64bit mov instructions */
119 static __always_inline void
sdsi_memcpy64_toio(u64 __iomem * to,const u64 * from,size_t count_bytes)120 sdsi_memcpy64_toio(u64 __iomem *to, const u64 *from, size_t count_bytes)
121 {
122 size_t count = count_bytes / sizeof(*to);
123 int i;
124
125 for (i = 0; i < count; i++)
126 writeq(from[i], &to[i]);
127 }
128
129 static __always_inline void
sdsi_memcpy64_fromio(u64 * to,const u64 __iomem * from,size_t count_bytes)130 sdsi_memcpy64_fromio(u64 *to, const u64 __iomem *from, size_t count_bytes)
131 {
132 size_t count = count_bytes / sizeof(*to);
133 int i;
134
135 for (i = 0; i < count; i++)
136 to[i] = readq(&from[i]);
137 }
138
sdsi_complete_transaction(struct sdsi_priv * priv)139 static inline void sdsi_complete_transaction(struct sdsi_priv *priv)
140 {
141 u64 control = FIELD_PREP(CTRL_COMPLETE, 1);
142
143 lockdep_assert_held(&priv->mb_lock);
144 writeq(control, priv->control_addr);
145 }
146
sdsi_status_to_errno(u32 status)147 static int sdsi_status_to_errno(u32 status)
148 {
149 switch (status) {
150 case SDSI_MBOX_CMD_SUCCESS:
151 return 0;
152 case SDSI_MBOX_CMD_TIMEOUT:
153 return -ETIMEDOUT;
154 default:
155 return -EIO;
156 }
157 }
158
sdsi_mbox_cmd_read(struct sdsi_priv * priv,struct sdsi_mbox_info * info,size_t * data_size)159 static int sdsi_mbox_cmd_read(struct sdsi_priv *priv, struct sdsi_mbox_info *info,
160 size_t *data_size)
161 {
162 struct device *dev = priv->dev;
163 u32 total, loop, eom, status, message_size;
164 u64 control;
165 int ret;
166
167 lockdep_assert_held(&priv->mb_lock);
168
169 /* Format and send the read command */
170 control = FIELD_PREP(CTRL_EOM, 1) |
171 FIELD_PREP(CTRL_SOM, 1) |
172 FIELD_PREP(CTRL_RUN_BUSY, 1) |
173 FIELD_PREP(CTRL_PACKET_SIZE, info->size);
174 writeq(control, priv->control_addr);
175
176 /* For reads, data sizes that are larger than the mailbox size are read in packets. */
177 total = 0;
178 loop = 0;
179 do {
180 void *buf = info->buffer + (SDSI_SIZE_MAILBOX * loop);
181 u32 packet_size;
182
183 /* Poll on ready bit */
184 ret = readq_poll_timeout(priv->control_addr, control, control & CTRL_READY,
185 MBOX_POLLING_PERIOD_US, MBOX_TIMEOUT_US);
186 if (ret)
187 break;
188
189 eom = FIELD_GET(CTRL_EOM, control);
190 status = FIELD_GET(CTRL_STATUS, control);
191 packet_size = FIELD_GET(CTRL_PACKET_SIZE, control);
192 message_size = FIELD_GET(CTRL_MSG_SIZE, control);
193
194 ret = sdsi_status_to_errno(status);
195 if (ret)
196 break;
197
198 /* Only the last packet can be less than the mailbox size. */
199 if (!eom && packet_size != SDSI_SIZE_MAILBOX) {
200 dev_err(dev, "Invalid packet size\n");
201 ret = -EPROTO;
202 break;
203 }
204
205 if (packet_size > SDSI_SIZE_MAILBOX) {
206 dev_err(dev, "Packet size too large\n");
207 ret = -EPROTO;
208 break;
209 }
210
211 sdsi_memcpy64_fromio(buf, priv->mbox_addr, round_up(packet_size, SDSI_SIZE_CMD));
212
213 total += packet_size;
214
215 sdsi_complete_transaction(priv);
216 } while (!eom && ++loop < MBOX_MAX_PACKETS);
217
218 if (ret) {
219 sdsi_complete_transaction(priv);
220 return ret;
221 }
222
223 if (!eom) {
224 dev_err(dev, "Exceeded read attempts\n");
225 return -EPROTO;
226 }
227
228 /* Message size check is only valid for multi-packet transfers */
229 if (loop && total != message_size)
230 dev_warn(dev, "Read count %u differs from expected count %u\n",
231 total, message_size);
232
233 *data_size = total;
234
235 return 0;
236 }
237
sdsi_mbox_cmd_write(struct sdsi_priv * priv,struct sdsi_mbox_info * info)238 static int sdsi_mbox_cmd_write(struct sdsi_priv *priv, struct sdsi_mbox_info *info)
239 {
240 u64 control;
241 u32 status;
242 int ret;
243
244 lockdep_assert_held(&priv->mb_lock);
245
246 /* Write rest of the payload */
247 sdsi_memcpy64_toio(priv->mbox_addr + SDSI_SIZE_CMD, info->payload + 1,
248 info->size - SDSI_SIZE_CMD);
249
250 /* Format and send the write command */
251 control = FIELD_PREP(CTRL_EOM, 1) |
252 FIELD_PREP(CTRL_SOM, 1) |
253 FIELD_PREP(CTRL_RUN_BUSY, 1) |
254 FIELD_PREP(CTRL_READ_WRITE, 1) |
255 FIELD_PREP(CTRL_PACKET_SIZE, info->size);
256 writeq(control, priv->control_addr);
257
258 /* Poll on ready bit */
259 ret = readq_poll_timeout(priv->control_addr, control, control & CTRL_READY,
260 MBOX_POLLING_PERIOD_US, MBOX_TIMEOUT_US);
261
262 if (ret)
263 goto release_mbox;
264
265 status = FIELD_GET(CTRL_STATUS, control);
266 ret = sdsi_status_to_errno(status);
267
268 release_mbox:
269 sdsi_complete_transaction(priv);
270
271 return ret;
272 }
273
sdsi_mbox_acquire(struct sdsi_priv * priv,struct sdsi_mbox_info * info)274 static int sdsi_mbox_acquire(struct sdsi_priv *priv, struct sdsi_mbox_info *info)
275 {
276 u64 control;
277 u32 owner;
278 int ret, retries = 0;
279
280 lockdep_assert_held(&priv->mb_lock);
281
282 /* Check mailbox is available */
283 control = readq(priv->control_addr);
284 owner = FIELD_GET(CTRL_OWNER, control);
285 if (owner != MBOX_OWNER_NONE)
286 return -EBUSY;
287
288 /*
289 * If there has been no recent transaction and no one owns the mailbox,
290 * we should acquire it in under 1ms. However, if we've accessed it
291 * recently it may take up to 2.1 seconds to acquire it again.
292 */
293 do {
294 /* Write first qword of payload */
295 writeq(info->payload[0], priv->mbox_addr);
296
297 /* Check for ownership */
298 ret = readq_poll_timeout(priv->control_addr, control,
299 FIELD_GET(CTRL_OWNER, control) == MBOX_OWNER_INBAND,
300 MBOX_POLLING_PERIOD_US, MBOX_TIMEOUT_ACQUIRE_US);
301
302 if (FIELD_GET(CTRL_OWNER, control) == MBOX_OWNER_NONE &&
303 retries++ < MBOX_ACQUIRE_NUM_RETRIES) {
304 msleep(MBOX_ACQUIRE_RETRY_DELAY_MS);
305 continue;
306 }
307
308 /* Either we got it or someone else did. */
309 break;
310 } while (true);
311
312 return ret;
313 }
314
sdsi_mbox_write(struct sdsi_priv * priv,struct sdsi_mbox_info * info)315 static int sdsi_mbox_write(struct sdsi_priv *priv, struct sdsi_mbox_info *info)
316 {
317 int ret;
318
319 lockdep_assert_held(&priv->mb_lock);
320
321 ret = sdsi_mbox_acquire(priv, info);
322 if (ret)
323 return ret;
324
325 return sdsi_mbox_cmd_write(priv, info);
326 }
327
sdsi_mbox_read(struct sdsi_priv * priv,struct sdsi_mbox_info * info,size_t * data_size)328 static int sdsi_mbox_read(struct sdsi_priv *priv, struct sdsi_mbox_info *info, size_t *data_size)
329 {
330 int ret;
331
332 lockdep_assert_held(&priv->mb_lock);
333
334 ret = sdsi_mbox_acquire(priv, info);
335 if (ret)
336 return ret;
337
338 return sdsi_mbox_cmd_read(priv, info, data_size);
339 }
340
sdsi_provision(struct sdsi_priv * priv,char * buf,size_t count,enum sdsi_command command)341 static ssize_t sdsi_provision(struct sdsi_priv *priv, char *buf, size_t count,
342 enum sdsi_command command)
343 {
344 struct sdsi_mbox_info info;
345 int ret;
346
347 if (count > (SDSI_SIZE_WRITE_MSG - SDSI_SIZE_CMD))
348 return -EOVERFLOW;
349
350 /* Qword aligned message + command qword */
351 info.size = round_up(count, SDSI_SIZE_CMD) + SDSI_SIZE_CMD;
352
353 info.payload = kzalloc(info.size, GFP_KERNEL);
354 if (!info.payload)
355 return -ENOMEM;
356
357 /* Copy message to payload buffer */
358 memcpy(info.payload, buf, count);
359
360 /* Command is last qword of payload buffer */
361 info.payload[(info.size - SDSI_SIZE_CMD) / SDSI_SIZE_CMD] = command;
362
363 ret = mutex_lock_interruptible(&priv->mb_lock);
364 if (ret)
365 goto free_payload;
366 ret = sdsi_mbox_write(priv, &info);
367 mutex_unlock(&priv->mb_lock);
368
369 free_payload:
370 kfree(info.payload);
371
372 if (ret)
373 return ret;
374
375 return count;
376 }
377
provision_akc_write(struct file * filp,struct kobject * kobj,struct bin_attribute * attr,char * buf,loff_t off,size_t count)378 static ssize_t provision_akc_write(struct file *filp, struct kobject *kobj,
379 struct bin_attribute *attr, char *buf, loff_t off,
380 size_t count)
381 {
382 struct device *dev = kobj_to_dev(kobj);
383 struct sdsi_priv *priv = dev_get_drvdata(dev);
384
385 if (off)
386 return -ESPIPE;
387
388 return sdsi_provision(priv, buf, count, SDSI_CMD_PROVISION_AKC);
389 }
390 static BIN_ATTR_WO(provision_akc, SDSI_SIZE_WRITE_MSG);
391
provision_cap_write(struct file * filp,struct kobject * kobj,struct bin_attribute * attr,char * buf,loff_t off,size_t count)392 static ssize_t provision_cap_write(struct file *filp, struct kobject *kobj,
393 struct bin_attribute *attr, char *buf, loff_t off,
394 size_t count)
395 {
396 struct device *dev = kobj_to_dev(kobj);
397 struct sdsi_priv *priv = dev_get_drvdata(dev);
398
399 if (off)
400 return -ESPIPE;
401
402 return sdsi_provision(priv, buf, count, SDSI_CMD_PROVISION_CAP);
403 }
404 static BIN_ATTR_WO(provision_cap, SDSI_SIZE_WRITE_MSG);
405
406 static ssize_t
certificate_read(u64 command,struct sdsi_priv * priv,char * buf,loff_t off,size_t count)407 certificate_read(u64 command, struct sdsi_priv *priv, char *buf, loff_t off,
408 size_t count)
409 {
410 struct sdsi_mbox_info info;
411 size_t size;
412 int ret;
413
414 if (off)
415 return 0;
416
417 /* Buffer for return data */
418 info.buffer = kmalloc(SDSI_SIZE_READ_MSG, GFP_KERNEL);
419 if (!info.buffer)
420 return -ENOMEM;
421
422 info.payload = &command;
423 info.size = sizeof(command);
424
425 ret = mutex_lock_interruptible(&priv->mb_lock);
426 if (ret)
427 goto free_buffer;
428 ret = sdsi_mbox_read(priv, &info, &size);
429 mutex_unlock(&priv->mb_lock);
430 if (ret < 0)
431 goto free_buffer;
432
433 if (size > count)
434 size = count;
435
436 memcpy(buf, info.buffer, size);
437
438 free_buffer:
439 kfree(info.buffer);
440
441 if (ret)
442 return ret;
443
444 return size;
445 }
446
447 static ssize_t
state_certificate_read(struct file * filp,struct kobject * kobj,struct bin_attribute * attr,char * buf,loff_t off,size_t count)448 state_certificate_read(struct file *filp, struct kobject *kobj,
449 struct bin_attribute *attr, char *buf, loff_t off,
450 size_t count)
451 {
452 struct device *dev = kobj_to_dev(kobj);
453 struct sdsi_priv *priv = dev_get_drvdata(dev);
454
455 return certificate_read(SDSI_CMD_READ_STATE, priv, buf, off, count);
456 }
457 static BIN_ATTR_ADMIN_RO(state_certificate, SDSI_SIZE_READ_MSG);
458
459 static ssize_t
meter_certificate_read(struct file * filp,struct kobject * kobj,struct bin_attribute * attr,char * buf,loff_t off,size_t count)460 meter_certificate_read(struct file *filp, struct kobject *kobj,
461 struct bin_attribute *attr, char *buf, loff_t off,
462 size_t count)
463 {
464 struct device *dev = kobj_to_dev(kobj);
465 struct sdsi_priv *priv = dev_get_drvdata(dev);
466
467 return certificate_read(SDSI_CMD_READ_METER, priv, buf, off, count);
468 }
469 static BIN_ATTR_ADMIN_RO(meter_certificate, SDSI_SIZE_READ_MSG);
470
registers_read(struct file * filp,struct kobject * kobj,struct bin_attribute * attr,char * buf,loff_t off,size_t count)471 static ssize_t registers_read(struct file *filp, struct kobject *kobj,
472 struct bin_attribute *attr, char *buf, loff_t off,
473 size_t count)
474 {
475 struct device *dev = kobj_to_dev(kobj);
476 struct sdsi_priv *priv = dev_get_drvdata(dev);
477 void __iomem *addr = priv->regs_addr;
478 int size = priv->registers_size;
479
480 /*
481 * The check below is performed by the sysfs caller based on the static
482 * file size. But this may be greater than the actual size which is based
483 * on the GUID. So check here again based on actual size before reading.
484 */
485 if (off >= size)
486 return 0;
487
488 if (off + count > size)
489 count = size - off;
490
491 memcpy_fromio(buf, addr + off, count);
492
493 return count;
494 }
495 static BIN_ATTR_ADMIN_RO(registers, SDSI_SIZE_REGS);
496
497 static struct bin_attribute *sdsi_bin_attrs[] = {
498 &bin_attr_registers,
499 &bin_attr_state_certificate,
500 &bin_attr_meter_certificate,
501 &bin_attr_provision_akc,
502 &bin_attr_provision_cap,
503 NULL
504 };
505
506 static umode_t
sdsi_battr_is_visible(struct kobject * kobj,struct bin_attribute * attr,int n)507 sdsi_battr_is_visible(struct kobject *kobj, struct bin_attribute *attr, int n)
508 {
509 struct device *dev = kobj_to_dev(kobj);
510 struct sdsi_priv *priv = dev_get_drvdata(dev);
511
512 /* Registers file is always readable if the device is present */
513 if (attr == &bin_attr_registers)
514 return attr->attr.mode;
515
516 /* All other attributes not visible if BIOS has not enabled On Demand */
517 if (!(priv->features & SDSI_FEATURE_SDSI))
518 return 0;
519
520 if (attr == &bin_attr_meter_certificate)
521 return (priv->features & SDSI_FEATURE_METERING) ?
522 attr->attr.mode : 0;
523
524 return attr->attr.mode;
525 }
526
guid_show(struct device * dev,struct device_attribute * attr,char * buf)527 static ssize_t guid_show(struct device *dev, struct device_attribute *attr, char *buf)
528 {
529 struct sdsi_priv *priv = dev_get_drvdata(dev);
530
531 return sysfs_emit(buf, "0x%x\n", priv->guid);
532 }
533 static DEVICE_ATTR_RO(guid);
534
535 static struct attribute *sdsi_attrs[] = {
536 &dev_attr_guid.attr,
537 NULL
538 };
539
540 static const struct attribute_group sdsi_group = {
541 .attrs = sdsi_attrs,
542 .bin_attrs = sdsi_bin_attrs,
543 .is_bin_visible = sdsi_battr_is_visible,
544 };
545 __ATTRIBUTE_GROUPS(sdsi);
546
sdsi_get_layout(struct sdsi_priv * priv,struct disc_table * table)547 static int sdsi_get_layout(struct sdsi_priv *priv, struct disc_table *table)
548 {
549 switch (table->guid) {
550 case SDSI_GUID_V1:
551 priv->control_size = GUID_V1_CNTRL_SIZE;
552 priv->registers_size = GUID_V1_REGS_SIZE;
553 break;
554 case SDSI_GUID_V2:
555 priv->control_size = GUID_V2_CNTRL_SIZE;
556 priv->registers_size = GUID_V2_REGS_SIZE;
557 break;
558 default:
559 dev_err(priv->dev, "Unrecognized GUID 0x%x\n", table->guid);
560 return -EINVAL;
561 }
562 return 0;
563 }
564
sdsi_map_mbox_registers(struct sdsi_priv * priv,struct pci_dev * parent,struct disc_table * disc_table,struct resource * disc_res)565 static int sdsi_map_mbox_registers(struct sdsi_priv *priv, struct pci_dev *parent,
566 struct disc_table *disc_table, struct resource *disc_res)
567 {
568 u32 access_type = FIELD_GET(DT_ACCESS_TYPE, disc_table->access_info);
569 u32 size = FIELD_GET(DT_SIZE, disc_table->access_info);
570 u32 tbir = FIELD_GET(DT_TBIR, disc_table->offset);
571 u32 offset = DT_OFFSET(disc_table->offset);
572 struct resource res = {};
573
574 /* Starting location of SDSi MMIO region based on access type */
575 switch (access_type) {
576 case ACCESS_TYPE_LOCAL:
577 if (tbir) {
578 dev_err(priv->dev, "Unsupported BAR index %u for access type %u\n",
579 tbir, access_type);
580 return -EINVAL;
581 }
582
583 /*
584 * For access_type LOCAL, the base address is as follows:
585 * base address = end of discovery region + base offset + 1
586 */
587 res.start = disc_res->end + offset + 1;
588 break;
589
590 case ACCESS_TYPE_BARID:
591 res.start = pci_resource_start(parent, tbir) + offset;
592 break;
593
594 default:
595 dev_err(priv->dev, "Unrecognized access_type %u\n", access_type);
596 return -EINVAL;
597 }
598
599 res.end = res.start + size * sizeof(u32) - 1;
600 res.flags = IORESOURCE_MEM;
601
602 priv->control_addr = devm_ioremap_resource(priv->dev, &res);
603 if (IS_ERR(priv->control_addr))
604 return PTR_ERR(priv->control_addr);
605
606 priv->mbox_addr = priv->control_addr + priv->control_size;
607 priv->regs_addr = priv->mbox_addr + SDSI_SIZE_MAILBOX;
608
609 priv->features = readq(priv->regs_addr + SDSI_ENABLED_FEATURES_OFFSET);
610
611 return 0;
612 }
613
sdsi_probe(struct auxiliary_device * auxdev,const struct auxiliary_device_id * id)614 static int sdsi_probe(struct auxiliary_device *auxdev, const struct auxiliary_device_id *id)
615 {
616 struct intel_vsec_device *intel_cap_dev = auxdev_to_ivdev(auxdev);
617 struct disc_table disc_table;
618 struct resource *disc_res;
619 void __iomem *disc_addr;
620 struct sdsi_priv *priv;
621 int ret;
622
623 priv = devm_kzalloc(&auxdev->dev, sizeof(*priv), GFP_KERNEL);
624 if (!priv)
625 return -ENOMEM;
626
627 priv->dev = &auxdev->dev;
628 mutex_init(&priv->mb_lock);
629 auxiliary_set_drvdata(auxdev, priv);
630
631 /* Get the SDSi discovery table */
632 disc_res = &intel_cap_dev->resource[0];
633 disc_addr = devm_ioremap_resource(&auxdev->dev, disc_res);
634 if (IS_ERR(disc_addr))
635 return PTR_ERR(disc_addr);
636
637 memcpy_fromio(&disc_table, disc_addr, DISC_TABLE_SIZE);
638
639 priv->guid = disc_table.guid;
640
641 /* Get guid based layout info */
642 ret = sdsi_get_layout(priv, &disc_table);
643 if (ret)
644 return ret;
645
646 /* Map the SDSi mailbox registers */
647 ret = sdsi_map_mbox_registers(priv, intel_cap_dev->pcidev, &disc_table, disc_res);
648 if (ret)
649 return ret;
650
651 return 0;
652 }
653
654 static const struct auxiliary_device_id sdsi_aux_id_table[] = {
655 { .name = "intel_vsec.sdsi" },
656 {}
657 };
658 MODULE_DEVICE_TABLE(auxiliary, sdsi_aux_id_table);
659
660 static struct auxiliary_driver sdsi_aux_driver = {
661 .driver = {
662 .dev_groups = sdsi_groups,
663 },
664 .id_table = sdsi_aux_id_table,
665 .probe = sdsi_probe,
666 /* No remove. All resources are handled under devm */
667 };
668 module_auxiliary_driver(sdsi_aux_driver);
669
670 MODULE_AUTHOR("David E. Box <david.e.box@linux.intel.com>");
671 MODULE_DESCRIPTION("Intel On Demand (SDSi) driver");
672 MODULE_LICENSE("GPL");
673