2 * This file is provided under a dual BSD/GPLv2 license. When using or
3 * redistributing this file, you may do so under either license.
7 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of version 2 of the GNU General Public License as
11 * published by the Free Software Foundation.
13 * This program is distributed in the hope that it will be useful, but
14 * WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * General Public License for more details.
18 * You should have received a copy of the GNU General Public License
19 * along with this program; if not, write to the Free Software
20 * Foundation, Inc., 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
21 * The full GNU General Public License is included in this distribution
22 * in the file called LICENSE.GPL.
26 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
27 * All rights reserved.
29 * Redistribution and use in source and binary forms, with or without
30 * modification, are permitted provided that the following conditions
33 * * Redistributions of source code must retain the above copyright
34 * notice, this list of conditions and the following disclaimer.
35 * * Redistributions in binary form must reproduce the above copyright
36 * notice, this list of conditions and the following disclaimer in
37 * the documentation and/or other materials provided with the
39 * * Neither the name of Intel Corporation nor the names of its
40 * contributors may be used to endorse or promote products derived
41 * from this software without specific prior written permission.
43 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
44 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
45 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
46 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
47 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
48 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
49 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
50 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
51 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
52 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
53 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
57 #include "scic_io_request.h"
58 #include "scic_remote_device.h"
59 #include "scic_port.h"
64 #include "probe_roms.h"
66 irqreturn_t isci_msix_isr(int vec, void *data)
68 struct isci_host *ihost = data;
69 struct scic_sds_controller *scic = ihost->core_controller;
71 if (scic_sds_controller_isr(scic))
72 tasklet_schedule(&ihost->completion_tasklet);
77 irqreturn_t isci_intx_isr(int vec, void *data)
79 struct pci_dev *pdev = data;
80 struct isci_host *ihost;
81 irqreturn_t ret = IRQ_NONE;
84 for_each_isci_host(i, ihost, pdev) {
85 struct scic_sds_controller *scic = ihost->core_controller;
87 if (scic_sds_controller_isr(scic)) {
88 tasklet_schedule(&ihost->completion_tasklet);
90 } else if (scic_sds_controller_error_isr(scic)) {
91 spin_lock(&ihost->scic_lock);
92 scic_sds_controller_error_handler(scic);
93 spin_unlock(&ihost->scic_lock);
101 irqreturn_t isci_error_isr(int vec, void *data)
103 struct isci_host *ihost = data;
104 struct scic_sds_controller *scic = ihost->core_controller;
106 if (scic_sds_controller_error_isr(scic))
107 scic_sds_controller_error_handler(scic);
113 * isci_host_start_complete() - This function is called by the core library,
114 * through the ISCI Module, to indicate controller start status.
115 * @isci_host: This parameter specifies the ISCI host object
116 * @completion_status: This parameter specifies the completion status from the
120 void isci_host_start_complete(struct isci_host *ihost, enum sci_status completion_status)
122 if (completion_status != SCI_SUCCESS)
123 dev_info(&ihost->pdev->dev,
124 "controller start timed out, continuing...\n");
125 isci_host_change_state(ihost, isci_ready);
126 clear_bit(IHOST_START_PENDING, &ihost->flags);
127 wake_up(&ihost->eventq);
130 int isci_host_scan_finished(struct Scsi_Host *shost, unsigned long time)
132 struct isci_host *ihost = SHOST_TO_SAS_HA(shost)->lldd_ha;
134 if (test_bit(IHOST_START_PENDING, &ihost->flags))
137 /* todo: use sas_flush_discovery once it is upstream */
138 scsi_flush_work(shost);
140 scsi_flush_work(shost);
142 dev_dbg(&ihost->pdev->dev,
143 "%s: ihost->status = %d, time = %ld\n",
144 __func__, isci_host_get_state(ihost), time);
150 void isci_host_scan_start(struct Scsi_Host *shost)
152 struct isci_host *ihost = SHOST_TO_SAS_HA(shost)->lldd_ha;
153 struct scic_sds_controller *scic = ihost->core_controller;
154 unsigned long tmo = scic_controller_get_suggested_start_timeout(scic);
156 set_bit(IHOST_START_PENDING, &ihost->flags);
158 spin_lock_irq(&ihost->scic_lock);
159 scic_controller_start(scic, tmo);
160 scic_controller_enable_interrupts(scic);
161 spin_unlock_irq(&ihost->scic_lock);
164 void isci_host_stop_complete(struct isci_host *ihost, enum sci_status completion_status)
166 isci_host_change_state(ihost, isci_stopped);
167 scic_controller_disable_interrupts(ihost->core_controller);
168 clear_bit(IHOST_STOP_PENDING, &ihost->flags);
169 wake_up(&ihost->eventq);
172 static struct coherent_memory_info *isci_host_alloc_mdl_struct(
173 struct isci_host *isci_host,
176 struct coherent_memory_info *mdl_struct;
177 void *uncached_address = NULL;
180 mdl_struct = devm_kzalloc(&isci_host->pdev->dev,
186 INIT_LIST_HEAD(&mdl_struct->node);
188 uncached_address = dmam_alloc_coherent(&isci_host->pdev->dev,
190 &mdl_struct->dma_handle,
192 if (!uncached_address)
195 /* memset the whole memory area. */
196 memset((char *)uncached_address, 0, size);
197 mdl_struct->vaddr = uncached_address;
198 mdl_struct->size = (size_t)size;
203 static void isci_host_build_mde(
204 struct sci_physical_memory_descriptor *mde_struct,
205 struct coherent_memory_info *mdl_struct)
207 unsigned long address = 0;
208 dma_addr_t dma_addr = 0;
210 address = (unsigned long)mdl_struct->vaddr;
211 dma_addr = mdl_struct->dma_handle;
213 /* to satisfy the alignment. */
214 if ((address % mde_struct->constant_memory_alignment) != 0) {
216 = (mde_struct->constant_memory_alignment
217 - (address % mde_struct->constant_memory_alignment));
218 address += align_offset;
219 dma_addr += align_offset;
222 mde_struct->virtual_address = (void *)address;
223 mde_struct->physical_address = dma_addr;
224 mdl_struct->mde = mde_struct;
227 static int isci_host_mdl_allocate_coherent(
228 struct isci_host *isci_host)
230 struct sci_physical_memory_descriptor *current_mde;
231 struct coherent_memory_info *mdl_struct;
234 struct sci_base_memory_descriptor_list *mdl_handle
235 = sci_controller_get_memory_descriptor_list_handle(
236 isci_host->core_controller);
238 sci_mdl_first_entry(mdl_handle);
240 current_mde = sci_mdl_get_current_entry(mdl_handle);
242 while (current_mde != NULL) {
244 size = (current_mde->constant_memory_size
245 + current_mde->constant_memory_alignment);
247 mdl_struct = isci_host_alloc_mdl_struct(isci_host, size);
251 list_add_tail(&mdl_struct->node, &isci_host->mdl_struct_list);
253 isci_host_build_mde(current_mde, mdl_struct);
255 sci_mdl_next_entry(mdl_handle);
256 current_mde = sci_mdl_get_current_entry(mdl_handle);
264 * isci_host_completion_routine() - This function is the delayed service
265 * routine that calls the sci core library's completion handler. It's
266 * scheduled as a tasklet from the interrupt service routine when interrupts
267 * in use, or set as the timeout function in polled mode.
268 * @data: This parameter specifies the ISCI host object
271 static void isci_host_completion_routine(unsigned long data)
273 struct isci_host *isci_host = (struct isci_host *)data;
274 struct list_head completed_request_list;
275 struct list_head errored_request_list;
276 struct list_head *current_position;
277 struct list_head *next_position;
278 struct isci_request *request;
279 struct isci_request *next_request;
280 struct sas_task *task;
282 INIT_LIST_HEAD(&completed_request_list);
283 INIT_LIST_HEAD(&errored_request_list);
285 spin_lock_irq(&isci_host->scic_lock);
287 scic_sds_controller_completion_handler(isci_host->core_controller);
289 /* Take the lists of completed I/Os from the host. */
291 list_splice_init(&isci_host->requests_to_complete,
292 &completed_request_list);
294 /* Take the list of errored I/Os from the host. */
295 list_splice_init(&isci_host->requests_to_errorback,
296 &errored_request_list);
298 spin_unlock_irq(&isci_host->scic_lock);
300 /* Process any completions in the lists. */
301 list_for_each_safe(current_position, next_position,
302 &completed_request_list) {
304 request = list_entry(current_position, struct isci_request,
306 task = isci_request_access_task(request);
308 /* Normal notification (task_done) */
309 dev_dbg(&isci_host->pdev->dev,
310 "%s: Normal - request/task = %p/%p\n",
315 /* Return the task to libsas */
318 task->lldd_task = NULL;
319 if (!(task->task_state_flags & SAS_TASK_STATE_ABORTED)) {
321 /* If the task is already in the abort path,
322 * the task_done callback cannot be called.
324 task->task_done(task);
327 /* Free the request object. */
328 isci_request_free(isci_host, request);
330 list_for_each_entry_safe(request, next_request, &errored_request_list,
333 task = isci_request_access_task(request);
335 /* Use sas_task_abort */
336 dev_warn(&isci_host->pdev->dev,
337 "%s: Error - request/task = %p/%p\n",
344 /* Put the task into the abort path if it's not there
347 if (!(task->task_state_flags & SAS_TASK_STATE_ABORTED))
348 sas_task_abort(task);
351 /* This is a case where the request has completed with a
352 * status such that it needed further target servicing,
353 * but the sas_task reference has already been removed
354 * from the request. Since it was errored, it was not
355 * being aborted, so there is nothing to do except free
359 spin_lock_irq(&isci_host->scic_lock);
360 /* Remove the request from the remote device's list
361 * of pending requests.
363 list_del_init(&request->dev_node);
364 spin_unlock_irq(&isci_host->scic_lock);
366 /* Free the request object. */
367 isci_request_free(isci_host, request);
373 void isci_host_deinit(struct isci_host *ihost)
375 struct scic_sds_controller *scic = ihost->core_controller;
378 isci_host_change_state(ihost, isci_stopping);
379 for (i = 0; i < SCI_MAX_PORTS; i++) {
380 struct isci_port *port = &ihost->isci_ports[i];
381 struct isci_remote_device *idev, *d;
383 list_for_each_entry_safe(idev, d, &port->remote_dev_list, node) {
384 isci_remote_device_change_state(idev, isci_stopping);
385 isci_remote_device_stop(ihost, idev);
389 set_bit(IHOST_STOP_PENDING, &ihost->flags);
391 spin_lock_irq(&ihost->scic_lock);
392 scic_controller_stop(scic, SCIC_CONTROLLER_STOP_TIMEOUT);
393 spin_unlock_irq(&ihost->scic_lock);
395 wait_for_stop(ihost);
396 scic_controller_reset(scic);
397 isci_timer_list_destroy(ihost);
400 static void __iomem *scu_base(struct isci_host *isci_host)
402 struct pci_dev *pdev = isci_host->pdev;
403 int id = isci_host->id;
405 return pcim_iomap_table(pdev)[SCI_SCU_BAR * 2] + SCI_SCU_BAR_SIZE * id;
408 static void __iomem *smu_base(struct isci_host *isci_host)
410 struct pci_dev *pdev = isci_host->pdev;
411 int id = isci_host->id;
413 return pcim_iomap_table(pdev)[SCI_SMU_BAR * 2] + SCI_SMU_BAR_SIZE * id;
416 static void isci_user_parameters_get(
417 struct isci_host *isci_host,
418 union scic_user_parameters *scic_user_params)
420 struct scic_sds_user_parameters *u = &scic_user_params->sds1;
423 for (i = 0; i < SCI_MAX_PHYS; i++) {
424 struct sci_phy_user_params *u_phy = &u->phys[i];
426 u_phy->max_speed_generation = phy_gen;
428 /* we are not exporting these for now */
429 u_phy->align_insertion_frequency = 0x7f;
430 u_phy->in_connection_align_insertion_frequency = 0xff;
431 u_phy->notify_enable_spin_up_insertion_frequency = 0x33;
434 u->stp_inactivity_timeout = stp_inactive_to;
435 u->ssp_inactivity_timeout = ssp_inactive_to;
436 u->stp_max_occupancy_timeout = stp_max_occ_to;
437 u->ssp_max_occupancy_timeout = ssp_max_occ_to;
438 u->no_outbound_task_timeout = no_outbound_task_to;
439 u->max_number_concurrent_device_spin_up = max_concurr_spinup;
442 int isci_host_init(struct isci_host *isci_host)
445 enum sci_status status;
446 struct scic_sds_controller *controller;
447 union scic_oem_parameters oem;
448 union scic_user_parameters scic_user_params;
449 struct isci_pci_info *pci_info = to_pci_info(isci_host->pdev);
451 isci_timer_list_construct(isci_host);
453 controller = scic_controller_alloc(&isci_host->pdev->dev);
456 dev_err(&isci_host->pdev->dev,
463 isci_host->core_controller = controller;
464 sci_object_set_association(isci_host->core_controller, isci_host);
465 spin_lock_init(&isci_host->state_lock);
466 spin_lock_init(&isci_host->scic_lock);
467 spin_lock_init(&isci_host->queue_lock);
468 init_waitqueue_head(&isci_host->eventq);
470 isci_host_change_state(isci_host, isci_starting);
471 isci_host->can_queue = ISCI_CAN_QUEUE_VAL;
473 status = scic_controller_construct(controller, scu_base(isci_host),
474 smu_base(isci_host));
476 if (status != SCI_SUCCESS) {
477 dev_err(&isci_host->pdev->dev,
478 "%s: scic_controller_construct failed - status = %x\n",
484 isci_host->sas_ha.dev = &isci_host->pdev->dev;
485 isci_host->sas_ha.lldd_ha = isci_host;
488 * grab initial values stored in the controller object for OEM and USER
491 isci_user_parameters_get(isci_host, &scic_user_params);
492 status = scic_user_parameters_set(isci_host->core_controller,
494 if (status != SCI_SUCCESS) {
495 dev_warn(&isci_host->pdev->dev,
496 "%s: scic_user_parameters_set failed\n",
501 scic_oem_parameters_get(controller, &oem);
503 /* grab any OEM parameters specified in orom */
504 if (pci_info->orom) {
505 status = isci_parse_oem_parameters(&oem,
508 if (status != SCI_SUCCESS) {
509 dev_warn(&isci_host->pdev->dev,
510 "parsing firmware oem parameters failed\n");
515 status = scic_oem_parameters_set(isci_host->core_controller, &oem);
516 if (status != SCI_SUCCESS) {
517 dev_warn(&isci_host->pdev->dev,
518 "%s: scic_oem_parameters_set failed\n",
523 tasklet_init(&isci_host->completion_tasklet,
524 isci_host_completion_routine, (unsigned long)isci_host);
526 INIT_LIST_HEAD(&(isci_host->mdl_struct_list));
528 INIT_LIST_HEAD(&isci_host->requests_to_complete);
529 INIT_LIST_HEAD(&isci_host->requests_to_errorback);
531 spin_lock_irq(&isci_host->scic_lock);
532 status = scic_controller_initialize(isci_host->core_controller);
533 spin_unlock_irq(&isci_host->scic_lock);
534 if (status != SCI_SUCCESS) {
535 dev_warn(&isci_host->pdev->dev,
536 "%s: scic_controller_initialize failed -"
542 /* populate mdl with dma memory. scu_mdl_allocate_coherent() */
543 err = isci_host_mdl_allocate_coherent(isci_host);
548 * keep the pool alloc size around, will use it for a bounds checking
549 * when trying to convert virtual addresses to physical addresses
551 isci_host->dma_pool_alloc_size = sizeof(struct isci_request) +
552 scic_io_request_get_object_size();
553 isci_host->dma_pool = dmam_pool_create(DRV_NAME, &isci_host->pdev->dev,
554 isci_host->dma_pool_alloc_size,
555 SLAB_HWCACHE_ALIGN, 0);
557 if (!isci_host->dma_pool)
560 for (i = 0; i < SCI_MAX_PORTS; i++)
561 isci_port_init(&isci_host->isci_ports[i], isci_host, i);
563 for (i = 0; i < SCI_MAX_PHYS; i++)
564 isci_phy_init(&isci_host->phys[i], isci_host, i);
566 for (i = 0; i < SCI_MAX_REMOTE_DEVICES; i++) {
567 struct isci_remote_device *idev = idev_by_id(isci_host, i);
569 INIT_LIST_HEAD(&idev->reqs_in_process);
570 INIT_LIST_HEAD(&idev->node);
571 spin_lock_init(&idev->state_lock);