2 * This file is provided under a dual BSD/GPLv2 license. When using or
3 * redistributing this file, you may do so under either license.
7 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of version 2 of the GNU General Public License as
11 * published by the Free Software Foundation.
13 * This program is distributed in the hope that it will be useful, but
14 * WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * General Public License for more details.
18 * You should have received a copy of the GNU General Public License
19 * along with this program; if not, write to the Free Software
20 * Foundation, Inc., 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
21 * The full GNU General Public License is included in this distribution
22 * in the file called LICENSE.GPL.
26 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
27 * All rights reserved.
29 * Redistribution and use in source and binary forms, with or without
30 * modification, are permitted provided that the following conditions
33 * * Redistributions of source code must retain the above copyright
34 * notice, this list of conditions and the following disclaimer.
35 * * Redistributions in binary form must reproduce the above copyright
36 * notice, this list of conditions and the following disclaimer in
37 * the documentation and/or other materials provided with the
39 * * Neither the name of Intel Corporation nor the names of its
40 * contributors may be used to endorse or promote products derived
41 * from this software without specific prior written permission.
43 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
44 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
45 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
46 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
47 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
48 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
49 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
50 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
51 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
52 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
53 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
56 #ifndef _ISCI_REQUEST_H_
57 #define _ISCI_REQUEST_H_
61 #include "scu_task_context.h"
64 * struct isci_request_status - This enum defines the possible states of an I/O
69 enum isci_request_status {
85 enum sci_request_protocol {
90 }; /* XXX remove me, use sas_task.{dev|task_proto} instead */;
92 struct scic_sds_stp_request {
98 struct scic_sds_stp_pio_request {
100 * Total transfer for the entire PIO request recorded at request constuction
103 * @todo Should we just decrement this value for each byte of data transitted
104 * or received to elemenate the current_transfer_bytes field?
106 u32 total_transfer_bytes;
109 * Total number of bytes received/transmitted in data frames since the start
110 * of the IO request. At the end of the IO request this should equal the
111 * total_transfer_bytes.
113 u32 current_transfer_bytes;
116 * The number of bytes requested in the in the PIO setup.
118 u32 pio_transfer_bytes;
121 * PIO Setup ending status value to tell us if we need to wait for another FIS
122 * or if the transfer is complete. On the receipt of a D2H FIS this will be
123 * the status field of that FIS.
128 * On receipt of a D2H FIS this will be the ending error field if the
129 * ending_status has the SATA_STATUS_ERR bit set.
133 struct scic_sds_request_pio_sgl {
134 struct scu_sgl_element_pair *sgl_pair;
142 * The number of bytes requested in the PIO setup before CDB data frame.
144 u32 device_preferred_cdb_length;
149 struct scic_sds_request {
151 * This field contains the information for the base request state machine.
153 struct sci_base_state_machine state_machine;
156 * This field simply points to the controller to which this IO request
159 struct scic_sds_controller *owning_controller;
162 * This field simply points to the remote device to which this IO request
165 struct scic_sds_remote_device *target_device;
168 * This field is utilized to determine if the SCI user is managing
169 * the IO tag for this request or if the core is managing it.
171 bool was_tag_assigned_by_user;
174 * This field indicates the IO tag for this request. The IO tag is
175 * comprised of the task_index and a sequence count. The sequence count
176 * is utilized to help identify tasks from one life to another.
181 * This field specifies the protocol being utilized for this
184 enum sci_request_protocol protocol;
187 * This field indicates the completion status taken from the SCUs
188 * completion code. It indicates the completion result for the SCU hardware.
193 * This field indicates the completion status returned to the SCI user. It
194 * indicates the users view of the io request completion.
199 * This field contains the value to be utilized when posting (e.g. Post_TC,
200 * Post_TC_Abort) this request to the silicon.
204 struct scu_task_context *task_context_buffer;
205 struct scu_task_context tc ____cacheline_aligned;
207 /* could be larger with sg chaining */
208 #define SCU_SGL_SIZE ((SCU_IO_REQUEST_SGE_COUNT + 1) / 2)
209 struct scu_sgl_element_pair sg_table[SCU_SGL_SIZE] __attribute__ ((aligned(32)));
212 * This field indicates if this request is a task management request or
215 bool is_task_management_request;
218 * This field is a pointer to the stored rx frame data. It is used in STP
219 * internal requests and SMP response frames. If this field is non-NULL the
220 * saved frame must be released on IO request completion.
222 * @todo In the future do we want to keep a list of RX frame buffers?
224 u32 saved_rx_frame_index;
227 * This field specifies the current state handlers in place for this
228 * IO Request object. This field is updated each time the request
231 const struct scic_sds_io_request_state_handler *state_handlers;
234 * This field in the recorded device sequence for the io request. This is
235 * recorded during the build operation and is compared in the start
236 * operation. If the sequence is different then there was a change of
237 * devices from the build to start operations.
244 struct ssp_cmd_iu cmd;
245 struct ssp_task_iu tmf;
248 struct ssp_response_iu rsp;
249 u8 rsp_buf[SSP_RESP_IU_MAX_SIZE];
259 struct scic_sds_stp_request req;
260 struct host_to_dev_fis cmd;
261 struct dev_to_host_fis rsp;
267 static inline struct scic_sds_request *to_sci_req(struct scic_sds_stp_request *stp_req)
269 struct scic_sds_request *sci_req;
271 sci_req = container_of(stp_req, typeof(*sci_req), stp.req);
275 struct isci_request {
276 enum isci_request_status status;
277 enum task_type ttype;
278 unsigned short io_tag;
279 bool complete_in_target;
282 union ttype_ptr_union {
283 struct sas_task *io_task_ptr; /* When ttype==io_task */
284 struct isci_tmf *tmf_task_ptr; /* When ttype==tmf_task */
286 struct isci_host *isci_host;
287 struct isci_remote_device *isci_device;
288 /* For use in the requests_to_{complete|abort} lists: */
289 struct list_head completed_node;
290 /* For use in the reqs_in_process list: */
291 struct list_head dev_node;
292 spinlock_t state_lock;
293 dma_addr_t request_daddr;
294 dma_addr_t zero_scatter_daddr;
296 unsigned int num_sg_entries; /* returned by pci_alloc_sg */
298 /** Note: "io_request_completion" is completed in two different ways
299 * depending on whether this is a TMF or regular request.
300 * - TMF requests are completed in the thread that started them;
301 * - regular requests are completed in the request completion callback
303 * This difference in operation allows the aborter of a TMF request
304 * to be sure that once the TMF request completes, the I/O that the
305 * TMF was aborting is guaranteed to have completed.
307 struct completion *io_request_completion;
308 struct scic_sds_request sci;
311 static inline struct isci_request *sci_req_to_ireq(struct scic_sds_request *sci_req)
313 struct isci_request *ireq = container_of(sci_req, typeof(*ireq), sci);
319 * enum sci_base_request_states - This enumeration depicts all the states for
320 * the common request state machine.
324 enum sci_base_request_states {
326 * Simply the initial state for the base request state machine.
328 SCI_BASE_REQUEST_STATE_INITIAL,
331 * This state indicates that the request has been constructed. This state
332 * is entered from the INITIAL state.
334 SCI_BASE_REQUEST_STATE_CONSTRUCTED,
337 * This state indicates that the request has been started. This state is
338 * entered from the CONSTRUCTED state.
340 SCI_BASE_REQUEST_STATE_STARTED,
342 SCIC_SDS_STP_REQUEST_STARTED_UDMA_AWAIT_TC_COMPLETION_SUBSTATE,
343 SCIC_SDS_STP_REQUEST_STARTED_UDMA_AWAIT_D2H_REG_FIS_SUBSTATE,
345 SCIC_SDS_STP_REQUEST_STARTED_NON_DATA_AWAIT_H2D_COMPLETION_SUBSTATE,
346 SCIC_SDS_STP_REQUEST_STARTED_NON_DATA_AWAIT_D2H_SUBSTATE,
348 SCIC_SDS_STP_REQUEST_STARTED_SOFT_RESET_AWAIT_H2D_ASSERTED_COMPLETION_SUBSTATE,
349 SCIC_SDS_STP_REQUEST_STARTED_SOFT_RESET_AWAIT_H2D_DIAGNOSTIC_COMPLETION_SUBSTATE,
350 SCIC_SDS_STP_REQUEST_STARTED_SOFT_RESET_AWAIT_D2H_RESPONSE_FRAME_SUBSTATE,
353 * While in this state the IO request object is waiting for the TC completion
354 * notification for the H2D Register FIS
356 SCIC_SDS_STP_REQUEST_STARTED_PIO_AWAIT_H2D_COMPLETION_SUBSTATE,
359 * While in this state the IO request object is waiting for either a PIO Setup
360 * FIS or a D2H register FIS. The type of frame received is based on the
361 * result of the prior frame and line conditions.
363 SCIC_SDS_STP_REQUEST_STARTED_PIO_AWAIT_FRAME_SUBSTATE,
366 * While in this state the IO request object is waiting for a DATA frame from
369 SCIC_SDS_STP_REQUEST_STARTED_PIO_DATA_IN_AWAIT_DATA_SUBSTATE,
372 * While in this state the IO request object is waiting to transmit the next data
373 * frame to the device.
375 SCIC_SDS_STP_REQUEST_STARTED_PIO_DATA_OUT_TRANSMIT_DATA_SUBSTATE,
378 * The AWAIT_TC_COMPLETION sub-state indicates that the started raw
379 * task management request is waiting for the transmission of the
380 * initial frame (i.e. command, task, etc.).
382 SCIC_SDS_IO_REQUEST_STARTED_TASK_MGMT_SUBSTATE_AWAIT_TC_COMPLETION,
385 * This sub-state indicates that the started task management request
386 * is waiting for the reception of an unsolicited frame
387 * (i.e. response IU).
389 SCIC_SDS_IO_REQUEST_STARTED_TASK_MGMT_SUBSTATE_AWAIT_TC_RESPONSE,
392 * This sub-state indicates that the started task management request
393 * is waiting for the reception of an unsolicited frame
394 * (i.e. response IU).
396 SCIC_SDS_SMP_REQUEST_STARTED_SUBSTATE_AWAIT_RESPONSE,
399 * The AWAIT_TC_COMPLETION sub-state indicates that the started SMP request is
400 * waiting for the transmission of the initial frame (i.e. command, task, etc.).
402 SCIC_SDS_SMP_REQUEST_STARTED_SUBSTATE_AWAIT_TC_COMPLETION,
405 * This state indicates that the request has completed.
406 * This state is entered from the STARTED state. This state is entered from
407 * the ABORTING state.
409 SCI_BASE_REQUEST_STATE_COMPLETED,
412 * This state indicates that the request is in the process of being
413 * terminated/aborted.
414 * This state is entered from the CONSTRUCTED state.
415 * This state is entered from the STARTED state.
417 SCI_BASE_REQUEST_STATE_ABORTING,
420 * Simply the final state for the base request state machine.
422 SCI_BASE_REQUEST_STATE_FINAL,
425 typedef enum sci_status (*scic_sds_io_request_handler_t)
426 (struct scic_sds_request *request);
427 typedef enum sci_status (*scic_sds_io_request_frame_handler_t)
428 (struct scic_sds_request *req, u32 frame);
429 typedef enum sci_status (*scic_sds_io_request_event_handler_t)
430 (struct scic_sds_request *req, u32 event);
431 typedef enum sci_status (*scic_sds_io_request_task_completion_handler_t)
432 (struct scic_sds_request *req, u32 completion_code);
435 * struct scic_sds_io_request_state_handler - This is the SDS core definition
436 * of the state handlers.
440 struct scic_sds_io_request_state_handler {
442 * The start_handler specifies the method invoked when a user attempts to
445 scic_sds_io_request_handler_t start_handler;
448 * The abort_handler specifies the method invoked when a user attempts to
451 scic_sds_io_request_handler_t abort_handler;
454 * The complete_handler specifies the method invoked when a user attempts to
455 * complete a request.
457 scic_sds_io_request_handler_t complete_handler;
459 scic_sds_io_request_task_completion_handler_t tc_completion_handler;
460 scic_sds_io_request_event_handler_t event_handler;
461 scic_sds_io_request_frame_handler_t frame_handler;
466 * scic_sds_request_get_controller() -
468 * This macro will return the controller for this io request object
470 #define scic_sds_request_get_controller(sci_req) \
471 ((sci_req)->owning_controller)
474 * scic_sds_request_get_device() -
476 * This macro will return the device for this io request object
478 #define scic_sds_request_get_device(sci_req) \
479 ((sci_req)->target_device)
482 * scic_sds_request_get_port() -
484 * This macro will return the port for this io request object
486 #define scic_sds_request_get_port(sci_req) \
487 scic_sds_remote_device_get_port(scic_sds_request_get_device(sci_req))
490 * scic_sds_request_get_post_context() -
492 * This macro returns the constructed post context result for the io request.
494 #define scic_sds_request_get_post_context(sci_req) \
495 ((sci_req)->post_context)
498 * scic_sds_request_get_task_context() -
500 * This is a helper macro to return the os handle for this request object.
502 #define scic_sds_request_get_task_context(request) \
503 ((request)->task_context_buffer)
506 * scic_sds_request_set_status() -
508 * This macro will set the scu hardware status and sci request completion
509 * status for an io request.
511 #define scic_sds_request_set_status(request, scu_status_code, sci_status_code) \
513 (request)->scu_status = (scu_status_code); \
514 (request)->sci_status = (sci_status_code); \
517 #define scic_sds_request_complete(a_request) \
518 ((a_request)->state_handlers->complete_handler(a_request))
521 extern enum sci_status
522 scic_sds_io_request_tc_completion(struct scic_sds_request *request, u32 completion_code);
527 * This macro zeros the hardware SGL element data
529 #define SCU_SGL_ZERO(scu_sge) \
531 (scu_sge).length = 0; \
532 (scu_sge).address_lower = 0; \
533 (scu_sge).address_upper = 0; \
534 (scu_sge).address_modifier = 0; \
540 * This macro copys the SGL Element data from the host os to the hardware SGL
543 #define SCU_SGL_COPY(scu_sge, os_sge) \
545 (scu_sge).length = sg_dma_len(sg); \
546 (scu_sge).address_upper = \
547 upper_32_bits(sg_dma_address(sg)); \
548 (scu_sge).address_lower = \
549 lower_32_bits(sg_dma_address(sg)); \
550 (scu_sge).address_modifier = 0; \
553 enum sci_status scic_sds_request_start(struct scic_sds_request *sci_req);
554 enum sci_status scic_sds_io_request_terminate(struct scic_sds_request *sci_req);
555 enum sci_status scic_sds_io_request_event_handler(struct scic_sds_request *sci_req,
557 enum sci_status scic_sds_io_request_frame_handler(struct scic_sds_request *sci_req,
559 enum sci_status scic_sds_task_request_terminate(struct scic_sds_request *sci_req);
561 /* XXX open code in caller */
562 static inline void *scic_request_get_virt_addr(struct scic_sds_request *sci_req,
563 dma_addr_t phys_addr)
565 struct isci_request *ireq = sci_req_to_ireq(sci_req);
568 BUG_ON(phys_addr < ireq->request_daddr);
570 offset = phys_addr - ireq->request_daddr;
572 BUG_ON(offset >= sizeof(*ireq));
574 return (char *)ireq + offset;
577 /* XXX open code in caller */
578 static inline dma_addr_t scic_io_request_get_dma_addr(struct scic_sds_request *sci_req,
581 struct isci_request *ireq = sci_req_to_ireq(sci_req);
583 char *requested_addr = (char *)virt_addr;
584 char *base_addr = (char *)ireq;
586 BUG_ON(requested_addr < base_addr);
587 BUG_ON((requested_addr - base_addr) >= sizeof(*ireq));
589 return ireq->request_daddr + (requested_addr - base_addr);
593 * This function gets the status of the request object.
594 * @request: This parameter points to the isci_request object
596 * status of the object as a isci_request_status enum.
599 enum isci_request_status isci_request_get_state(
600 struct isci_request *isci_request)
602 BUG_ON(isci_request == NULL);
604 /*probably a bad sign... */
605 if (isci_request->status == unallocated)
606 dev_warn(&isci_request->isci_host->pdev->dev,
607 "%s: isci_request->status == unallocated\n",
610 return isci_request->status;
615 * isci_request_change_state() - This function sets the status of the request
617 * @request: This parameter points to the isci_request object
618 * @status: This Parameter is the new status of the object
621 static inline enum isci_request_status isci_request_change_state(
622 struct isci_request *isci_request,
623 enum isci_request_status status)
625 enum isci_request_status old_state;
628 dev_dbg(&isci_request->isci_host->pdev->dev,
629 "%s: isci_request = %p, state = 0x%x\n",
634 BUG_ON(isci_request == NULL);
636 spin_lock_irqsave(&isci_request->state_lock, flags);
637 old_state = isci_request->status;
638 isci_request->status = status;
639 spin_unlock_irqrestore(&isci_request->state_lock, flags);
645 * isci_request_change_started_to_newstate() - This function sets the status of
646 * the request object.
647 * @request: This parameter points to the isci_request object
648 * @status: This Parameter is the new status of the object
650 * state previous to any change.
652 static inline enum isci_request_status isci_request_change_started_to_newstate(
653 struct isci_request *isci_request,
654 struct completion *completion_ptr,
655 enum isci_request_status newstate)
657 enum isci_request_status old_state;
660 spin_lock_irqsave(&isci_request->state_lock, flags);
662 old_state = isci_request->status;
664 if (old_state == started || old_state == aborting) {
665 BUG_ON(isci_request->io_request_completion != NULL);
667 isci_request->io_request_completion = completion_ptr;
668 isci_request->status = newstate;
670 spin_unlock_irqrestore(&isci_request->state_lock, flags);
672 dev_dbg(&isci_request->isci_host->pdev->dev,
673 "%s: isci_request = %p, old_state = 0x%x\n",
682 * isci_request_change_started_to_aborted() - This function sets the status of
683 * the request object.
684 * @request: This parameter points to the isci_request object
685 * @completion_ptr: This parameter is saved as the kernel completion structure
686 * signalled when the old request completes.
688 * state previous to any change.
690 static inline enum isci_request_status isci_request_change_started_to_aborted(
691 struct isci_request *isci_request,
692 struct completion *completion_ptr)
694 return isci_request_change_started_to_newstate(
695 isci_request, completion_ptr, aborted
699 * isci_request_free() - This function frees the request object.
700 * @isci_host: This parameter specifies the ISCI host object
701 * @isci_request: This parameter points to the isci_request object
704 static inline void isci_request_free(
705 struct isci_host *isci_host,
706 struct isci_request *isci_request)
711 /* release the dma memory if we fail. */
712 dma_pool_free(isci_host->dma_pool, isci_request,
713 isci_request->request_daddr);
717 /* #define ISCI_REQUEST_VALIDATE_ACCESS
720 #ifdef ISCI_REQUEST_VALIDATE_ACCESS
723 struct sas_task *isci_request_access_task(struct isci_request *isci_request)
725 BUG_ON(isci_request->ttype != io_task);
726 return isci_request->ttype_ptr.io_task_ptr;
730 struct isci_tmf *isci_request_access_tmf(struct isci_request *isci_request)
732 BUG_ON(isci_request->ttype != tmf_task);
733 return isci_request->ttype_ptr.tmf_task_ptr;
736 #else /* not ISCI_REQUEST_VALIDATE_ACCESS */
738 #define isci_request_access_task(RequestPtr) \
739 ((RequestPtr)->ttype_ptr.io_task_ptr)
741 #define isci_request_access_tmf(RequestPtr) \
742 ((RequestPtr)->ttype_ptr.tmf_task_ptr)
744 #endif /* not ISCI_REQUEST_VALIDATE_ACCESS */
747 int isci_request_alloc_tmf(
748 struct isci_host *isci_host,
749 struct isci_tmf *isci_tmf,
750 struct isci_request **isci_request,
751 struct isci_remote_device *isci_device,
755 int isci_request_execute(
756 struct isci_host *isci_host,
757 struct sas_task *task,
758 struct isci_request **request,
762 * isci_request_unmap_sgl() - This function unmaps the DMA address of a given
764 * @request: This parameter points to the isci_request object
765 * @*pdev: This Parameter is the pci_device struct for the controller
768 static inline void isci_request_unmap_sgl(
769 struct isci_request *request,
770 struct pci_dev *pdev)
772 struct sas_task *task = isci_request_access_task(request);
774 dev_dbg(&request->isci_host->pdev->dev,
775 "%s: request = %p, task = %p,\n"
776 "task->data_dir = %d, is_sata = %d\n ",
781 sas_protocol_ata(task->task_proto));
783 if ((task->data_dir != PCI_DMA_NONE) &&
784 !sas_protocol_ata(task->task_proto)) {
785 if (task->num_scatter == 0)
786 /* 0 indicates a single dma address */
789 request->zero_scatter_daddr,
790 task->total_xfer_len,
794 else /* unmap the sgl dma addresses */
798 request->num_sg_entries,
805 * isci_request_io_request_get_next_sge() - This function is called by the sci
806 * core to retrieve the next sge for a given request.
807 * @request: This parameter is the isci_request object.
808 * @current_sge_address: This parameter is the last sge retrieved by the sci
809 * core for this request.
811 * pointer to the next sge for specified request.
813 static inline void *isci_request_io_request_get_next_sge(
814 struct isci_request *request,
815 void *current_sge_address)
817 struct sas_task *task = isci_request_access_task(request);
820 dev_dbg(&request->isci_host->pdev->dev,
822 "current_sge_address = %p, "
823 "num_scatter = %d\n",
829 if (!current_sge_address) /* First time through.. */
830 ret = task->scatter; /* always task->scatter */
831 else if (task->num_scatter == 0) /* Next element, if num_scatter == 0 */
832 ret = NULL; /* there is only one element. */
834 ret = sg_next(current_sge_address); /* sg_next returns NULL
835 * for the last element
838 dev_dbg(&request->isci_host->pdev->dev,
839 "%s: next sge address = %p\n",
846 void isci_terminate_pending_requests(struct isci_host *isci_host,
847 struct isci_remote_device *isci_device,
848 enum isci_request_status new_request_state);
849 enum sci_status scic_task_request_construct(struct scic_sds_controller *scic,
850 struct scic_sds_remote_device *sci_dev,
852 struct scic_sds_request *sci_req);
853 enum sci_status scic_task_request_construct_ssp(struct scic_sds_request *sci_req);
854 enum sci_status scic_task_request_construct_sata(struct scic_sds_request *sci_req);
855 enum sci_status scic_sds_stp_udma_request_construct(struct scic_sds_request *sci_req,
857 enum dma_data_direction dir);
858 void scic_stp_io_request_set_ncq_tag(struct scic_sds_request *sci_req, u16 ncq_tag);
859 void scic_sds_smp_request_copy_response(struct scic_sds_request *sci_req);
860 #endif /* !defined(_ISCI_REQUEST_H_) */