DPDK  23.11.0
rte_cryptodev.h
Go to the documentation of this file.
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2015-2020 Intel Corporation.
3  */
4 
5 #ifndef _RTE_CRYPTODEV_H_
6 #define _RTE_CRYPTODEV_H_
7 
17 #ifdef __cplusplus
18 extern "C" {
19 #endif
20 
21 #include <rte_compat.h>
22 #include "rte_kvargs.h"
23 #include "rte_crypto.h"
24 #include <rte_common.h>
25 #include <rte_rcu_qsbr.h>
26 
27 #include "rte_cryptodev_trace_fp.h"
28 
29 extern const char **rte_cyptodev_names;
30 
31 /* Logging Macros */
32 
33 #define CDEV_LOG_ERR(...) \
34  RTE_LOG(ERR, CRYPTODEV, \
35  RTE_FMT("%s() line %u: " RTE_FMT_HEAD(__VA_ARGS__,) "\n", \
36  __func__, __LINE__, RTE_FMT_TAIL(__VA_ARGS__,)))
37 
38 #define CDEV_LOG_INFO(...) \
39  RTE_LOG(INFO, CRYPTODEV, \
40  RTE_FMT(RTE_FMT_HEAD(__VA_ARGS__,) "\n", \
41  RTE_FMT_TAIL(__VA_ARGS__,)))
42 
43 #define CDEV_LOG_DEBUG(...) \
44  RTE_LOG(DEBUG, CRYPTODEV, \
45  RTE_FMT("%s() line %u: " RTE_FMT_HEAD(__VA_ARGS__,) "\n", \
46  __func__, __LINE__, RTE_FMT_TAIL(__VA_ARGS__,)))
47 
48 #define CDEV_PMD_TRACE(...) \
49  RTE_LOG(DEBUG, CRYPTODEV, \
50  RTE_FMT("[%s] %s: " RTE_FMT_HEAD(__VA_ARGS__,) "\n", \
51  dev, __func__, RTE_FMT_TAIL(__VA_ARGS__,)))
52 
66 #define rte_crypto_op_ctod_offset(c, t, o) \
67  ((t)((char *)(c) + (o)))
68 
80 #define rte_crypto_op_ctophys_offset(c, o) \
81  (rte_iova_t)((c)->phys_addr + (o))
82 
87  uint16_t min;
88  uint16_t max;
89  uint16_t increment;
95 };
96 
102 #define RTE_CRYPTO_CIPHER_DATA_UNIT_LEN_512_BYTES RTE_BIT32(0)
103 #define RTE_CRYPTO_CIPHER_DATA_UNIT_LEN_4096_BYTES RTE_BIT32(1)
104 #define RTE_CRYPTO_CIPHER_DATA_UNIT_LEN_1_MEGABYTES RTE_BIT32(2)
105 
112  union {
113  struct {
116  uint16_t block_size;
126  } auth;
128  struct {
131  uint16_t block_size;
137  uint32_t dataunit_set;
143  } cipher;
145  struct {
148  uint16_t block_size;
158  } aead;
159  };
160 };
161 
169  uint32_t op_types;
178  __extension__
179  union {
185  uint8_t internal_rng;
190  };
191 
192  uint64_t hash_algos;
194 };
195 
201 };
202 
203 
206  enum rte_crypto_op_type op;
209  union {
214  };
215 };
216 
219  enum rte_crypto_sym_xform_type type;
220  union {
221  enum rte_crypto_cipher_algorithm cipher;
222  enum rte_crypto_auth_algorithm auth;
223  enum rte_crypto_aead_algorithm aead;
224  } algo;
225 };
226 
234 };
235 
248  const struct rte_cryptodev_sym_capability_idx *idx);
249 
262  const struct rte_cryptodev_asym_capability_idx *idx);
263 
276 int
278  const struct rte_cryptodev_symmetric_capability *capability,
279  uint16_t key_size, uint16_t iv_size);
280 
294 int
296  const struct rte_cryptodev_symmetric_capability *capability,
297  uint16_t key_size, uint16_t digest_size, uint16_t iv_size);
298 
313 int
315  const struct rte_cryptodev_symmetric_capability *capability,
316  uint16_t key_size, uint16_t digest_size, uint16_t aad_size,
317  uint16_t iv_size);
318 
329 int
331  const struct rte_cryptodev_asymmetric_xform_capability *capability,
332  enum rte_crypto_asym_op_type op_type);
333 
344 int
346  const struct rte_cryptodev_asymmetric_xform_capability *capability,
347  uint16_t modlen);
348 
359 bool
361  const struct rte_cryptodev_asymmetric_xform_capability *capability,
362  enum rte_crypto_auth_algorithm hash);
363 
375 int
377  const char *algo_string);
378 
390 int
392  const char *algo_string);
393 
405 int
407  const char *algo_string);
408 
420 int
422  const char *xform_string);
423 
433 __rte_experimental
434 const char *
436 
446 __rte_experimental
447 const char *
449 
459 __rte_experimental
460 const char *
462 
472 __rte_experimental
473 const char *
475 
476 
478 #define RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() \
479  { RTE_CRYPTO_OP_TYPE_UNDEFINED }
480 
481 
490 #define RTE_CRYPTODEV_FF_SYMMETRIC_CRYPTO (1ULL << 0)
492 #define RTE_CRYPTODEV_FF_ASYMMETRIC_CRYPTO (1ULL << 1)
494 #define RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING (1ULL << 2)
496 #define RTE_CRYPTODEV_FF_CPU_SSE (1ULL << 3)
498 #define RTE_CRYPTODEV_FF_CPU_AVX (1ULL << 4)
500 #define RTE_CRYPTODEV_FF_CPU_AVX2 (1ULL << 5)
502 #define RTE_CRYPTODEV_FF_CPU_AESNI (1ULL << 6)
504 #define RTE_CRYPTODEV_FF_HW_ACCELERATED (1ULL << 7)
508 #define RTE_CRYPTODEV_FF_CPU_AVX512 (1ULL << 8)
510 #define RTE_CRYPTODEV_FF_IN_PLACE_SGL (1ULL << 9)
514 #define RTE_CRYPTODEV_FF_OOP_SGL_IN_SGL_OUT (1ULL << 10)
518 #define RTE_CRYPTODEV_FF_OOP_SGL_IN_LB_OUT (1ULL << 11)
523 #define RTE_CRYPTODEV_FF_OOP_LB_IN_SGL_OUT (1ULL << 12)
527 #define RTE_CRYPTODEV_FF_OOP_LB_IN_LB_OUT (1ULL << 13)
529 #define RTE_CRYPTODEV_FF_CPU_NEON (1ULL << 14)
531 #define RTE_CRYPTODEV_FF_CPU_ARM_CE (1ULL << 15)
533 #define RTE_CRYPTODEV_FF_SECURITY (1ULL << 16)
535 #define RTE_CRYPTODEV_FF_RSA_PRIV_OP_KEY_EXP (1ULL << 17)
537 #define RTE_CRYPTODEV_FF_RSA_PRIV_OP_KEY_QT (1ULL << 18)
539 #define RTE_CRYPTODEV_FF_DIGEST_ENCRYPTED (1ULL << 19)
541 #define RTE_CRYPTODEV_FF_ASYM_SESSIONLESS (1ULL << 20)
543 #define RTE_CRYPTODEV_FF_SYM_CPU_CRYPTO (1ULL << 21)
545 #define RTE_CRYPTODEV_FF_SYM_SESSIONLESS (1ULL << 22)
547 #define RTE_CRYPTODEV_FF_NON_BYTE_ALIGNED_DATA (1ULL << 23)
549 #define RTE_CRYPTODEV_FF_SYM_RAW_DP (1ULL << 24)
551 #define RTE_CRYPTODEV_FF_CIPHER_MULTIPLE_DATA_UNITS (1ULL << 25)
553 #define RTE_CRYPTODEV_FF_CIPHER_WRAPPED_KEY (1ULL << 26)
555 #define RTE_CRYPTODEV_FF_SECURITY_INNER_CSUM (1ULL << 27)
557 #define RTE_CRYPTODEV_FF_SECURITY_RX_INJECT (1ULL << 28)
568 const char *
570 
572 /* Structure rte_cryptodev_info 8< */
574  const char *driver_name;
575  uint8_t driver_id;
576  struct rte_device *device;
578  uint64_t feature_flags;
593  struct {
594  unsigned max_nb_sessions;
599  } sym;
600 };
601 /* >8 End of structure rte_cryptodev_info. */
602 
603 #define RTE_CRYPTODEV_DETACHED (0)
604 #define RTE_CRYPTODEV_ATTACHED (1)
605 
611 };
612 
614 /* Structure rte_cryptodev_qp_conf 8<*/
616  uint32_t nb_descriptors;
619 };
620 /* >8 End of structure rte_cryptodev_qp_conf. */
621 
643 typedef uint16_t (*rte_cryptodev_callback_fn)(uint16_t dev_id, uint16_t qp_id,
644  struct rte_crypto_op **ops, uint16_t nb_ops, void *user_param);
645 
655 typedef void (*rte_cryptodev_cb_fn)(uint8_t dev_id,
656  enum rte_cryptodev_event_type event, void *cb_arg);
657 
658 
661  uint64_t enqueued_count;
663  uint64_t dequeued_count;
670 };
671 
672 #define RTE_CRYPTODEV_NAME_MAX_LEN (64)
684 int
685 rte_cryptodev_get_dev_id(const char *name);
686 
697 const char *
698 rte_cryptodev_name_get(uint8_t dev_id);
699 
707 uint8_t
709 
718 uint8_t
720 
732 uint8_t
733 rte_cryptodev_devices_get(const char *driver_name, uint8_t *devices,
734  uint8_t nb_devices);
735 /*
736  * Return the NUMA socket to which a device is connected
737  *
738  * @param dev_id
739  * The identifier of the device
740  * @return
741  * The NUMA socket id to which the device is connected or
742  * a default of zero if the socket could not be determined.
743  * -1 if returned is the dev_id value is out of range.
744  */
745 int
746 rte_cryptodev_socket_id(uint8_t dev_id);
747 
749 /* Structure rte_cryptodev_config 8< */
751  int socket_id;
752  uint16_t nb_queue_pairs;
754  uint64_t ff_disable;
761 };
762 /* >8 End of structure rte_cryptodev_config. */
763 
778 int
779 rte_cryptodev_configure(uint8_t dev_id, struct rte_cryptodev_config *config);
780 
796 int
797 rte_cryptodev_start(uint8_t dev_id);
798 
805 void
806 rte_cryptodev_stop(uint8_t dev_id);
807 
817 int
818 rte_cryptodev_close(uint8_t dev_id);
819 
841 int
842 rte_cryptodev_queue_pair_setup(uint8_t dev_id, uint16_t queue_pair_id,
843  const struct rte_cryptodev_qp_conf *qp_conf, int socket_id);
844 
858 int
859 rte_cryptodev_get_qp_status(uint8_t dev_id, uint16_t queue_pair_id);
860 
868 uint16_t
870 
871 
883 int
884 rte_cryptodev_stats_get(uint8_t dev_id, struct rte_cryptodev_stats *stats);
885 
891 void
893 
907 void
908 rte_cryptodev_info_get(uint8_t dev_id, struct rte_cryptodev_info *dev_info);
909 
910 
924 int
926  enum rte_cryptodev_event_type event,
927  rte_cryptodev_cb_fn cb_fn, void *cb_arg);
928 
942 int
944  enum rte_cryptodev_event_type event,
945  rte_cryptodev_cb_fn cb_fn, void *cb_arg);
946 
962 __rte_experimental
963 int
964 rte_cryptodev_queue_pair_event_error_query(uint8_t dev_id, uint16_t qp_id);
965 
966 struct rte_cryptodev_callback;
967 
969 RTE_TAILQ_HEAD(rte_cryptodev_cb_list, rte_cryptodev_callback);
970 
976  RTE_ATOMIC(struct rte_cryptodev_cb *) next;
980  void *arg;
982 };
983 
988 struct rte_cryptodev_cb_rcu {
989  RTE_ATOMIC(struct rte_cryptodev_cb *) next;
991  struct rte_rcu_qsbr *qsbr;
993 };
994 
1004 void *
1006 
1036 struct rte_mempool *
1037 rte_cryptodev_sym_session_pool_create(const char *name, uint32_t nb_elts,
1038  uint32_t elt_size, uint32_t cache_size, uint16_t priv_size,
1039  int socket_id);
1040 
1041 
1062 struct rte_mempool *
1063 rte_cryptodev_asym_session_pool_create(const char *name, uint32_t nb_elts,
1064  uint32_t cache_size, uint16_t user_data_size, int socket_id);
1065 
1082 void *
1084  struct rte_crypto_sym_xform *xforms,
1085  struct rte_mempool *mp);
1103 int
1105  struct rte_crypto_asym_xform *xforms, struct rte_mempool *mp,
1106  void **session);
1107 
1120 int
1122  void *sess);
1123 
1135 int
1136 rte_cryptodev_asym_session_free(uint8_t dev_id, void *sess);
1137 
1144 unsigned int
1146 
1158 unsigned int
1160 
1171 unsigned int
1173 
1182 unsigned int
1184 
1194 
1203 const char *rte_cryptodev_driver_name_get(uint8_t driver_id);
1204 
1217 int
1219  void *data,
1220  uint16_t size);
1221 
1222 #define CRYPTO_SESS_OPAQUE_DATA_OFF 0
1226 static inline uint64_t
1228 {
1229  return *((uint64_t *)sess + CRYPTO_SESS_OPAQUE_DATA_OFF);
1230 }
1231 
1235 static inline void
1236 rte_cryptodev_sym_session_opaque_data_set(void *sess, uint64_t opaque)
1237 {
1238  uint64_t *data;
1239  data = (((uint64_t *)sess) + CRYPTO_SESS_OPAQUE_DATA_OFF);
1240  *data = opaque;
1241 }
1242 
1253 void *
1255 
1269 int
1270 rte_cryptodev_asym_session_set_user_data(void *sess, void *data, uint16_t size);
1271 
1282 void *
1284 
1297 uint32_t
1299  void *sess, union rte_crypto_sym_ofs ofs,
1300  struct rte_crypto_sym_vec *vec);
1301 
1311 int
1313 
1329 int
1330 rte_cryptodev_session_event_mdata_set(uint8_t dev_id, void *sess,
1331  enum rte_crypto_op_type op_type,
1332  enum rte_crypto_op_sess_type sess_type,
1333  void *ev_mdata, uint16_t size);
1334 
1339 union rte_cryptodev_session_ctx {void *crypto_sess;
1340  struct rte_crypto_sym_xform *xform;
1341  struct rte_security_session *sec_sess;
1342 };
1343 
1370  void *qp, uint8_t *drv_ctx, struct rte_crypto_sym_vec *vec,
1371  union rte_crypto_sym_ofs ofs, void *user_data[], int *enqueue_status);
1372 
1395  void *qp, uint8_t *drv_ctx, struct rte_crypto_vec *data_vec,
1396  uint16_t n_data_vecs, union rte_crypto_sym_ofs ofs,
1397  struct rte_crypto_va_iova_ptr *iv,
1398  struct rte_crypto_va_iova_ptr *digest,
1399  struct rte_crypto_va_iova_ptr *aad_or_auth_iv,
1400  void *user_data);
1401 
1413 typedef int (*cryptodev_sym_raw_operation_done_t)(void *qp, uint8_t *drv_ctx,
1414  uint32_t n);
1415 
1425 typedef uint32_t (*rte_cryptodev_raw_get_dequeue_count_t)(void *user_data);
1426 
1435 typedef void (*rte_cryptodev_raw_post_dequeue_t)(void *user_data,
1436  uint32_t index, uint8_t is_op_success);
1437 
1479 typedef uint32_t (*cryptodev_sym_raw_dequeue_burst_t)(void *qp,
1480  uint8_t *drv_ctx,
1481  rte_cryptodev_raw_get_dequeue_count_t get_dequeue_count,
1482  uint32_t max_nb_to_dequeue,
1483  rte_cryptodev_raw_post_dequeue_t post_dequeue,
1484  void **out_user_data, uint8_t is_user_data_array,
1485  uint32_t *n_success, int *dequeue_status);
1486 
1510 typedef void * (*cryptodev_sym_raw_dequeue_t)(
1511  void *qp, uint8_t *drv_ctx, int *dequeue_status,
1512  enum rte_crypto_op_status *op_status);
1513 
1520  void *qp_data;
1521 
1523  cryptodev_sym_raw_enqueue_burst_t enqueue_burst;
1526  cryptodev_sym_raw_dequeue_burst_t dequeue_burst;
1528 
1529  /* Driver specific context data */
1530  __extension__ uint8_t drv_ctx_data[];
1531 };
1532 
1554 int
1555 rte_cryptodev_configure_raw_dp_ctx(uint8_t dev_id, uint16_t qp_id,
1556  struct rte_crypto_raw_dp_ctx *ctx,
1557  enum rte_crypto_op_sess_type sess_type,
1558  union rte_cryptodev_session_ctx session_ctx,
1559  uint8_t is_update);
1560 
1585 uint32_t
1587  struct rte_crypto_sym_vec *vec, union rte_crypto_sym_ofs ofs,
1588  void **user_data, int *enqueue_status);
1589 
1610 __rte_experimental
1611 static __rte_always_inline int
1613  struct rte_crypto_vec *data_vec, uint16_t n_data_vecs,
1614  union rte_crypto_sym_ofs ofs,
1615  struct rte_crypto_va_iova_ptr *iv,
1616  struct rte_crypto_va_iova_ptr *digest,
1617  struct rte_crypto_va_iova_ptr *aad_or_auth_iv,
1618  void *user_data)
1619 {
1620  return (*ctx->enqueue)(ctx->qp_data, ctx->drv_ctx_data, data_vec,
1621  n_data_vecs, ofs, iv, digest, aad_or_auth_iv, user_data);
1622 }
1623 
1634 int
1636  uint32_t n);
1637 
1679 uint32_t
1681  rte_cryptodev_raw_get_dequeue_count_t get_dequeue_count,
1682  uint32_t max_nb_to_dequeue,
1683  rte_cryptodev_raw_post_dequeue_t post_dequeue,
1684  void **out_user_data, uint8_t is_user_data_array,
1685  uint32_t *n_success, int *dequeue_status);
1686 
1710 __rte_experimental
1711 static __rte_always_inline void *
1713  int *dequeue_status, enum rte_crypto_op_status *op_status)
1714 {
1715  return (*ctx->dequeue)(ctx->qp_data, ctx->drv_ctx_data, dequeue_status,
1716  op_status);
1717 }
1718 
1728 int
1730  uint32_t n);
1731 
1767 struct rte_cryptodev_cb *
1769  uint16_t qp_id,
1771  void *cb_arg);
1772 
1795  uint16_t qp_id,
1796  struct rte_cryptodev_cb *cb);
1797 
1832 struct rte_cryptodev_cb *
1834  uint16_t qp_id,
1836  void *cb_arg);
1837 
1860  uint16_t qp_id,
1861  struct rte_cryptodev_cb *cb);
1862 
1863 #include <rte_cryptodev_core.h>
1900 static inline uint16_t
1901 rte_cryptodev_dequeue_burst(uint8_t dev_id, uint16_t qp_id,
1902  struct rte_crypto_op **ops, uint16_t nb_ops)
1903 {
1904  const struct rte_crypto_fp_ops *fp_ops;
1905  void *qp;
1906 
1907  rte_cryptodev_trace_dequeue_burst(dev_id, qp_id, (void **)ops, nb_ops);
1908 
1909  fp_ops = &rte_crypto_fp_ops[dev_id];
1910  qp = fp_ops->qp.data[qp_id];
1911 
1912  nb_ops = fp_ops->dequeue_burst(qp, ops, nb_ops);
1913 
1914 #ifdef RTE_CRYPTO_CALLBACKS
1915  if (unlikely(fp_ops->qp.deq_cb != NULL)) {
1916  struct rte_cryptodev_cb_rcu *list;
1917  struct rte_cryptodev_cb *cb;
1918 
1919  /* rte_memory_order_release memory order was used when the
1920  * call back was inserted into the list.
1921  * Since there is a clear dependency between loading
1922  * cb and cb->fn/cb->next, rte_memory_order_acquire memory order is
1923  * not required.
1924  */
1925  list = &fp_ops->qp.deq_cb[qp_id];
1926  rte_rcu_qsbr_thread_online(list->qsbr, 0);
1927  cb = rte_atomic_load_explicit(&list->next, rte_memory_order_relaxed);
1928 
1929  while (cb != NULL) {
1930  nb_ops = cb->fn(dev_id, qp_id, ops, nb_ops,
1931  cb->arg);
1932  cb = cb->next;
1933  };
1934 
1935  rte_rcu_qsbr_thread_offline(list->qsbr, 0);
1936  }
1937 #endif
1938  return nb_ops;
1939 }
1940 
1972 static inline uint16_t
1973 rte_cryptodev_enqueue_burst(uint8_t dev_id, uint16_t qp_id,
1974  struct rte_crypto_op **ops, uint16_t nb_ops)
1975 {
1976  const struct rte_crypto_fp_ops *fp_ops;
1977  void *qp;
1978 
1979  fp_ops = &rte_crypto_fp_ops[dev_id];
1980  qp = fp_ops->qp.data[qp_id];
1981 #ifdef RTE_CRYPTO_CALLBACKS
1982  if (unlikely(fp_ops->qp.enq_cb != NULL)) {
1983  struct rte_cryptodev_cb_rcu *list;
1984  struct rte_cryptodev_cb *cb;
1985 
1986  /* rte_memory_order_release memory order was used when the
1987  * call back was inserted into the list.
1988  * Since there is a clear dependency between loading
1989  * cb and cb->fn/cb->next, rte_memory_order_acquire memory order is
1990  * not required.
1991  */
1992  list = &fp_ops->qp.enq_cb[qp_id];
1993  rte_rcu_qsbr_thread_online(list->qsbr, 0);
1994  cb = rte_atomic_load_explicit(&list->next, rte_memory_order_relaxed);
1995 
1996  while (cb != NULL) {
1997  nb_ops = cb->fn(dev_id, qp_id, ops, nb_ops,
1998  cb->arg);
1999  cb = cb->next;
2000  };
2001 
2002  rte_rcu_qsbr_thread_offline(list->qsbr, 0);
2003  }
2004 #endif
2005 
2006  rte_cryptodev_trace_enqueue_burst(dev_id, qp_id, (void **)ops, nb_ops);
2007  return fp_ops->enqueue_burst(qp, ops, nb_ops);
2008 }
2009 
2010 
2011 
2012 #ifdef __cplusplus
2013 }
2014 #endif
2015 
2016 #endif /* _RTE_CRYPTODEV_H_ */
#define unlikely(x)
#define __rte_always_inline
Definition: rte_common.h:331
rte_crypto_op_sess_type
Definition: rte_crypto.h:61
rte_crypto_op_type
Definition: rte_crypto.h:28
rte_crypto_op_status
Definition: rte_crypto.h:38
rte_crypto_asym_op_type
rte_crypto_asym_xform_type
rte_crypto_auth_algorithm
rte_crypto_sym_xform_type
rte_crypto_aead_algorithm
rte_crypto_cipher_algorithm
static uint16_t rte_cryptodev_dequeue_burst(uint8_t dev_id, uint16_t qp_id, struct rte_crypto_op **ops, uint16_t nb_ops)
int rte_cryptodev_asym_session_create(uint8_t dev_id, struct rte_crypto_asym_xform *xforms, struct rte_mempool *mp, void **session)
void * rte_cryptodev_sym_session_create(uint8_t dev_id, struct rte_crypto_sym_xform *xforms, struct rte_mempool *mp)
uint32_t(* cryptodev_sym_raw_dequeue_burst_t)(void *qp, uint8_t *drv_ctx, rte_cryptodev_raw_get_dequeue_count_t get_dequeue_count, uint32_t max_nb_to_dequeue, rte_cryptodev_raw_post_dequeue_t post_dequeue, void **out_user_data, uint8_t is_user_data_array, uint32_t *n_success, int *dequeue_status)
int rte_cryptodev_close(uint8_t dev_id)
rte_cryptodev_event_type
@ RTE_CRYPTODEV_EVENT_ERROR
@ RTE_CRYPTODEV_EVENT_UNKNOWN
@ RTE_CRYPTODEV_EVENT_MAX
uint8_t rte_cryptodev_devices_get(const char *driver_name, uint8_t *devices, uint8_t nb_devices)
uint8_t rte_cryptodev_count(void)
static uint16_t rte_cryptodev_enqueue_burst(uint8_t dev_id, uint16_t qp_id, struct rte_crypto_op **ops, uint16_t nb_ops)
int rte_cryptodev_start(uint8_t dev_id)
bool rte_cryptodev_asym_xform_capability_check_hash(const struct rte_cryptodev_asymmetric_xform_capability *capability, enum rte_crypto_auth_algorithm hash)
__rte_experimental const char * rte_cryptodev_get_auth_algo_string(enum rte_crypto_auth_algorithm algo_enum)
uint32_t rte_cryptodev_raw_enqueue_burst(struct rte_crypto_raw_dp_ctx *ctx, struct rte_crypto_sym_vec *vec, union rte_crypto_sym_ofs ofs, void **user_data, int *enqueue_status)
static __rte_experimental __rte_always_inline void * rte_cryptodev_raw_dequeue(struct rte_crypto_raw_dp_ctx *ctx, int *dequeue_status, enum rte_crypto_op_status *op_status)
int rte_cryptodev_asym_xform_capability_check_modlen(const struct rte_cryptodev_asymmetric_xform_capability *capability, uint16_t modlen)
static uint64_t rte_cryptodev_sym_session_opaque_data_get(void *sess)
int rte_cryptodev_configure_raw_dp_ctx(uint8_t dev_id, uint16_t qp_id, struct rte_crypto_raw_dp_ctx *ctx, enum rte_crypto_op_sess_type sess_type, union rte_cryptodev_session_ctx session_ctx, uint8_t is_update)
unsigned int rte_cryptodev_is_valid_dev(uint8_t dev_id)
int rte_cryptodev_session_event_mdata_set(uint8_t dev_id, void *sess, enum rte_crypto_op_type op_type, enum rte_crypto_op_sess_type sess_type, void *ev_mdata, uint16_t size)
uint32_t rte_cryptodev_raw_dequeue_burst(struct rte_crypto_raw_dp_ctx *ctx, rte_cryptodev_raw_get_dequeue_count_t get_dequeue_count, uint32_t max_nb_to_dequeue, rte_cryptodev_raw_post_dequeue_t post_dequeue, void **out_user_data, uint8_t is_user_data_array, uint32_t *n_success, int *dequeue_status)
unsigned int rte_cryptodev_sym_get_private_session_size(uint8_t dev_id)
int rte_cryptodev_get_auth_algo_enum(enum rte_crypto_auth_algorithm *algo_enum, const char *algo_string)
void rte_cryptodev_stop(uint8_t dev_id)
RTE_TAILQ_HEAD(rte_cryptodev_cb_list, rte_cryptodev_callback)
void * rte_cryptodev_asym_session_get_user_data(void *sess)
int rte_cryptodev_sym_session_set_user_data(void *sess, void *data, uint16_t size)
int rte_cryptodev_callback_unregister(uint8_t dev_id, enum rte_cryptodev_event_type event, rte_cryptodev_cb_fn cb_fn, void *cb_arg)
static void rte_cryptodev_sym_session_opaque_data_set(void *sess, uint64_t opaque)
int rte_cryptodev_remove_enq_callback(uint8_t dev_id, uint16_t qp_id, struct rte_cryptodev_cb *cb)
__rte_experimental const char * rte_cryptodev_get_aead_algo_string(enum rte_crypto_aead_algorithm algo_enum)
int rte_cryptodev_remove_deq_callback(uint8_t dev_id, uint16_t qp_id, struct rte_cryptodev_cb *cb)
int rte_cryptodev_sym_capability_check_auth(const struct rte_cryptodev_symmetric_capability *capability, uint16_t key_size, uint16_t digest_size, uint16_t iv_size)
void * rte_cryptodev_sym_session_get_user_data(void *sess)
int rte_cryptodev_configure(uint8_t dev_id, struct rte_cryptodev_config *config)
int rte_cryptodev_raw_dequeue_done(struct rte_crypto_raw_dp_ctx *ctx, uint32_t n)
struct rte_mempool * rte_cryptodev_asym_session_pool_create(const char *name, uint32_t nb_elts, uint32_t cache_size, uint16_t user_data_size, int socket_id)
int rte_cryptodev_callback_register(uint8_t dev_id, enum rte_cryptodev_event_type event, rte_cryptodev_cb_fn cb_fn, void *cb_arg)
uint16_t rte_cryptodev_queue_pair_count(uint8_t dev_id)
__rte_experimental const char * rte_cryptodev_asym_get_xform_string(enum rte_crypto_asym_xform_type xform_enum)
void(* rte_cryptodev_cb_fn)(uint8_t dev_id, enum rte_cryptodev_event_type event, void *cb_arg)
const char * rte_cryptodev_get_feature_name(uint64_t flag)
uint32_t(* cryptodev_sym_raw_enqueue_burst_t)(void *qp, uint8_t *drv_ctx, struct rte_crypto_sym_vec *vec, union rte_crypto_sym_ofs ofs, void *user_data[], int *enqueue_status)
void rte_cryptodev_stats_reset(uint8_t dev_id)
static __rte_experimental __rte_always_inline int rte_cryptodev_raw_enqueue(struct rte_crypto_raw_dp_ctx *ctx, struct rte_crypto_vec *data_vec, uint16_t n_data_vecs, union rte_crypto_sym_ofs ofs, struct rte_crypto_va_iova_ptr *iv, struct rte_crypto_va_iova_ptr *digest, struct rte_crypto_va_iova_ptr *aad_or_auth_iv, void *user_data)
int rte_cryptodev_asym_session_free(uint8_t dev_id, void *sess)
int rte_cryptodev_asym_get_xform_enum(enum rte_crypto_asym_xform_type *xform_enum, const char *xform_string)
void * rte_cryptodev_get_sec_ctx(uint8_t dev_id)
const char * rte_cryptodev_name_get(uint8_t dev_id)
void(* rte_cryptodev_raw_post_dequeue_t)(void *user_data, uint32_t index, uint8_t is_op_success)
__rte_experimental const char * rte_cryptodev_get_cipher_algo_string(enum rte_crypto_cipher_algorithm algo_enum)
int rte_cryptodev_driver_id_get(const char *name)
struct rte_mempool * rte_cryptodev_sym_session_pool_create(const char *name, uint32_t nb_elts, uint32_t elt_size, uint32_t cache_size, uint16_t priv_size, int socket_id)
int rte_cryptodev_get_dev_id(const char *name)
unsigned int rte_cryptodev_asym_get_private_session_size(uint8_t dev_id)
int rte_cryptodev_stats_get(uint8_t dev_id, struct rte_cryptodev_stats *stats)
struct rte_cryptodev_cb * rte_cryptodev_add_enq_callback(uint8_t dev_id, uint16_t qp_id, rte_cryptodev_callback_fn cb_fn, void *cb_arg)
int rte_cryptodev_get_cipher_algo_enum(enum rte_crypto_cipher_algorithm *algo_enum, const char *algo_string)
unsigned int rte_cryptodev_asym_get_header_session_size(void)
uint32_t(* rte_cryptodev_raw_get_dequeue_count_t)(void *user_data)
__rte_experimental int rte_cryptodev_queue_pair_event_error_query(uint8_t dev_id, uint16_t qp_id)
void *(* cryptodev_sym_raw_dequeue_t)(void *qp, uint8_t *drv_ctx, int *dequeue_status, enum rte_crypto_op_status *op_status)
int(* cryptodev_sym_raw_enqueue_t)(void *qp, uint8_t *drv_ctx, struct rte_crypto_vec *data_vec, uint16_t n_data_vecs, union rte_crypto_sym_ofs ofs, struct rte_crypto_va_iova_ptr *iv, struct rte_crypto_va_iova_ptr *digest, struct rte_crypto_va_iova_ptr *aad_or_auth_iv, void *user_data)
int rte_cryptodev_raw_enqueue_done(struct rte_crypto_raw_dp_ctx *ctx, uint32_t n)
int rte_cryptodev_sym_capability_check_aead(const struct rte_cryptodev_symmetric_capability *capability, uint16_t key_size, uint16_t digest_size, uint16_t aad_size, uint16_t iv_size)
int rte_cryptodev_asym_session_set_user_data(void *sess, void *data, uint16_t size)
int rte_cryptodev_sym_capability_check_cipher(const struct rte_cryptodev_symmetric_capability *capability, uint16_t key_size, uint16_t iv_size)
uint8_t rte_cryptodev_device_count_by_driver(uint8_t driver_id)
const struct rte_cryptodev_asymmetric_xform_capability * rte_cryptodev_asym_capability_get(uint8_t dev_id, const struct rte_cryptodev_asym_capability_idx *idx)
int rte_cryptodev_get_aead_algo_enum(enum rte_crypto_aead_algorithm *algo_enum, const char *algo_string)
uint16_t(* rte_cryptodev_callback_fn)(uint16_t dev_id, uint16_t qp_id, struct rte_crypto_op **ops, uint16_t nb_ops, void *user_param)
int rte_cryptodev_get_raw_dp_ctx_size(uint8_t dev_id)
struct rte_cryptodev_cb * rte_cryptodev_add_deq_callback(uint8_t dev_id, uint16_t qp_id, rte_cryptodev_callback_fn cb_fn, void *cb_arg)
int rte_cryptodev_queue_pair_setup(uint8_t dev_id, uint16_t queue_pair_id, const struct rte_cryptodev_qp_conf *qp_conf, int socket_id)
int(* cryptodev_sym_raw_operation_done_t)(void *qp, uint8_t *drv_ctx, uint32_t n)
int rte_cryptodev_asym_xform_capability_check_optype(const struct rte_cryptodev_asymmetric_xform_capability *capability, enum rte_crypto_asym_op_type op_type)
uint32_t rte_cryptodev_sym_cpu_crypto_process(uint8_t dev_id, void *sess, union rte_crypto_sym_ofs ofs, struct rte_crypto_sym_vec *vec)
int rte_cryptodev_get_qp_status(uint8_t dev_id, uint16_t queue_pair_id)
const struct rte_cryptodev_symmetric_capability * rte_cryptodev_sym_capability_get(uint8_t dev_id, const struct rte_cryptodev_sym_capability_idx *idx)
void rte_cryptodev_info_get(uint8_t dev_id, struct rte_cryptodev_info *dev_info)
int rte_cryptodev_sym_session_free(uint8_t dev_id, void *sess)
const char * rte_cryptodev_driver_name_get(uint8_t driver_id)
static __rte_always_inline void rte_rcu_qsbr_thread_online(struct rte_rcu_qsbr *v, unsigned int thread_id)
Definition: rte_rcu_qsbr.h:296
static __rte_always_inline void rte_rcu_qsbr_thread_offline(struct rte_rcu_qsbr *v, unsigned int thread_id)
Definition: rte_rcu_qsbr.h:349
enum rte_crypto_asym_xform_type type
struct rte_crypto_param_range modlen
enum rte_crypto_asym_xform_type xform_type
enum rte_crypto_op_type op
struct rte_cryptodev_symmetric_capability sym
struct rte_cryptodev_asymmetric_capability asym
struct rte_cryptodev_cb * next
rte_cryptodev_callback_fn fn
unsigned max_nb_queue_pairs
struct rte_device * device
uint16_t min_mbuf_headroom_req
const struct rte_cryptodev_capabilities * capabilities
uint16_t min_mbuf_tailroom_req
const char * driver_name
unsigned max_nb_sessions
struct rte_mempool * mp_session
uint64_t enqueue_err_count
uint64_t dequeue_err_count
enum rte_crypto_auth_algorithm algo
struct rte_crypto_param_range iv_size
struct rte_crypto_param_range digest_size
struct rte_cryptodev_symmetric_capability::@101::@103 auth
struct rte_cryptodev_symmetric_capability::@101::@104 cipher
struct rte_crypto_param_range aad_size
struct rte_crypto_param_range key_size
enum rte_crypto_sym_xform_type xform_type
char name[RTE_MEMPOOL_NAMESIZE]
Definition: rte_mempool.h:221
uint32_t size
Definition: rte_mempool.h:230
uint32_t cache_size
Definition: rte_mempool.h:231
uint32_t elt_size
Definition: rte_mempool.h:234