aboutsummaryrefslogtreecommitdiff
path: root/include/crypto
diff options
context:
space:
mode:
authorHerbert Xu2007-12-14 22:28:14 +0800
committerHerbert Xu2008-01-11 08:16:55 +1100
commite7cd2514ea506f06bd4f7b13a9b62afd60f9c73b (patch)
tree66c46a54a2841091165a0b8083329c41f3504847 /include/crypto
parent4726204200327c04a77b819e2c653c063f1bc6ab (diff)
[CRYPTO] chainiv: Avoid lock spinning where possible
This patch makes chainiv avoid spinning by postponing requests on lock contention if the user allows the use of asynchronous algorithms. If a synchronous algorithm is requested then we behave as before. This should improve IPsec performance on SMP when two CPUs attempt to transmit over the same SA. Currently one of them will spin doing nothing waiting for the other CPU to finish its encryption. This patch makes it postpone the request and get on with other work. If only one CPU is transmitting for a given SA, then we will process the request synchronously as before. Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
Diffstat (limited to 'include/crypto')
-rw-r--r--include/crypto/internal/skcipher.h13
1 files changed, 13 insertions, 0 deletions
diff --git a/include/crypto/internal/skcipher.h b/include/crypto/internal/skcipher.h
index 0053f34764ff..2ba42cd7d6aa 100644
--- a/include/crypto/internal/skcipher.h
+++ b/include/crypto/internal/skcipher.h
@@ -70,6 +70,19 @@ static inline struct crypto_ablkcipher *skcipher_geniv_cipher(
return crypto_ablkcipher_crt(geniv)->base;
}
+static inline int skcipher_enqueue_givcrypt(
+ struct crypto_queue *queue, struct skcipher_givcrypt_request *request)
+{
+ return ablkcipher_enqueue_request(queue, &request->creq);
+}
+
+static inline struct skcipher_givcrypt_request *skcipher_dequeue_givcrypt(
+ struct crypto_queue *queue)
+{
+ return container_of(ablkcipher_dequeue_request(queue),
+ struct skcipher_givcrypt_request, creq);
+}
+
static inline void *skcipher_givcrypt_reqctx(
struct skcipher_givcrypt_request *req)
{