/* * Copyright (c) 2006 Oracle. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * */ #include <linux/kernel.h> #include "rds.h" #include "ib.h" /* * Locking for IB rings. * We assume that allocation is always protected by a mutex * in the caller (this is a valid assumption for the current * implementation). * * Freeing always happens in an interrupt, and hence only * races with allocations, but not with other free()s. * * The interaction between allocation and freeing is that * the alloc code has to determine the number of free entries. * To this end, we maintain two counters; an allocation counter * and a free counter. Both are allowed to run freely, and wrap * around. * The number of used entries is always (alloc_ctr - free_ctr) % NR. * * The current implementation makes free_ctr atomic. When the * caller finds an allocation fails, it should set an "alloc fail" * bit and retry the allocation. The "alloc fail" bit essentially tells * the CQ completion handlers to wake it up after freeing some * more entries. */ /* * This only happens on shutdown. */ DECLARE_WAIT_QUEUE_HEAD(…); void rds_ib_ring_init(struct rds_ib_work_ring *ring, u32 nr) { … } static inline u32 __rds_ib_ring_used(struct rds_ib_work_ring *ring) { … } void rds_ib_ring_resize(struct rds_ib_work_ring *ring, u32 nr) { … } static int __rds_ib_ring_empty(struct rds_ib_work_ring *ring) { … } u32 rds_ib_ring_alloc(struct rds_ib_work_ring *ring, u32 val, u32 *pos) { … } void rds_ib_ring_free(struct rds_ib_work_ring *ring, u32 val) { … } void rds_ib_ring_unalloc(struct rds_ib_work_ring *ring, u32 val) { … } int rds_ib_ring_empty(struct rds_ib_work_ring *ring) { … } int rds_ib_ring_low(struct rds_ib_work_ring *ring) { … } /* * returns the oldest allocated ring entry. This will be the next one * freed. This can't be called if there are none allocated. */ u32 rds_ib_ring_oldest(struct rds_ib_work_ring *ring) { … } /* * returns the number of completed work requests. */ u32 rds_ib_ring_completed(struct rds_ib_work_ring *ring, u32 wr_id, u32 oldest) { … }