staging: ramster: delete the driver
authorGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Fri, 10 Feb 2012 17:45:22 +0000 (09:45 -0800)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Fri, 10 Feb 2012 17:45:22 +0000 (09:45 -0800)
Turns out it's not quite ready to be included, thanks to some other work
done in the zcache and zram code, which breaks this driver.

So, delete it for now, per the recommendation of Dan.

Acked-by: Dan Magenheimer <dan.magenheimer@oracle.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
30 files changed:
drivers/staging/Kconfig
drivers/staging/Makefile
drivers/staging/ramster/Kconfig [deleted file]
drivers/staging/ramster/Makefile [deleted file]
drivers/staging/ramster/TODO [deleted file]
drivers/staging/ramster/cluster/Makefile [deleted file]
drivers/staging/ramster/cluster/heartbeat.c [deleted file]
drivers/staging/ramster/cluster/heartbeat.h [deleted file]
drivers/staging/ramster/cluster/masklog.c [deleted file]
drivers/staging/ramster/cluster/masklog.h [deleted file]
drivers/staging/ramster/cluster/netdebug.c [deleted file]
drivers/staging/ramster/cluster/nodemanager.c [deleted file]
drivers/staging/ramster/cluster/nodemanager.h [deleted file]
drivers/staging/ramster/cluster/ocfs2_heartbeat.h [deleted file]
drivers/staging/ramster/cluster/ocfs2_nodemanager.h [deleted file]
drivers/staging/ramster/cluster/quorum.c [deleted file]
drivers/staging/ramster/cluster/quorum.h [deleted file]
drivers/staging/ramster/cluster/sys.c [deleted file]
drivers/staging/ramster/cluster/sys.h [deleted file]
drivers/staging/ramster/cluster/tcp.c [deleted file]
drivers/staging/ramster/cluster/tcp.h [deleted file]
drivers/staging/ramster/cluster/tcp_internal.h [deleted file]
drivers/staging/ramster/cluster/ver.c [deleted file]
drivers/staging/ramster/cluster/ver.h [deleted file]
drivers/staging/ramster/ramster.h [deleted file]
drivers/staging/ramster/ramster_o2net.c [deleted file]
drivers/staging/ramster/tmem.c [deleted file]
drivers/staging/ramster/tmem.h [deleted file]
drivers/staging/ramster/zcache-main.c [deleted file]
drivers/staging/ramster/zcache.h [deleted file]

index 783bc0a..f916a3f 100644 (file)
@@ -136,6 +136,4 @@ source "drivers/staging/android/Kconfig"
 
 source "drivers/staging/telephony/Kconfig"
 
-source "drivers/staging/ramster/Kconfig"
-
 endif # STAGING
index 077a3f3..f6d4a43 100644 (file)
@@ -57,5 +57,4 @@ obj-$(CONFIG_INTEL_MEI)               += mei/
 obj-$(CONFIG_MFD_NVEC)         += nvec/
 obj-$(CONFIG_DRM_OMAP)         += omapdrm/
 obj-$(CONFIG_ANDROID)          += android/
-obj-$(CONFIG_RAMSTER)          += ramster/
 obj-$(CONFIG_PHONE)            += telephony/
diff --git a/drivers/staging/ramster/Kconfig b/drivers/staging/ramster/Kconfig
deleted file mode 100644 (file)
index 3957b87..0000000
+++ /dev/null
@@ -1,14 +0,0 @@
-config RAMSTER
-       tristate "Cross-machine RAM capacity sharing, aka peer-to-peer tmem"
-       depends on (CLEANCACHE || FRONTSWAP) && CONFIGFS_FS && !OCFS2_FS && !ZCACHE && !HIGHMEM
-       select XVMALLOC
-       select LZO_COMPRESS
-       select LZO_DECOMPRESS
-       default n
-       help
-         RAMster allows RAM on other machines in a cluster to be utilized
-         dynamically and symmetrically instead of swapping to a local swap
-         disk, thus improving performance on memory-constrained workloads
-         while minimizing total RAM across the cluster.  RAMster, like
-         zcache, compresses swap pages into local RAM, but then remotifies
-         the compressed pages to another node in the RAMster cluster.
diff --git a/drivers/staging/ramster/Makefile b/drivers/staging/ramster/Makefile
deleted file mode 100644 (file)
index e6c4a2e..0000000
+++ /dev/null
@@ -1,2 +0,0 @@
-obj-$(CONFIG_RAMSTER)  +=      zcache-main.o tmem.o
-obj-$(CONFIG_RAMSTER)  +=      ramster_o2net.o cluster/
diff --git a/drivers/staging/ramster/TODO b/drivers/staging/ramster/TODO
deleted file mode 100644 (file)
index d4268f0..0000000
+++ /dev/null
@@ -1,9 +0,0 @@
-For this staging driver, RAMster duplicates code from fs/ocfs2/cluster
-and from drivers/staging/zcache, then incorporates changes to the local
-copy of the code.  Before RAMster can be promoted from staging, this code
-duplication must be resolved.  Specifically, we will first need to work with
-the ocfs2 maintainers to split out the ocfs2 core cluster code so that
-it can be easily included by another subsystem, even if ocfs2 is not
-configured, and also to merge the handful of functional changes required.
-Second, the zcache and RAMster drivers should be either merged or reorganized
-to separate out common code.
diff --git a/drivers/staging/ramster/cluster/Makefile b/drivers/staging/ramster/cluster/Makefile
deleted file mode 100644 (file)
index 3fc8550..0000000
+++ /dev/null
@@ -1,5 +0,0 @@
-#obj-$(CONFIG_OCFS2_FS) += ocfs2_nodemanager.o
-obj-$(CONFIG_RAMSTER) += ocfs2_nodemanager.o
-
-ocfs2_nodemanager-objs := heartbeat.o masklog.o sys.o nodemanager.o \
-       quorum.o tcp.o netdebug.o ver.o
diff --git a/drivers/staging/ramster/cluster/heartbeat.c b/drivers/staging/ramster/cluster/heartbeat.c
deleted file mode 100644 (file)
index 1f72143..0000000
+++ /dev/null
@@ -1,2687 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2004, 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#include <linux/kernel.h>
-#include <linux/sched.h>
-#include <linux/jiffies.h>
-#include <linux/module.h>
-#include <linux/fs.h>
-#include <linux/bio.h>
-#include <linux/blkdev.h>
-#include <linux/delay.h>
-#include <linux/file.h>
-#include <linux/kthread.h>
-#include <linux/configfs.h>
-#include <linux/random.h>
-#include <linux/crc32.h>
-#include <linux/time.h>
-#include <linux/debugfs.h>
-#include <linux/slab.h>
-
-#include "heartbeat.h"
-#include "tcp.h"
-#include "nodemanager.h"
-#include "quorum.h"
-
-#include "masklog.h"
-
-
-/*
- * The first heartbeat pass had one global thread that would serialize all hb
- * callback calls.  This global serializing sem should only be removed once
- * we've made sure that all callees can deal with being called concurrently
- * from multiple hb region threads.
- */
-static DECLARE_RWSEM(o2hb_callback_sem);
-
-/*
- * multiple hb threads are watching multiple regions.  A node is live
- * whenever any of the threads sees activity from the node in its region.
- */
-static DEFINE_SPINLOCK(o2hb_live_lock);
-static struct list_head o2hb_live_slots[O2NM_MAX_NODES];
-static unsigned long o2hb_live_node_bitmap[BITS_TO_LONGS(O2NM_MAX_NODES)];
-static LIST_HEAD(o2hb_node_events);
-static DECLARE_WAIT_QUEUE_HEAD(o2hb_steady_queue);
-
-/*
- * In global heartbeat, we maintain a series of region bitmaps.
- *     - o2hb_region_bitmap allows us to limit the region number to max region.
- *     - o2hb_live_region_bitmap tracks live regions (seen steady iterations).
- *     - o2hb_quorum_region_bitmap tracks live regions that have seen all nodes
- *             heartbeat on it.
- *     - o2hb_failed_region_bitmap tracks the regions that have seen io timeouts.
- */
-static unsigned long o2hb_region_bitmap[BITS_TO_LONGS(O2NM_MAX_REGIONS)];
-static unsigned long o2hb_live_region_bitmap[BITS_TO_LONGS(O2NM_MAX_REGIONS)];
-static unsigned long o2hb_quorum_region_bitmap[BITS_TO_LONGS(O2NM_MAX_REGIONS)];
-static unsigned long o2hb_failed_region_bitmap[BITS_TO_LONGS(O2NM_MAX_REGIONS)];
-
-#define O2HB_DB_TYPE_LIVENODES         0
-#define O2HB_DB_TYPE_LIVEREGIONS       1
-#define O2HB_DB_TYPE_QUORUMREGIONS     2
-#define O2HB_DB_TYPE_FAILEDREGIONS     3
-#define O2HB_DB_TYPE_REGION_LIVENODES  4
-#define O2HB_DB_TYPE_REGION_NUMBER     5
-#define O2HB_DB_TYPE_REGION_ELAPSED_TIME       6
-#define O2HB_DB_TYPE_REGION_PINNED     7
-struct o2hb_debug_buf {
-       int db_type;
-       int db_size;
-       int db_len;
-       void *db_data;
-};
-
-static struct o2hb_debug_buf *o2hb_db_livenodes;
-static struct o2hb_debug_buf *o2hb_db_liveregions;
-static struct o2hb_debug_buf *o2hb_db_quorumregions;
-static struct o2hb_debug_buf *o2hb_db_failedregions;
-
-#define O2HB_DEBUG_DIR                 "o2hb"
-#define O2HB_DEBUG_LIVENODES           "livenodes"
-#define O2HB_DEBUG_LIVEREGIONS         "live_regions"
-#define O2HB_DEBUG_QUORUMREGIONS       "quorum_regions"
-#define O2HB_DEBUG_FAILEDREGIONS       "failed_regions"
-#define O2HB_DEBUG_REGION_NUMBER       "num"
-#define O2HB_DEBUG_REGION_ELAPSED_TIME "elapsed_time_in_ms"
-#define O2HB_DEBUG_REGION_PINNED       "pinned"
-
-static struct dentry *o2hb_debug_dir;
-static struct dentry *o2hb_debug_livenodes;
-static struct dentry *o2hb_debug_liveregions;
-static struct dentry *o2hb_debug_quorumregions;
-static struct dentry *o2hb_debug_failedregions;
-
-static LIST_HEAD(o2hb_all_regions);
-
-static struct o2hb_callback {
-       struct list_head list;
-} o2hb_callbacks[O2HB_NUM_CB];
-
-static struct o2hb_callback *hbcall_from_type(enum o2hb_callback_type type);
-
-#define O2HB_DEFAULT_BLOCK_BITS       9
-
-enum o2hb_heartbeat_modes {
-       O2HB_HEARTBEAT_LOCAL            = 0,
-       O2HB_HEARTBEAT_GLOBAL,
-       O2HB_HEARTBEAT_NUM_MODES,
-};
-
-char *o2hb_heartbeat_mode_desc[O2HB_HEARTBEAT_NUM_MODES] = {
-               "local",        /* O2HB_HEARTBEAT_LOCAL */
-               "global",       /* O2HB_HEARTBEAT_GLOBAL */
-};
-
-unsigned int o2hb_dead_threshold = O2HB_DEFAULT_DEAD_THRESHOLD;
-unsigned int o2hb_heartbeat_mode = O2HB_HEARTBEAT_LOCAL;
-
-/*
- * o2hb_dependent_users tracks the number of registered callbacks that depend
- * on heartbeat. o2net and o2dlm are two entities that register this callback.
- * However only o2dlm depends on the heartbeat. It does not want the heartbeat
- * to stop while a dlm domain is still active.
- */
-unsigned int o2hb_dependent_users;
-
-/*
- * In global heartbeat mode, all regions are pinned if there are one or more
- * dependent users and the quorum region count is <= O2HB_PIN_CUT_OFF. All
- * regions are unpinned if the region count exceeds the cut off or the number
- * of dependent users falls to zero.
- */
-#define O2HB_PIN_CUT_OFF               3
-
-/*
- * In local heartbeat mode, we assume the dlm domain name to be the same as
- * region uuid. This is true for domains created for the file system but not
- * necessarily true for userdlm domains. This is a known limitation.
- *
- * In global heartbeat mode, we pin/unpin all o2hb regions. This solution
- * works for both file system and userdlm domains.
- */
-static int o2hb_region_pin(const char *region_uuid);
-static void o2hb_region_unpin(const char *region_uuid);
-
-/* Only sets a new threshold if there are no active regions.
- *
- * No locking or otherwise interesting code is required for reading
- * o2hb_dead_threshold as it can't change once regions are active and
- * it's not interesting to anyone until then anyway. */
-static void o2hb_dead_threshold_set(unsigned int threshold)
-{
-       if (threshold > O2HB_MIN_DEAD_THRESHOLD) {
-               spin_lock(&o2hb_live_lock);
-               if (list_empty(&o2hb_all_regions))
-                       o2hb_dead_threshold = threshold;
-               spin_unlock(&o2hb_live_lock);
-       }
-}
-
-static int o2hb_global_hearbeat_mode_set(unsigned int hb_mode)
-{
-       int ret = -1;
-
-       if (hb_mode < O2HB_HEARTBEAT_NUM_MODES) {
-               spin_lock(&o2hb_live_lock);
-               if (list_empty(&o2hb_all_regions)) {
-                       o2hb_heartbeat_mode = hb_mode;
-                       ret = 0;
-               }
-               spin_unlock(&o2hb_live_lock);
-       }
-
-       return ret;
-}
-
-struct o2hb_node_event {
-       struct list_head        hn_item;
-       enum o2hb_callback_type hn_event_type;
-       struct o2nm_node        *hn_node;
-       int                     hn_node_num;
-};
-
-struct o2hb_disk_slot {
-       struct o2hb_disk_heartbeat_block *ds_raw_block;
-       u8                      ds_node_num;
-       u64                     ds_last_time;
-       u64                     ds_last_generation;
-       u16                     ds_equal_samples;
-       u16                     ds_changed_samples;
-       struct list_head        ds_live_item;
-};
-
-/* each thread owns a region.. when we're asked to tear down the region
- * we ask the thread to stop, who cleans up the region */
-struct o2hb_region {
-       struct config_item      hr_item;
-
-       struct list_head        hr_all_item;
-       unsigned                hr_unclean_stop:1,
-                               hr_aborted_start:1,
-                               hr_item_pinned:1,
-                               hr_item_dropped:1;
-
-       /* protected by the hr_callback_sem */
-       struct task_struct      *hr_task;
-
-       unsigned int            hr_blocks;
-       unsigned long long      hr_start_block;
-
-       unsigned int            hr_block_bits;
-       unsigned int            hr_block_bytes;
-
-       unsigned int            hr_slots_per_page;
-       unsigned int            hr_num_pages;
-
-       struct page             **hr_slot_data;
-       struct block_device     *hr_bdev;
-       struct o2hb_disk_slot   *hr_slots;
-
-       /* live node map of this region */
-       unsigned long           hr_live_node_bitmap[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       unsigned int            hr_region_num;
-
-       struct dentry           *hr_debug_dir;
-       struct dentry           *hr_debug_livenodes;
-       struct dentry           *hr_debug_regnum;
-       struct dentry           *hr_debug_elapsed_time;
-       struct dentry           *hr_debug_pinned;
-       struct o2hb_debug_buf   *hr_db_livenodes;
-       struct o2hb_debug_buf   *hr_db_regnum;
-       struct o2hb_debug_buf   *hr_db_elapsed_time;
-       struct o2hb_debug_buf   *hr_db_pinned;
-
-       /* let the person setting up hb wait for it to return until it
-        * has reached a 'steady' state.  This will be fixed when we have
-        * a more complete api that doesn't lead to this sort of fragility. */
-       atomic_t                hr_steady_iterations;
-
-       /* terminate o2hb thread if it does not reach steady state
-        * (hr_steady_iterations == 0) within hr_unsteady_iterations */
-       atomic_t                hr_unsteady_iterations;
-
-       char                    hr_dev_name[BDEVNAME_SIZE];
-
-       unsigned int            hr_timeout_ms;
-
-       /* randomized as the region goes up and down so that a node
-        * recognizes a node going up and down in one iteration */
-       u64                     hr_generation;
-
-       struct delayed_work     hr_write_timeout_work;
-       unsigned long           hr_last_timeout_start;
-
-       /* Used during o2hb_check_slot to hold a copy of the block
-        * being checked because we temporarily have to zero out the
-        * crc field. */
-       struct o2hb_disk_heartbeat_block *hr_tmp_block;
-};
-
-struct o2hb_bio_wait_ctxt {
-       atomic_t          wc_num_reqs;
-       struct completion wc_io_complete;
-       int               wc_error;
-};
-
-static int o2hb_pop_count(void *map, int count)
-{
-       int i = -1, pop = 0;
-
-       while ((i = find_next_bit(map, count, i + 1)) < count)
-               pop++;
-       return pop;
-}
-
-static void o2hb_write_timeout(struct work_struct *work)
-{
-       int failed, quorum;
-       unsigned long flags;
-       struct o2hb_region *reg =
-               container_of(work, struct o2hb_region,
-                            hr_write_timeout_work.work);
-
-       mlog(ML_ERROR, "Heartbeat write timeout to device %s after %u "
-            "milliseconds\n", reg->hr_dev_name,
-            jiffies_to_msecs(jiffies - reg->hr_last_timeout_start));
-
-       if (o2hb_global_heartbeat_active()) {
-               spin_lock_irqsave(&o2hb_live_lock, flags);
-               if (test_bit(reg->hr_region_num, o2hb_quorum_region_bitmap))
-                       set_bit(reg->hr_region_num, o2hb_failed_region_bitmap);
-               failed = o2hb_pop_count(&o2hb_failed_region_bitmap,
-                                       O2NM_MAX_REGIONS);
-               quorum = o2hb_pop_count(&o2hb_quorum_region_bitmap,
-                                       O2NM_MAX_REGIONS);
-               spin_unlock_irqrestore(&o2hb_live_lock, flags);
-
-               mlog(ML_HEARTBEAT, "Number of regions %d, failed regions %d\n",
-                    quorum, failed);
-
-               /*
-                * Fence if the number of failed regions >= half the number
-                * of  quorum regions
-                */
-               if ((failed << 1) < quorum)
-                       return;
-       }
-
-       o2quo_disk_timeout();
-}
-
-static void o2hb_arm_write_timeout(struct o2hb_region *reg)
-{
-       /* Arm writeout only after thread reaches steady state */
-       if (atomic_read(&reg->hr_steady_iterations) != 0)
-               return;
-
-       mlog(ML_HEARTBEAT, "Queue write timeout for %u ms\n",
-            O2HB_MAX_WRITE_TIMEOUT_MS);
-
-       if (o2hb_global_heartbeat_active()) {
-               spin_lock(&o2hb_live_lock);
-               clear_bit(reg->hr_region_num, o2hb_failed_region_bitmap);
-               spin_unlock(&o2hb_live_lock);
-       }
-       cancel_delayed_work(&reg->hr_write_timeout_work);
-       reg->hr_last_timeout_start = jiffies;
-       schedule_delayed_work(&reg->hr_write_timeout_work,
-                             msecs_to_jiffies(O2HB_MAX_WRITE_TIMEOUT_MS));
-}
-
-static void o2hb_disarm_write_timeout(struct o2hb_region *reg)
-{
-       cancel_delayed_work_sync(&reg->hr_write_timeout_work);
-}
-
-static inline void o2hb_bio_wait_init(struct o2hb_bio_wait_ctxt *wc)
-{
-       atomic_set(&wc->wc_num_reqs, 1);
-       init_completion(&wc->wc_io_complete);
-       wc->wc_error = 0;
-}
-
-/* Used in error paths too */
-static inline void o2hb_bio_wait_dec(struct o2hb_bio_wait_ctxt *wc,
-                                    unsigned int num)
-{
-       /* sadly atomic_sub_and_test() isn't available on all platforms.  The
-        * good news is that the fast path only completes one at a time */
-       while(num--) {
-               if (atomic_dec_and_test(&wc->wc_num_reqs)) {
-                       BUG_ON(num > 0);
-                       complete(&wc->wc_io_complete);
-               }
-       }
-}
-
-static void o2hb_wait_on_io(struct o2hb_region *reg,
-                           struct o2hb_bio_wait_ctxt *wc)
-{
-       o2hb_bio_wait_dec(wc, 1);
-       wait_for_completion(&wc->wc_io_complete);
-}
-
-static void o2hb_bio_end_io(struct bio *bio,
-                          int error)
-{
-       struct o2hb_bio_wait_ctxt *wc = bio->bi_private;
-
-       if (error) {
-               mlog(ML_ERROR, "IO Error %d\n", error);
-               wc->wc_error = error;
-       }
-
-       o2hb_bio_wait_dec(wc, 1);
-       bio_put(bio);
-}
-
-/* Setup a Bio to cover I/O against num_slots slots starting at
- * start_slot. */
-static struct bio *o2hb_setup_one_bio(struct o2hb_region *reg,
-                                     struct o2hb_bio_wait_ctxt *wc,
-                                     unsigned int *current_slot,
-                                     unsigned int max_slots)
-{
-       int len, current_page;
-       unsigned int vec_len, vec_start;
-       unsigned int bits = reg->hr_block_bits;
-       unsigned int spp = reg->hr_slots_per_page;
-       unsigned int cs = *current_slot;
-       struct bio *bio;
-       struct page *page;
-
-       /* Testing has shown this allocation to take long enough under
-        * GFP_KERNEL that the local node can get fenced. It would be
-        * nicest if we could pre-allocate these bios and avoid this
-        * all together. */
-       bio = bio_alloc(GFP_ATOMIC, 16);
-       if (!bio) {
-               mlog(ML_ERROR, "Could not alloc slots BIO!\n");
-               bio = ERR_PTR(-ENOMEM);
-               goto bail;
-       }
-
-       /* Must put everything in 512 byte sectors for the bio... */
-       bio->bi_sector = (reg->hr_start_block + cs) << (bits - 9);
-       bio->bi_bdev = reg->hr_bdev;
-       bio->bi_private = wc;
-       bio->bi_end_io = o2hb_bio_end_io;
-
-       vec_start = (cs << bits) % PAGE_CACHE_SIZE;
-       while(cs < max_slots) {
-               current_page = cs / spp;
-               page = reg->hr_slot_data[current_page];
-
-               vec_len = min(PAGE_CACHE_SIZE - vec_start,
-                             (max_slots-cs) * (PAGE_CACHE_SIZE/spp) );
-
-               mlog(ML_HB_BIO, "page %d, vec_len = %u, vec_start = %u\n",
-                    current_page, vec_len, vec_start);
-
-               len = bio_add_page(bio, page, vec_len, vec_start);
-               if (len != vec_len) break;
-
-               cs += vec_len / (PAGE_CACHE_SIZE/spp);
-               vec_start = 0;
-       }
-
-bail:
-       *current_slot = cs;
-       return bio;
-}
-
-static int o2hb_read_slots(struct o2hb_region *reg,
-                          unsigned int max_slots)
-{
-       unsigned int current_slot=0;
-       int status;
-       struct o2hb_bio_wait_ctxt wc;
-       struct bio *bio;
-
-       o2hb_bio_wait_init(&wc);
-
-       while(current_slot < max_slots) {
-               bio = o2hb_setup_one_bio(reg, &wc, &current_slot, max_slots);
-               if (IS_ERR(bio)) {
-                       status = PTR_ERR(bio);
-                       mlog_errno(status);
-                       goto bail_and_wait;
-               }
-
-               atomic_inc(&wc.wc_num_reqs);
-               submit_bio(READ, bio);
-       }
-
-       status = 0;
-
-bail_and_wait:
-       o2hb_wait_on_io(reg, &wc);
-       if (wc.wc_error && !status)
-               status = wc.wc_error;
-
-       return status;
-}
-
-static int o2hb_issue_node_write(struct o2hb_region *reg,
-                                struct o2hb_bio_wait_ctxt *write_wc)
-{
-       int status;
-       unsigned int slot;
-       struct bio *bio;
-
-       o2hb_bio_wait_init(write_wc);
-
-       slot = o2nm_this_node();
-
-       bio = o2hb_setup_one_bio(reg, write_wc, &slot, slot+1);
-       if (IS_ERR(bio)) {
-               status = PTR_ERR(bio);
-               mlog_errno(status);
-               goto bail;
-       }
-
-       atomic_inc(&write_wc->wc_num_reqs);
-       submit_bio(WRITE, bio);
-
-       status = 0;
-bail:
-       return status;
-}
-
-static u32 o2hb_compute_block_crc_le(struct o2hb_region *reg,
-                                    struct o2hb_disk_heartbeat_block *hb_block)
-{
-       __le32 old_cksum;
-       u32 ret;
-
-       /* We want to compute the block crc with a 0 value in the
-        * hb_cksum field. Save it off here and replace after the
-        * crc. */
-       old_cksum = hb_block->hb_cksum;
-       hb_block->hb_cksum = 0;
-
-       ret = crc32_le(0, (unsigned char *) hb_block, reg->hr_block_bytes);
-
-       hb_block->hb_cksum = old_cksum;
-
-       return ret;
-}
-
-static void o2hb_dump_slot(struct o2hb_disk_heartbeat_block *hb_block)
-{
-       mlog(ML_ERROR, "Dump slot information: seq = 0x%llx, node = %u, "
-            "cksum = 0x%x, generation 0x%llx\n",
-            (long long)le64_to_cpu(hb_block->hb_seq),
-            hb_block->hb_node, le32_to_cpu(hb_block->hb_cksum),
-            (long long)le64_to_cpu(hb_block->hb_generation));
-}
-
-static int o2hb_verify_crc(struct o2hb_region *reg,
-                          struct o2hb_disk_heartbeat_block *hb_block)
-{
-       u32 read, computed;
-
-       read = le32_to_cpu(hb_block->hb_cksum);
-       computed = o2hb_compute_block_crc_le(reg, hb_block);
-
-       return read == computed;
-}
-
-/*
- * Compare the slot data with what we wrote in the last iteration.
- * If the match fails, print an appropriate error message. This is to
- * detect errors like... another node hearting on the same slot,
- * flaky device that is losing writes, etc.
- * Returns 1 if check succeeds, 0 otherwise.
- */
-static int o2hb_check_own_slot(struct o2hb_region *reg)
-{
-       struct o2hb_disk_slot *slot;
-       struct o2hb_disk_heartbeat_block *hb_block;
-       char *errstr;
-
-       slot = &reg->hr_slots[o2nm_this_node()];
-       /* Don't check on our 1st timestamp */
-       if (!slot->ds_last_time)
-               return 0;
-
-       hb_block = slot->ds_raw_block;
-       if (le64_to_cpu(hb_block->hb_seq) == slot->ds_last_time &&
-           le64_to_cpu(hb_block->hb_generation) == slot->ds_last_generation &&
-           hb_block->hb_node == slot->ds_node_num)
-               return 1;
-
-#define ERRSTR1                "Another node is heartbeating on device"
-#define ERRSTR2                "Heartbeat generation mismatch on device"
-#define ERRSTR3                "Heartbeat sequence mismatch on device"
-
-       if (hb_block->hb_node != slot->ds_node_num)
-               errstr = ERRSTR1;
-       else if (le64_to_cpu(hb_block->hb_generation) !=
-                slot->ds_last_generation)
-               errstr = ERRSTR2;
-       else
-               errstr = ERRSTR3;
-
-       mlog(ML_ERROR, "%s (%s): expected(%u:0x%llx, 0x%llx), "
-            "ondisk(%u:0x%llx, 0x%llx)\n", errstr, reg->hr_dev_name,
-            slot->ds_node_num, (unsigned long long)slot->ds_last_generation,
-            (unsigned long long)slot->ds_last_time, hb_block->hb_node,
-            (unsigned long long)le64_to_cpu(hb_block->hb_generation),
-            (unsigned long long)le64_to_cpu(hb_block->hb_seq));
-
-       return 0;
-}
-
-static inline void o2hb_prepare_block(struct o2hb_region *reg,
-                                     u64 generation)
-{
-       int node_num;
-       u64 cputime;
-       struct o2hb_disk_slot *slot;
-       struct o2hb_disk_heartbeat_block *hb_block;
-
-       node_num = o2nm_this_node();
-       slot = &reg->hr_slots[node_num];
-
-       hb_block = (struct o2hb_disk_heartbeat_block *)slot->ds_raw_block;
-       memset(hb_block, 0, reg->hr_block_bytes);
-       /* TODO: time stuff */
-       cputime = CURRENT_TIME.tv_sec;
-       if (!cputime)
-               cputime = 1;
-
-       hb_block->hb_seq = cpu_to_le64(cputime);
-       hb_block->hb_node = node_num;
-       hb_block->hb_generation = cpu_to_le64(generation);
-       hb_block->hb_dead_ms = cpu_to_le32(o2hb_dead_threshold * O2HB_REGION_TIMEOUT_MS);
-
-       /* This step must always happen last! */
-       hb_block->hb_cksum = cpu_to_le32(o2hb_compute_block_crc_le(reg,
-                                                                  hb_block));
-
-       mlog(ML_HB_BIO, "our node generation = 0x%llx, cksum = 0x%x\n",
-            (long long)generation,
-            le32_to_cpu(hb_block->hb_cksum));
-}
-
-static void o2hb_fire_callbacks(struct o2hb_callback *hbcall,
-                               struct o2nm_node *node,
-                               int idx)
-{
-       struct list_head *iter;
-       struct o2hb_callback_func *f;
-
-       list_for_each(iter, &hbcall->list) {
-               f = list_entry(iter, struct o2hb_callback_func, hc_item);
-               mlog(ML_HEARTBEAT, "calling funcs %p\n", f);
-               (f->hc_func)(node, idx, f->hc_data);
-       }
-}
-
-/* Will run the list in order until we process the passed event */
-static void o2hb_run_event_list(struct o2hb_node_event *queued_event)
-{
-       int empty;
-       struct o2hb_callback *hbcall;
-       struct o2hb_node_event *event;
-
-       spin_lock(&o2hb_live_lock);
-       empty = list_empty(&queued_event->hn_item);
-       spin_unlock(&o2hb_live_lock);
-       if (empty)
-               return;
-
-       /* Holding callback sem assures we don't alter the callback
-        * lists when doing this, and serializes ourselves with other
-        * processes wanting callbacks. */
-       down_write(&o2hb_callback_sem);
-
-       spin_lock(&o2hb_live_lock);
-       while (!list_empty(&o2hb_node_events)
-              && !list_empty(&queued_event->hn_item)) {
-               event = list_entry(o2hb_node_events.next,
-                                  struct o2hb_node_event,
-                                  hn_item);
-               list_del_init(&event->hn_item);
-               spin_unlock(&o2hb_live_lock);
-
-               mlog(ML_HEARTBEAT, "Node %s event for %d\n",
-                    event->hn_event_type == O2HB_NODE_UP_CB ? "UP" : "DOWN",
-                    event->hn_node_num);
-
-               hbcall = hbcall_from_type(event->hn_event_type);
-
-               /* We should *never* have gotten on to the list with a
-                * bad type... This isn't something that we should try
-                * to recover from. */
-               BUG_ON(IS_ERR(hbcall));
-
-               o2hb_fire_callbacks(hbcall, event->hn_node, event->hn_node_num);
-
-               spin_lock(&o2hb_live_lock);
-       }
-       spin_unlock(&o2hb_live_lock);
-
-       up_write(&o2hb_callback_sem);
-}
-
-static void o2hb_queue_node_event(struct o2hb_node_event *event,
-                                 enum o2hb_callback_type type,
-                                 struct o2nm_node *node,
-                                 int node_num)
-{
-       assert_spin_locked(&o2hb_live_lock);
-
-       BUG_ON((!node) && (type != O2HB_NODE_DOWN_CB));
-
-       event->hn_event_type = type;
-       event->hn_node = node;
-       event->hn_node_num = node_num;
-
-       mlog(ML_HEARTBEAT, "Queue node %s event for node %d\n",
-            type == O2HB_NODE_UP_CB ? "UP" : "DOWN", node_num);
-
-       list_add_tail(&event->hn_item, &o2hb_node_events);
-}
-
-static void o2hb_shutdown_slot(struct o2hb_disk_slot *slot)
-{
-       struct o2hb_node_event event =
-               { .hn_item = LIST_HEAD_INIT(event.hn_item), };
-       struct o2nm_node *node;
-
-       node = o2nm_get_node_by_num(slot->ds_node_num);
-       if (!node)
-               return;
-
-       spin_lock(&o2hb_live_lock);
-       if (!list_empty(&slot->ds_live_item)) {
-               mlog(ML_HEARTBEAT, "Shutdown, node %d leaves region\n",
-                    slot->ds_node_num);
-
-               list_del_init(&slot->ds_live_item);
-
-               if (list_empty(&o2hb_live_slots[slot->ds_node_num])) {
-                       clear_bit(slot->ds_node_num, o2hb_live_node_bitmap);
-
-                       o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node,
-                                             slot->ds_node_num);
-               }
-       }
-       spin_unlock(&o2hb_live_lock);
-
-       o2hb_run_event_list(&event);
-
-       o2nm_node_put(node);
-}
-
-static void o2hb_set_quorum_device(struct o2hb_region *reg)
-{
-       if (!o2hb_global_heartbeat_active())
-               return;
-
-       /* Prevent race with o2hb_heartbeat_group_drop_item() */
-       if (kthread_should_stop())
-               return;
-
-       /* Tag region as quorum only after thread reaches steady state */
-       if (atomic_read(&reg->hr_steady_iterations) != 0)
-               return;
-
-       spin_lock(&o2hb_live_lock);
-
-       if (test_bit(reg->hr_region_num, o2hb_quorum_region_bitmap))
-               goto unlock;
-
-       /*
-        * A region can be added to the quorum only when it sees all
-        * live nodes heartbeat on it. In other words, the region has been
-        * added to all nodes.
-        */
-       if (memcmp(reg->hr_live_node_bitmap, o2hb_live_node_bitmap,
-                  sizeof(o2hb_live_node_bitmap)))
-               goto unlock;
-
-       printk(KERN_NOTICE "o2hb: Region %s (%s) is now a quorum device\n",
-              config_item_name(&reg->hr_item), reg->hr_dev_name);
-
-       set_bit(reg->hr_region_num, o2hb_quorum_region_bitmap);
-
-       /*
-        * If global heartbeat active, unpin all regions if the
-        * region count > CUT_OFF
-        */
-       if (o2hb_pop_count(&o2hb_quorum_region_bitmap,
-                          O2NM_MAX_REGIONS) > O2HB_PIN_CUT_OFF)
-               o2hb_region_unpin(NULL);
-unlock:
-       spin_unlock(&o2hb_live_lock);
-}
-
-static int o2hb_check_slot(struct o2hb_region *reg,
-                          struct o2hb_disk_slot *slot)
-{
-       int changed = 0, gen_changed = 0;
-       struct o2hb_node_event event =
-               { .hn_item = LIST_HEAD_INIT(event.hn_item), };
-       struct o2nm_node *node;
-       struct o2hb_disk_heartbeat_block *hb_block = reg->hr_tmp_block;
-       u64 cputime;
-       unsigned int dead_ms = o2hb_dead_threshold * O2HB_REGION_TIMEOUT_MS;
-       unsigned int slot_dead_ms;
-       int tmp;
-
-       memcpy(hb_block, slot->ds_raw_block, reg->hr_block_bytes);
-
-       /*
-        * If a node is no longer configured but is still in the livemap, we
-        * may need to clear that bit from the livemap.
-        */
-       node = o2nm_get_node_by_num(slot->ds_node_num);
-       if (!node) {
-               spin_lock(&o2hb_live_lock);
-               tmp = test_bit(slot->ds_node_num, o2hb_live_node_bitmap);
-               spin_unlock(&o2hb_live_lock);
-               if (!tmp)
-                       return 0;
-       }
-
-       if (!o2hb_verify_crc(reg, hb_block)) {
-               /* all paths from here will drop o2hb_live_lock for
-                * us. */
-               spin_lock(&o2hb_live_lock);
-
-               /* Don't print an error on the console in this case -
-                * a freshly formatted heartbeat area will not have a
-                * crc set on it. */
-               if (list_empty(&slot->ds_live_item))
-                       goto out;
-
-               /* The node is live but pushed out a bad crc. We
-                * consider it a transient miss but don't populate any
-                * other values as they may be junk. */
-               mlog(ML_ERROR, "Node %d has written a bad crc to %s\n",
-                    slot->ds_node_num, reg->hr_dev_name);
-               o2hb_dump_slot(hb_block);
-
-               slot->ds_equal_samples++;
-               goto fire_callbacks;
-       }
-
-       /* we don't care if these wrap.. the state transitions below
-        * clear at the right places */
-       cputime = le64_to_cpu(hb_block->hb_seq);
-       if (slot->ds_last_time != cputime)
-               slot->ds_changed_samples++;
-       else
-               slot->ds_equal_samples++;
-       slot->ds_last_time = cputime;
-
-       /* The node changed heartbeat generations. We assume this to
-        * mean it dropped off but came back before we timed out. We
-        * want to consider it down for the time being but don't want
-        * to lose any changed_samples state we might build up to
-        * considering it live again. */
-       if (slot->ds_last_generation != le64_to_cpu(hb_block->hb_generation)) {
-               gen_changed = 1;
-               slot->ds_equal_samples = 0;
-               mlog(ML_HEARTBEAT, "Node %d changed generation (0x%llx "
-                    "to 0x%llx)\n", slot->ds_node_num,
-                    (long long)slot->ds_last_generation,
-                    (long long)le64_to_cpu(hb_block->hb_generation));
-       }
-
-       slot->ds_last_generation = le64_to_cpu(hb_block->hb_generation);
-
-       mlog(ML_HEARTBEAT, "Slot %d gen 0x%llx cksum 0x%x "
-            "seq %llu last %llu changed %u equal %u\n",
-            slot->ds_node_num, (long long)slot->ds_last_generation,
-            le32_to_cpu(hb_block->hb_cksum),
-            (unsigned long long)le64_to_cpu(hb_block->hb_seq),
-            (unsigned long long)slot->ds_last_time, slot->ds_changed_samples,
-            slot->ds_equal_samples);
-
-       spin_lock(&o2hb_live_lock);
-
-fire_callbacks:
-       /* dead nodes only come to life after some number of
-        * changes at any time during their dead time */
-       if (list_empty(&slot->ds_live_item) &&
-           slot->ds_changed_samples >= O2HB_LIVE_THRESHOLD) {
-               mlog(ML_HEARTBEAT, "Node %d (id 0x%llx) joined my region\n",
-                    slot->ds_node_num, (long long)slot->ds_last_generation);
-
-               set_bit(slot->ds_node_num, reg->hr_live_node_bitmap);
-
-               /* first on the list generates a callback */
-               if (list_empty(&o2hb_live_slots[slot->ds_node_num])) {
-                       mlog(ML_HEARTBEAT, "o2hb: Add node %d to live nodes "
-                            "bitmap\n", slot->ds_node_num);
-                       set_bit(slot->ds_node_num, o2hb_live_node_bitmap);
-
-                       o2hb_queue_node_event(&event, O2HB_NODE_UP_CB, node,
-                                             slot->ds_node_num);
-
-                       changed = 1;
-               }
-
-               list_add_tail(&slot->ds_live_item,
-                             &o2hb_live_slots[slot->ds_node_num]);
-
-               slot->ds_equal_samples = 0;
-
-               /* We want to be sure that all nodes agree on the
-                * number of milliseconds before a node will be
-                * considered dead. The self-fencing timeout is
-                * computed from this value, and a discrepancy might
-                * result in heartbeat calling a node dead when it
-                * hasn't self-fenced yet. */
-               slot_dead_ms = le32_to_cpu(hb_block->hb_dead_ms);
-               if (slot_dead_ms && slot_dead_ms != dead_ms) {
-                       /* TODO: Perhaps we can fail the region here. */
-                       mlog(ML_ERROR, "Node %d on device %s has a dead count "
-                            "of %u ms, but our count is %u ms.\n"
-                            "Please double check your configuration values "
-                            "for 'O2CB_HEARTBEAT_THRESHOLD'\n",
-                            slot->ds_node_num, reg->hr_dev_name, slot_dead_ms,
-                            dead_ms);
-               }
-               goto out;
-       }
-
-       /* if the list is dead, we're done.. */
-       if (list_empty(&slot->ds_live_item))
-               goto out;
-
-       /* live nodes only go dead after enough consequtive missed
-        * samples..  reset the missed counter whenever we see
-        * activity */
-       if (slot->ds_equal_samples >= o2hb_dead_threshold || gen_changed) {
-               mlog(ML_HEARTBEAT, "Node %d left my region\n",
-                    slot->ds_node_num);
-
-               clear_bit(slot->ds_node_num, reg->hr_live_node_bitmap);
-
-               /* last off the live_slot generates a callback */
-               list_del_init(&slot->ds_live_item);
-               if (list_empty(&o2hb_live_slots[slot->ds_node_num])) {
-                       mlog(ML_HEARTBEAT, "o2hb: Remove node %d from live "
-                            "nodes bitmap\n", slot->ds_node_num);
-                       clear_bit(slot->ds_node_num, o2hb_live_node_bitmap);
-
-                       /* node can be null */
-                       o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB,
-                                             node, slot->ds_node_num);
-
-                       changed = 1;
-               }
-
-               /* We don't clear this because the node is still
-                * actually writing new blocks. */
-               if (!gen_changed)
-                       slot->ds_changed_samples = 0;
-               goto out;
-       }
-       if (slot->ds_changed_samples) {
-               slot->ds_changed_samples = 0;
-               slot->ds_equal_samples = 0;
-       }
-out:
-       spin_unlock(&o2hb_live_lock);
-
-       o2hb_run_event_list(&event);
-
-       if (node)
-               o2nm_node_put(node);
-       return changed;
-}
-
-/* This could be faster if we just implmented a find_last_bit, but I
- * don't think the circumstances warrant it. */
-static int o2hb_highest_node(unsigned long *nodes,
-                            int numbits)
-{
-       int highest, node;
-
-       highest = numbits;
-       node = -1;
-       while ((node = find_next_bit(nodes, numbits, node + 1)) != -1) {
-               if (node >= numbits)
-                       break;
-
-               highest = node;
-       }
-
-       return highest;
-}
-
-static int o2hb_do_disk_heartbeat(struct o2hb_region *reg)
-{
-       int i, ret, highest_node;
-       int membership_change = 0, own_slot_ok = 0;
-       unsigned long configured_nodes[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       unsigned long live_node_bitmap[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       struct o2hb_bio_wait_ctxt write_wc;
-
-       ret = o2nm_configured_node_map(configured_nodes,
-                                      sizeof(configured_nodes));
-       if (ret) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       /*
-        * If a node is not configured but is in the livemap, we still need
-        * to read the slot so as to be able to remove it from the livemap.
-        */
-       o2hb_fill_node_map(live_node_bitmap, sizeof(live_node_bitmap));
-       i = -1;
-       while ((i = find_next_bit(live_node_bitmap,
-                                 O2NM_MAX_NODES, i + 1)) < O2NM_MAX_NODES) {
-               set_bit(i, configured_nodes);
-       }
-
-       highest_node = o2hb_highest_node(configured_nodes, O2NM_MAX_NODES);
-       if (highest_node >= O2NM_MAX_NODES) {
-               mlog(ML_NOTICE, "o2hb: No configured nodes found!\n");
-               ret = -EINVAL;
-               goto bail;
-       }
-
-       /* No sense in reading the slots of nodes that don't exist
-        * yet. Of course, if the node definitions have holes in them
-        * then we're reading an empty slot anyway... Consider this
-        * best-effort. */
-       ret = o2hb_read_slots(reg, highest_node + 1);
-       if (ret < 0) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       /* With an up to date view of the slots, we can check that no
-        * other node has been improperly configured to heartbeat in
-        * our slot. */
-       own_slot_ok = o2hb_check_own_slot(reg);
-
-       /* fill in the proper info for our next heartbeat */
-       o2hb_prepare_block(reg, reg->hr_generation);
-
-       ret = o2hb_issue_node_write(reg, &write_wc);
-       if (ret < 0) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       i = -1;
-       while((i = find_next_bit(configured_nodes,
-                                O2NM_MAX_NODES, i + 1)) < O2NM_MAX_NODES) {
-               membership_change |= o2hb_check_slot(reg, &reg->hr_slots[i]);
-       }
-
-       /*
-        * We have to be sure we've advertised ourselves on disk
-        * before we can go to steady state.  This ensures that
-        * people we find in our steady state have seen us.
-        */
-       o2hb_wait_on_io(reg, &write_wc);
-       if (write_wc.wc_error) {
-               /* Do not re-arm the write timeout on I/O error - we
-                * can't be sure that the new block ever made it to
-                * disk */
-               mlog(ML_ERROR, "Write error %d on device \"%s\"\n",
-                    write_wc.wc_error, reg->hr_dev_name);
-               ret = write_wc.wc_error;
-               goto bail;
-       }
-
-       /* Skip disarming the timeout if own slot has stale/bad data */
-       if (own_slot_ok) {
-               o2hb_set_quorum_device(reg);
-               o2hb_arm_write_timeout(reg);
-       }
-
-bail:
-       /* let the person who launched us know when things are steady */
-       if (atomic_read(&reg->hr_steady_iterations) != 0) {
-               if (!ret && own_slot_ok && !membership_change) {
-                       if (atomic_dec_and_test(&reg->hr_steady_iterations))
-                               wake_up(&o2hb_steady_queue);
-               }
-       }
-
-       if (atomic_read(&reg->hr_steady_iterations) != 0) {
-               if (atomic_dec_and_test(&reg->hr_unsteady_iterations)) {
-                       printk(KERN_NOTICE "o2hb: Unable to stabilize "
-                              "heartbeart on region %s (%s)\n",
-                              config_item_name(&reg->hr_item),
-                              reg->hr_dev_name);
-                       atomic_set(&reg->hr_steady_iterations, 0);
-                       reg->hr_aborted_start = 1;
-                       wake_up(&o2hb_steady_queue);
-                       ret = -EIO;
-               }
-       }
-
-       return ret;
-}
-
-/* Subtract b from a, storing the result in a. a *must* have a larger
- * value than b. */
-static void o2hb_tv_subtract(struct timeval *a,
-                            struct timeval *b)
-{
-       /* just return 0 when a is after b */
-       if (a->tv_sec < b->tv_sec ||
-           (a->tv_sec == b->tv_sec && a->tv_usec < b->tv_usec)) {
-               a->tv_sec = 0;
-               a->tv_usec = 0;
-               return;
-       }
-
-       a->tv_sec -= b->tv_sec;
-       a->tv_usec -= b->tv_usec;
-       while ( a->tv_usec < 0 ) {
-               a->tv_sec--;
-               a->tv_usec += 1000000;
-       }
-}
-
-static unsigned int o2hb_elapsed_msecs(struct timeval *start,
-                                      struct timeval *end)
-{
-       struct timeval res = *end;
-
-       o2hb_tv_subtract(&res, start);
-
-       return res.tv_sec * 1000 + res.tv_usec / 1000;
-}
-
-/*
- * we ride the region ref that the region dir holds.  before the region
- * dir is removed and drops it ref it will wait to tear down this
- * thread.
- */
-static int o2hb_thread(void *data)
-{
-       int i, ret;
-       struct o2hb_region *reg = data;
-       struct o2hb_bio_wait_ctxt write_wc;
-       struct timeval before_hb, after_hb;
-       unsigned int elapsed_msec;
-
-       mlog(ML_HEARTBEAT|ML_KTHREAD, "hb thread running\n");
-
-       set_user_nice(current, -20);
-
-       /* Pin node */
-       o2nm_depend_this_node();
-
-       while (!kthread_should_stop() &&
-              !reg->hr_unclean_stop && !reg->hr_aborted_start) {
-               /* We track the time spent inside
-                * o2hb_do_disk_heartbeat so that we avoid more than
-                * hr_timeout_ms between disk writes. On busy systems
-                * this should result in a heartbeat which is less
-                * likely to time itself out. */
-               do_gettimeofday(&before_hb);
-
-               ret = o2hb_do_disk_heartbeat(reg);
-
-               do_gettimeofday(&after_hb);
-               elapsed_msec = o2hb_elapsed_msecs(&before_hb, &after_hb);
-
-               mlog(ML_HEARTBEAT,
-                    "start = %lu.%lu, end = %lu.%lu, msec = %u\n",
-                    before_hb.tv_sec, (unsigned long) before_hb.tv_usec,
-                    after_hb.tv_sec, (unsigned long) after_hb.tv_usec,
-                    elapsed_msec);
-
-               if (!kthread_should_stop() &&
-                   elapsed_msec < reg->hr_timeout_ms) {
-                       /* the kthread api has blocked signals for us so no
-                        * need to record the return value. */
-                       msleep_interruptible(reg->hr_timeout_ms - elapsed_msec);
-               }
-       }
-
-       o2hb_disarm_write_timeout(reg);
-
-       /* unclean stop is only used in very bad situation */
-       for(i = 0; !reg->hr_unclean_stop && i < reg->hr_blocks; i++)
-               o2hb_shutdown_slot(&reg->hr_slots[i]);
-
-       /* Explicit down notification - avoid forcing the other nodes
-        * to timeout on this region when we could just as easily
-        * write a clear generation - thus indicating to them that
-        * this node has left this region.
-        */
-       if (!reg->hr_unclean_stop && !reg->hr_aborted_start) {
-               o2hb_prepare_block(reg, 0);
-               ret = o2hb_issue_node_write(reg, &write_wc);
-               if (ret == 0)
-                       o2hb_wait_on_io(reg, &write_wc);
-               else
-                       mlog_errno(ret);
-       }
-
-       /* Unpin node */
-       o2nm_undepend_this_node();
-
-       mlog(ML_HEARTBEAT|ML_KTHREAD, "o2hb thread exiting\n");
-
-       return 0;
-}
-
-#ifdef CONFIG_DEBUG_FS
-static int o2hb_debug_open(struct inode *inode, struct file *file)
-{
-       struct o2hb_debug_buf *db = inode->i_private;
-       struct o2hb_region *reg;
-       unsigned long map[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       unsigned long lts;
-       char *buf = NULL;
-       int i = -1;
-       int out = 0;
-
-       /* max_nodes should be the largest bitmap we pass here */
-       BUG_ON(sizeof(map) < db->db_size);
-
-       buf = kmalloc(PAGE_SIZE, GFP_KERNEL);
-       if (!buf)
-               goto bail;
-
-       switch (db->db_type) {
-       case O2HB_DB_TYPE_LIVENODES:
-       case O2HB_DB_TYPE_LIVEREGIONS:
-       case O2HB_DB_TYPE_QUORUMREGIONS:
-       case O2HB_DB_TYPE_FAILEDREGIONS:
-               spin_lock(&o2hb_live_lock);
-               memcpy(map, db->db_data, db->db_size);
-               spin_unlock(&o2hb_live_lock);
-               break;
-
-       case O2HB_DB_TYPE_REGION_LIVENODES:
-               spin_lock(&o2hb_live_lock);
-               reg = (struct o2hb_region *)db->db_data;
-               memcpy(map, reg->hr_live_node_bitmap, db->db_size);
-               spin_unlock(&o2hb_live_lock);
-               break;
-
-       case O2HB_DB_TYPE_REGION_NUMBER:
-               reg = (struct o2hb_region *)db->db_data;
-               out += snprintf(buf + out, PAGE_SIZE - out, "%d\n",
-                               reg->hr_region_num);
-               goto done;
-
-       case O2HB_DB_TYPE_REGION_ELAPSED_TIME:
-               reg = (struct o2hb_region *)db->db_data;
-               lts = reg->hr_last_timeout_start;
-               /* If 0, it has never been set before */
-               if (lts)
-                       lts = jiffies_to_msecs(jiffies - lts);
-               out += snprintf(buf + out, PAGE_SIZE - out, "%lu\n", lts);
-               goto done;
-
-       case O2HB_DB_TYPE_REGION_PINNED:
-               reg = (struct o2hb_region *)db->db_data;
-               out += snprintf(buf + out, PAGE_SIZE - out, "%u\n",
-                               !!reg->hr_item_pinned);
-               goto done;
-
-       default:
-               goto done;
-       }
-
-       while ((i = find_next_bit(map, db->db_len, i + 1)) < db->db_len)
-               out += snprintf(buf + out, PAGE_SIZE - out, "%d ", i);
-       out += snprintf(buf + out, PAGE_SIZE - out, "\n");
-
-done:
-       i_size_write(inode, out);
-
-       file->private_data = buf;
-
-       return 0;
-bail:
-       return -ENOMEM;
-}
-
-static int o2hb_debug_release(struct inode *inode, struct file *file)
-{
-       kfree(file->private_data);
-       return 0;
-}
-
-static ssize_t o2hb_debug_read(struct file *file, char __user *buf,
-                                size_t nbytes, loff_t *ppos)
-{
-       return simple_read_from_buffer(buf, nbytes, ppos, file->private_data,
-                                      i_size_read(file->f_mapping->host));
-}
-#else
-static int o2hb_debug_open(struct inode *inode, struct file *file)
-{
-       return 0;
-}
-static int o2hb_debug_release(struct inode *inode, struct file *file)
-{
-       return 0;
-}
-static ssize_t o2hb_debug_read(struct file *file, char __user *buf,
-                              size_t nbytes, loff_t *ppos)
-{
-       return 0;
-}
-#endif  /* CONFIG_DEBUG_FS */
-
-static const struct file_operations o2hb_debug_fops = {
-       .open =         o2hb_debug_open,
-       .release =      o2hb_debug_release,
-       .read =         o2hb_debug_read,
-       .llseek =       generic_file_llseek,
-};
-
-void o2hb_exit(void)
-{
-       kfree(o2hb_db_livenodes);
-       kfree(o2hb_db_liveregions);
-       kfree(o2hb_db_quorumregions);
-       kfree(o2hb_db_failedregions);
-       debugfs_remove(o2hb_debug_failedregions);
-       debugfs_remove(o2hb_debug_quorumregions);
-       debugfs_remove(o2hb_debug_liveregions);
-       debugfs_remove(o2hb_debug_livenodes);
-       debugfs_remove(o2hb_debug_dir);
-}
-
-static struct dentry *o2hb_debug_create(const char *name, struct dentry *dir,
-                                       struct o2hb_debug_buf **db, int db_len,
-                                       int type, int size, int len, void *data)
-{
-       *db = kmalloc(db_len, GFP_KERNEL);
-       if (!*db)
-               return NULL;
-
-       (*db)->db_type = type;
-       (*db)->db_size = size;
-       (*db)->db_len = len;
-       (*db)->db_data = data;
-
-       return debugfs_create_file(name, S_IFREG|S_IRUSR, dir, *db,
-                                  &o2hb_debug_fops);
-}
-
-static int o2hb_debug_init(void)
-{
-       int ret = -ENOMEM;
-
-       o2hb_debug_dir = debugfs_create_dir(O2HB_DEBUG_DIR, NULL);
-       if (!o2hb_debug_dir) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       o2hb_debug_livenodes = o2hb_debug_create(O2HB_DEBUG_LIVENODES,
-                                                o2hb_debug_dir,
-                                                &o2hb_db_livenodes,
-                                                sizeof(*o2hb_db_livenodes),
-                                                O2HB_DB_TYPE_LIVENODES,
-                                                sizeof(o2hb_live_node_bitmap),
-                                                O2NM_MAX_NODES,
-                                                o2hb_live_node_bitmap);
-       if (!o2hb_debug_livenodes) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       o2hb_debug_liveregions = o2hb_debug_create(O2HB_DEBUG_LIVEREGIONS,
-                                                  o2hb_debug_dir,
-                                                  &o2hb_db_liveregions,
-                                                  sizeof(*o2hb_db_liveregions),
-                                                  O2HB_DB_TYPE_LIVEREGIONS,
-                                                  sizeof(o2hb_live_region_bitmap),
-                                                  O2NM_MAX_REGIONS,
-                                                  o2hb_live_region_bitmap);
-       if (!o2hb_debug_liveregions) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       o2hb_debug_quorumregions =
-                       o2hb_debug_create(O2HB_DEBUG_QUORUMREGIONS,
-                                         o2hb_debug_dir,
-                                         &o2hb_db_quorumregions,
-                                         sizeof(*o2hb_db_quorumregions),
-                                         O2HB_DB_TYPE_QUORUMREGIONS,
-                                         sizeof(o2hb_quorum_region_bitmap),
-                                         O2NM_MAX_REGIONS,
-                                         o2hb_quorum_region_bitmap);
-       if (!o2hb_debug_quorumregions) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       o2hb_debug_failedregions =
-                       o2hb_debug_create(O2HB_DEBUG_FAILEDREGIONS,
-                                         o2hb_debug_dir,
-                                         &o2hb_db_failedregions,
-                                         sizeof(*o2hb_db_failedregions),
-                                         O2HB_DB_TYPE_FAILEDREGIONS,
-                                         sizeof(o2hb_failed_region_bitmap),
-                                         O2NM_MAX_REGIONS,
-                                         o2hb_failed_region_bitmap);
-       if (!o2hb_debug_failedregions) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       ret = 0;
-bail:
-       if (ret)
-               o2hb_exit();
-
-       return ret;
-}
-
-int o2hb_init(void)
-{
-       int i;
-
-       for (i = 0; i < ARRAY_SIZE(o2hb_callbacks); i++)
-               INIT_LIST_HEAD(&o2hb_callbacks[i].list);
-
-       for (i = 0; i < ARRAY_SIZE(o2hb_live_slots); i++)
-               INIT_LIST_HEAD(&o2hb_live_slots[i]);
-
-       INIT_LIST_HEAD(&o2hb_node_events);
-
-       memset(o2hb_live_node_bitmap, 0, sizeof(o2hb_live_node_bitmap));
-       memset(o2hb_region_bitmap, 0, sizeof(o2hb_region_bitmap));
-       memset(o2hb_live_region_bitmap, 0, sizeof(o2hb_live_region_bitmap));
-       memset(o2hb_quorum_region_bitmap, 0, sizeof(o2hb_quorum_region_bitmap));
-       memset(o2hb_failed_region_bitmap, 0, sizeof(o2hb_failed_region_bitmap));
-
-       o2hb_dependent_users = 0;
-
-       return o2hb_debug_init();
-}
-
-/* if we're already in a callback then we're already serialized by the sem */
-static void o2hb_fill_node_map_from_callback(unsigned long *map,
-                                            unsigned bytes)
-{
-       BUG_ON(bytes < (BITS_TO_LONGS(O2NM_MAX_NODES) * sizeof(unsigned long)));
-
-       memcpy(map, &o2hb_live_node_bitmap, bytes);
-}
-
-/*
- * get a map of all nodes that are heartbeating in any regions
- */
-void o2hb_fill_node_map(unsigned long *map, unsigned bytes)
-{
-       /* callers want to serialize this map and callbacks so that they
-        * can trust that they don't miss nodes coming to the party */
-       down_read(&o2hb_callback_sem);
-       spin_lock(&o2hb_live_lock);
-       o2hb_fill_node_map_from_callback(map, bytes);
-       spin_unlock(&o2hb_live_lock);
-       up_read(&o2hb_callback_sem);
-}
-EXPORT_SYMBOL_GPL(o2hb_fill_node_map);
-
-/*
- * heartbeat configfs bits.  The heartbeat set is a default set under
- * the cluster set in nodemanager.c.
- */
-
-static struct o2hb_region *to_o2hb_region(struct config_item *item)
-{
-       return item ? container_of(item, struct o2hb_region, hr_item) : NULL;
-}
-
-/* drop_item only drops its ref after killing the thread, nothing should
- * be using the region anymore.  this has to clean up any state that
- * attributes might have built up. */
-static void o2hb_region_release(struct config_item *item)
-{
-       int i;
-       struct page *page;
-       struct o2hb_region *reg = to_o2hb_region(item);
-
-       mlog(ML_HEARTBEAT, "hb region release (%s)\n", reg->hr_dev_name);
-
-       if (reg->hr_tmp_block)
-               kfree(reg->hr_tmp_block);
-
-       if (reg->hr_slot_data) {
-               for (i = 0; i < reg->hr_num_pages; i++) {
-                       page = reg->hr_slot_data[i];
-                       if (page)
-                               __free_page(page);
-               }
-               kfree(reg->hr_slot_data);
-       }
-
-       if (reg->hr_bdev)
-               blkdev_put(reg->hr_bdev, FMODE_READ|FMODE_WRITE);
-
-       if (reg->hr_slots)
-               kfree(reg->hr_slots);
-
-       kfree(reg->hr_db_regnum);
-       kfree(reg->hr_db_livenodes);
-       debugfs_remove(reg->hr_debug_livenodes);
-       debugfs_remove(reg->hr_debug_regnum);
-       debugfs_remove(reg->hr_debug_elapsed_time);
-       debugfs_remove(reg->hr_debug_pinned);
-       debugfs_remove(reg->hr_debug_dir);
-
-       spin_lock(&o2hb_live_lock);
-       list_del(&reg->hr_all_item);
-       spin_unlock(&o2hb_live_lock);
-
-       kfree(reg);
-}
-
-static int o2hb_read_block_input(struct o2hb_region *reg,
-                                const char *page,
-                                size_t count,
-                                unsigned long *ret_bytes,
-                                unsigned int *ret_bits)
-{
-       unsigned long bytes;
-       char *p = (char *)page;
-
-       bytes = simple_strtoul(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       /* Heartbeat and fs min / max block sizes are the same. */
-       if (bytes > 4096 || bytes < 512)
-               return -ERANGE;
-       if (hweight16(bytes) != 1)
-               return -EINVAL;
-
-       if (ret_bytes)
-               *ret_bytes = bytes;
-       if (ret_bits)
-               *ret_bits = ffs(bytes) - 1;
-
-       return 0;
-}
-
-static ssize_t o2hb_region_block_bytes_read(struct o2hb_region *reg,
-                                           char *page)
-{
-       return sprintf(page, "%u\n", reg->hr_block_bytes);
-}
-
-static ssize_t o2hb_region_block_bytes_write(struct o2hb_region *reg,
-                                            const char *page,
-                                            size_t count)
-{
-       int status;
-       unsigned long block_bytes;
-       unsigned int block_bits;
-
-       if (reg->hr_bdev)
-               return -EINVAL;
-
-       status = o2hb_read_block_input(reg, page, count,
-                                      &block_bytes, &block_bits);
-       if (status)
-               return status;
-
-       reg->hr_block_bytes = (unsigned int)block_bytes;
-       reg->hr_block_bits = block_bits;
-
-       return count;
-}
-
-static ssize_t o2hb_region_start_block_read(struct o2hb_region *reg,
-                                           char *page)
-{
-       return sprintf(page, "%llu\n", reg->hr_start_block);
-}
-
-static ssize_t o2hb_region_start_block_write(struct o2hb_region *reg,
-                                            const char *page,
-                                            size_t count)
-{
-       unsigned long long tmp;
-       char *p = (char *)page;
-
-       if (reg->hr_bdev)
-               return -EINVAL;
-
-       tmp = simple_strtoull(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       reg->hr_start_block = tmp;
-
-       return count;
-}
-
-static ssize_t o2hb_region_blocks_read(struct o2hb_region *reg,
-                                      char *page)
-{
-       return sprintf(page, "%d\n", reg->hr_blocks);
-}
-
-static ssize_t o2hb_region_blocks_write(struct o2hb_region *reg,
-                                       const char *page,
-                                       size_t count)
-{
-       unsigned long tmp;
-       char *p = (char *)page;
-
-       if (reg->hr_bdev)
-               return -EINVAL;
-
-       tmp = simple_strtoul(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       if (tmp > O2NM_MAX_NODES || tmp == 0)
-               return -ERANGE;
-
-       reg->hr_blocks = (unsigned int)tmp;
-
-       return count;
-}
-
-static ssize_t o2hb_region_dev_read(struct o2hb_region *reg,
-                                   char *page)
-{
-       unsigned int ret = 0;
-
-       if (reg->hr_bdev)
-               ret = sprintf(page, "%s\n", reg->hr_dev_name);
-
-       return ret;
-}
-
-static void o2hb_init_region_params(struct o2hb_region *reg)
-{
-       reg->hr_slots_per_page = PAGE_CACHE_SIZE >> reg->hr_block_bits;
-       reg->hr_timeout_ms = O2HB_REGION_TIMEOUT_MS;
-
-       mlog(ML_HEARTBEAT, "hr_start_block = %llu, hr_blocks = %u\n",
-            reg->hr_start_block, reg->hr_blocks);
-       mlog(ML_HEARTBEAT, "hr_block_bytes = %u, hr_block_bits = %u\n",
-            reg->hr_block_bytes, reg->hr_block_bits);
-       mlog(ML_HEARTBEAT, "hr_timeout_ms = %u\n", reg->hr_timeout_ms);
-       mlog(ML_HEARTBEAT, "dead threshold = %u\n", o2hb_dead_threshold);
-}
-
-static int o2hb_map_slot_data(struct o2hb_region *reg)
-{
-       int i, j;
-       unsigned int last_slot;
-       unsigned int spp = reg->hr_slots_per_page;
-       struct page *page;
-       char *raw;
-       struct o2hb_disk_slot *slot;
-
-       reg->hr_tmp_block = kmalloc(reg->hr_block_bytes, GFP_KERNEL);
-       if (reg->hr_tmp_block == NULL) {
-               mlog_errno(-ENOMEM);
-               return -ENOMEM;
-       }
-
-       reg->hr_slots = kcalloc(reg->hr_blocks,
-                               sizeof(struct o2hb_disk_slot), GFP_KERNEL);
-       if (reg->hr_slots == NULL) {
-               mlog_errno(-ENOMEM);
-               return -ENOMEM;
-       }
-
-       for(i = 0; i < reg->hr_blocks; i++) {
-               slot = &reg->hr_slots[i];
-               slot->ds_node_num = i;
-               INIT_LIST_HEAD(&slot->ds_live_item);
-               slot->ds_raw_block = NULL;
-       }
-
-       reg->hr_num_pages = (reg->hr_blocks + spp - 1) / spp;
-       mlog(ML_HEARTBEAT, "Going to require %u pages to cover %u blocks "
-                          "at %u blocks per page\n",
-            reg->hr_num_pages, reg->hr_blocks, spp);
-
-       reg->hr_slot_data = kcalloc(reg->hr_num_pages, sizeof(struct page *),
-                                   GFP_KERNEL);
-       if (!reg->hr_slot_data) {
-               mlog_errno(-ENOMEM);
-               return -ENOMEM;
-       }
-
-       for(i = 0; i < reg->hr_num_pages; i++) {
-               page = alloc_page(GFP_KERNEL);
-               if (!page) {
-                       mlog_errno(-ENOMEM);
-                       return -ENOMEM;
-               }
-
-               reg->hr_slot_data[i] = page;
-
-               last_slot = i * spp;
-               raw = page_address(page);
-               for (j = 0;
-                    (j < spp) && ((j + last_slot) < reg->hr_blocks);
-                    j++) {
-                       BUG_ON((j + last_slot) >= reg->hr_blocks);
-
-                       slot = &reg->hr_slots[j + last_slot];
-                       slot->ds_raw_block =
-                               (struct o2hb_disk_heartbeat_block *) raw;
-
-                       raw += reg->hr_block_bytes;
-               }
-       }
-
-       return 0;
-}
-
-/* Read in all the slots available and populate the tracking
- * structures so that we can start with a baseline idea of what's
- * there. */
-static int o2hb_populate_slot_data(struct o2hb_region *reg)
-{
-       int ret, i;
-       struct o2hb_disk_slot *slot;
-       struct o2hb_disk_heartbeat_block *hb_block;
-
-       ret = o2hb_read_slots(reg, reg->hr_blocks);
-       if (ret) {
-               mlog_errno(ret);
-               goto out;
-       }
-
-       /* We only want to get an idea of the values initially in each
-        * slot, so we do no verification - o2hb_check_slot will
-        * actually determine if each configured slot is valid and
-        * whether any values have changed. */
-       for(i = 0; i < reg->hr_blocks; i++) {
-               slot = &reg->hr_slots[i];
-               hb_block = (struct o2hb_disk_heartbeat_block *) slot->ds_raw_block;
-
-               /* Only fill the values that o2hb_check_slot uses to
-                * determine changing slots */
-               slot->ds_last_time = le64_to_cpu(hb_block->hb_seq);
-               slot->ds_last_generation = le64_to_cpu(hb_block->hb_generation);
-       }
-
-out:
-       return ret;
-}
-
-/* this is acting as commit; we set up all of hr_bdev and hr_task or nothing */
-static ssize_t o2hb_region_dev_write(struct o2hb_region *reg,
-                                    const char *page,
-                                    size_t count)
-{
-       struct task_struct *hb_task;
-       long fd;
-       int sectsize;
-       char *p = (char *)page;
-       struct file *filp = NULL;
-       struct inode *inode = NULL;
-       ssize_t ret = -EINVAL;
-       int live_threshold;
-
-       if (reg->hr_bdev)
-               goto out;
-
-       /* We can't heartbeat without having had our node number
-        * configured yet. */
-       if (o2nm_this_node() == O2NM_MAX_NODES)
-               goto out;
-
-       fd = simple_strtol(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               goto out;
-
-       if (fd < 0 || fd >= INT_MAX)
-               goto out;
-
-       filp = fget(fd);
-       if (filp == NULL)
-               goto out;
-
-       if (reg->hr_blocks == 0 || reg->hr_start_block == 0 ||
-           reg->hr_block_bytes == 0)
-               goto out;
-
-       inode = igrab(filp->f_mapping->host);
-       if (inode == NULL)
-               goto out;
-
-       if (!S_ISBLK(inode->i_mode))
-               goto out;
-
-       reg->hr_bdev = I_BDEV(filp->f_mapping->host);
-       ret = blkdev_get(reg->hr_bdev, FMODE_WRITE | FMODE_READ, NULL);
-       if (ret) {
-               reg->hr_bdev = NULL;
-               goto out;
-       }
-       inode = NULL;
-
-       bdevname(reg->hr_bdev, reg->hr_dev_name);
-
-       sectsize = bdev_logical_block_size(reg->hr_bdev);
-       if (sectsize != reg->hr_block_bytes) {
-               mlog(ML_ERROR,
-                    "blocksize %u incorrect for device, expected %d",
-                    reg->hr_block_bytes, sectsize);
-               ret = -EINVAL;
-               goto out;
-       }
-
-       o2hb_init_region_params(reg);
-
-       /* Generation of zero is invalid */
-       do {
-               get_random_bytes(&reg->hr_generation,
-                                sizeof(reg->hr_generation));
-       } while (reg->hr_generation == 0);
-
-       ret = o2hb_map_slot_data(reg);
-       if (ret) {
-               mlog_errno(ret);
-               goto out;
-       }
-
-       ret = o2hb_populate_slot_data(reg);
-       if (ret) {
-               mlog_errno(ret);
-               goto out;
-       }
-
-       INIT_DELAYED_WORK(&reg->hr_write_timeout_work, o2hb_write_timeout);
-
-       /*
-        * A node is considered live after it has beat LIVE_THRESHOLD
-        * times.  We're not steady until we've given them a chance
-        * _after_ our first read.
-        * The default threshold is bare minimum so as to limit the delay
-        * during mounts. For global heartbeat, the threshold doubled for the
-        * first region.
-        */
-       live_threshold = O2HB_LIVE_THRESHOLD;
-       if (o2hb_global_heartbeat_active()) {
-               spin_lock(&o2hb_live_lock);
-               if (o2hb_pop_count(&o2hb_region_bitmap, O2NM_MAX_REGIONS) == 1)
-                       live_threshold <<= 1;
-               spin_unlock(&o2hb_live_lock);
-       }
-       ++live_threshold;
-       atomic_set(&reg->hr_steady_iterations, live_threshold);
-       /* unsteady_iterations is double the steady_iterations */
-       atomic_set(&reg->hr_unsteady_iterations, (live_threshold << 1));
-
-       hb_task = kthread_run(o2hb_thread, reg, "o2hb-%s",
-                             reg->hr_item.ci_name);
-       if (IS_ERR(hb_task)) {
-               ret = PTR_ERR(hb_task);
-               mlog_errno(ret);
-               goto out;
-       }
-
-       spin_lock(&o2hb_live_lock);
-       reg->hr_task = hb_task;
-       spin_unlock(&o2hb_live_lock);
-
-       ret = wait_event_interruptible(o2hb_steady_queue,
-                               atomic_read(&reg->hr_steady_iterations) == 0);
-       if (ret) {
-               atomic_set(&reg->hr_steady_iterations, 0);
-               reg->hr_aborted_start = 1;
-       }
-
-       if (reg->hr_aborted_start) {
-               ret = -EIO;
-               goto out;
-       }
-
-       /* Ok, we were woken.  Make sure it wasn't by drop_item() */
-       spin_lock(&o2hb_live_lock);
-       hb_task = reg->hr_task;
-       if (o2hb_global_heartbeat_active())
-               set_bit(reg->hr_region_num, o2hb_live_region_bitmap);
-       spin_unlock(&o2hb_live_lock);
-
-       if (hb_task)
-               ret = count;
-       else
-               ret = -EIO;
-
-       if (hb_task && o2hb_global_heartbeat_active())
-               printk(KERN_NOTICE "o2hb: Heartbeat started on region %s (%s)\n",
-                      config_item_name(&reg->hr_item), reg->hr_dev_name);
-
-out:
-       if (filp)
-               fput(filp);
-       if (inode)
-               iput(inode);
-       if (ret < 0) {
-               if (reg->hr_bdev) {
-                       blkdev_put(reg->hr_bdev, FMODE_READ|FMODE_WRITE);
-                       reg->hr_bdev = NULL;
-               }
-       }
-       return ret;
-}
-
-static ssize_t o2hb_region_pid_read(struct o2hb_region *reg,
-                                      char *page)
-{
-       pid_t pid = 0;
-
-       spin_lock(&o2hb_live_lock);
-       if (reg->hr_task)
-               pid = task_pid_nr(reg->hr_task);
-       spin_unlock(&o2hb_live_lock);
-
-       if (!pid)
-               return 0;
-
-       return sprintf(page, "%u\n", pid);
-}
-
-struct o2hb_region_attribute {
-       struct configfs_attribute attr;
-       ssize_t (*show)(struct o2hb_region *, char *);
-       ssize_t (*store)(struct o2hb_region *, const char *, size_t);
-};
-
-static struct o2hb_region_attribute o2hb_region_attr_block_bytes = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "block_bytes",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2hb_region_block_bytes_read,
-       .store  = o2hb_region_block_bytes_write,
-};
-
-static struct o2hb_region_attribute o2hb_region_attr_start_block = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "start_block",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2hb_region_start_block_read,
-       .store  = o2hb_region_start_block_write,
-};
-
-static struct o2hb_region_attribute o2hb_region_attr_blocks = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "blocks",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2hb_region_blocks_read,
-       .store  = o2hb_region_blocks_write,
-};
-
-static struct o2hb_region_attribute o2hb_region_attr_dev = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "dev",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2hb_region_dev_read,
-       .store  = o2hb_region_dev_write,
-};
-
-static struct o2hb_region_attribute o2hb_region_attr_pid = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "pid",
-                   .ca_mode = S_IRUGO | S_IRUSR },
-       .show   = o2hb_region_pid_read,
-};
-
-static struct configfs_attribute *o2hb_region_attrs[] = {
-       &o2hb_region_attr_block_bytes.attr,
-       &o2hb_region_attr_start_block.attr,
-       &o2hb_region_attr_blocks.attr,
-       &o2hb_region_attr_dev.attr,
-       &o2hb_region_attr_pid.attr,
-       NULL,
-};
-
-static ssize_t o2hb_region_show(struct config_item *item,
-                               struct configfs_attribute *attr,
-                               char *page)
-{
-       struct o2hb_region *reg = to_o2hb_region(item);
-       struct o2hb_region_attribute *o2hb_region_attr =
-               container_of(attr, struct o2hb_region_attribute, attr);
-       ssize_t ret = 0;
-
-       if (o2hb_region_attr->show)
-               ret = o2hb_region_attr->show(reg, page);
-       return ret;
-}
-
-static ssize_t o2hb_region_store(struct config_item *item,
-                                struct configfs_attribute *attr,
-                                const char *page, size_t count)
-{
-       struct o2hb_region *reg = to_o2hb_region(item);
-       struct o2hb_region_attribute *o2hb_region_attr =
-               container_of(attr, struct o2hb_region_attribute, attr);
-       ssize_t ret = -EINVAL;
-
-       if (o2hb_region_attr->store)
-               ret = o2hb_region_attr->store(reg, page, count);
-       return ret;
-}
-
-static struct configfs_item_operations o2hb_region_item_ops = {
-       .release                = o2hb_region_release,
-       .show_attribute         = o2hb_region_show,
-       .store_attribute        = o2hb_region_store,
-};
-
-static struct config_item_type o2hb_region_type = {
-       .ct_item_ops    = &o2hb_region_item_ops,
-       .ct_attrs       = o2hb_region_attrs,
-       .ct_owner       = THIS_MODULE,
-};
-
-/* heartbeat set */
-
-struct o2hb_heartbeat_group {
-       struct config_group hs_group;
-       /* some stuff? */
-};
-
-static struct o2hb_heartbeat_group *to_o2hb_heartbeat_group(struct config_group *group)
-{
-       return group ?
-               container_of(group, struct o2hb_heartbeat_group, hs_group)
-               : NULL;
-}
-
-static int o2hb_debug_region_init(struct o2hb_region *reg, struct dentry *dir)
-{
-       int ret = -ENOMEM;
-
-       reg->hr_debug_dir =
-               debugfs_create_dir(config_item_name(&reg->hr_item), dir);
-       if (!reg->hr_debug_dir) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       reg->hr_debug_livenodes =
-                       o2hb_debug_create(O2HB_DEBUG_LIVENODES,
-                                         reg->hr_debug_dir,
-                                         &(reg->hr_db_livenodes),
-                                         sizeof(*(reg->hr_db_livenodes)),
-                                         O2HB_DB_TYPE_REGION_LIVENODES,
-                                         sizeof(reg->hr_live_node_bitmap),
-                                         O2NM_MAX_NODES, reg);
-       if (!reg->hr_debug_livenodes) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       reg->hr_debug_regnum =
-                       o2hb_debug_create(O2HB_DEBUG_REGION_NUMBER,
-                                         reg->hr_debug_dir,
-                                         &(reg->hr_db_regnum),
-                                         sizeof(*(reg->hr_db_regnum)),
-                                         O2HB_DB_TYPE_REGION_NUMBER,
-                                         0, O2NM_MAX_NODES, reg);
-       if (!reg->hr_debug_regnum) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       reg->hr_debug_elapsed_time =
-                       o2hb_debug_create(O2HB_DEBUG_REGION_ELAPSED_TIME,
-                                         reg->hr_debug_dir,
-                                         &(reg->hr_db_elapsed_time),
-                                         sizeof(*(reg->hr_db_elapsed_time)),
-                                         O2HB_DB_TYPE_REGION_ELAPSED_TIME,
-                                         0, 0, reg);
-       if (!reg->hr_debug_elapsed_time) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       reg->hr_debug_pinned =
-                       o2hb_debug_create(O2HB_DEBUG_REGION_PINNED,
-                                         reg->hr_debug_dir,
-                                         &(reg->hr_db_pinned),
-                                         sizeof(*(reg->hr_db_pinned)),
-                                         O2HB_DB_TYPE_REGION_PINNED,
-                                         0, 0, reg);
-       if (!reg->hr_debug_pinned) {
-               mlog_errno(ret);
-               goto bail;
-       }
-
-       ret = 0;
-bail:
-       return ret;
-}
-
-static struct config_item *o2hb_heartbeat_group_make_item(struct config_group *group,
-                                                         const char *name)
-{
-       struct o2hb_region *reg = NULL;
-       int ret;
-
-       reg = kzalloc(sizeof(struct o2hb_region), GFP_KERNEL);
-       if (reg == NULL)
-               return ERR_PTR(-ENOMEM);
-
-       if (strlen(name) > O2HB_MAX_REGION_NAME_LEN) {
-               ret = -ENAMETOOLONG;
-               goto free;
-       }
-
-       spin_lock(&o2hb_live_lock);
-       reg->hr_region_num = 0;
-       if (o2hb_global_heartbeat_active()) {
-               reg->hr_region_num = find_first_zero_bit(o2hb_region_bitmap,
-                                                        O2NM_MAX_REGIONS);
-               if (reg->hr_region_num >= O2NM_MAX_REGIONS) {
-                       spin_unlock(&o2hb_live_lock);
-                       ret = -EFBIG;
-                       goto free;
-               }
-               set_bit(reg->hr_region_num, o2hb_region_bitmap);
-       }
-       list_add_tail(&reg->hr_all_item, &o2hb_all_regions);
-       spin_unlock(&o2hb_live_lock);
-
-       config_item_init_type_name(&reg->hr_item, name, &o2hb_region_type);
-
-       ret = o2hb_debug_region_init(reg, o2hb_debug_dir);
-       if (ret) {
-               config_item_put(&reg->hr_item);
-               goto free;
-       }
-
-       return &reg->hr_item;
-free:
-       kfree(reg);
-       return ERR_PTR(ret);
-}
-
-static void o2hb_heartbeat_group_drop_item(struct config_group *group,
-                                          struct config_item *item)
-{
-       struct task_struct *hb_task;
-       struct o2hb_region *reg = to_o2hb_region(item);
-       int quorum_region = 0;
-
-       /* stop the thread when the user removes the region dir */
-       spin_lock(&o2hb_live_lock);
-       hb_task = reg->hr_task;
-       reg->hr_task = NULL;
-       reg->hr_item_dropped = 1;
-       spin_unlock(&o2hb_live_lock);
-
-       if (hb_task)
-               kthread_stop(hb_task);
-
-       if (o2hb_global_heartbeat_active()) {
-               spin_lock(&o2hb_live_lock);
-               clear_bit(reg->hr_region_num, o2hb_region_bitmap);
-               clear_bit(reg->hr_region_num, o2hb_live_region_bitmap);
-               if (test_bit(reg->hr_region_num, o2hb_quorum_region_bitmap))
-                       quorum_region = 1;
-               clear_bit(reg->hr_region_num, o2hb_quorum_region_bitmap);
-               spin_unlock(&o2hb_live_lock);
-               printk(KERN_NOTICE "o2hb: Heartbeat %s on region %s (%s)\n",
-                      ((atomic_read(&reg->hr_steady_iterations) == 0) ?
-                       "stopped" : "start aborted"), config_item_name(item),
-                      reg->hr_dev_name);
-       }
-
-       /*
-        * If we're racing a dev_write(), we need to wake them.  They will
-        * check reg->hr_task
-        */
-       if (atomic_read(&reg->hr_steady_iterations) != 0) {
-               reg->hr_aborted_start = 1;
-               atomic_set(&reg->hr_steady_iterations, 0);
-               wake_up(&o2hb_steady_queue);
-       }
-
-       config_item_put(item);
-
-       if (!o2hb_global_heartbeat_active() || !quorum_region)
-               return;
-
-       /*
-        * If global heartbeat active and there are dependent users,
-        * pin all regions if quorum region count <= CUT_OFF
-        */
-       spin_lock(&o2hb_live_lock);
-
-       if (!o2hb_dependent_users)
-               goto unlock;
-
-       if (o2hb_pop_count(&o2hb_quorum_region_bitmap,
-                          O2NM_MAX_REGIONS) <= O2HB_PIN_CUT_OFF)
-               o2hb_region_pin(NULL);
-
-unlock:
-       spin_unlock(&o2hb_live_lock);
-}
-
-struct o2hb_heartbeat_group_attribute {
-       struct configfs_attribute attr;
-       ssize_t (*show)(struct o2hb_heartbeat_group *, char *);
-       ssize_t (*store)(struct o2hb_heartbeat_group *, const char *, size_t);
-};
-
-static ssize_t o2hb_heartbeat_group_show(struct config_item *item,
-                                        struct configfs_attribute *attr,
-                                        char *page)
-{
-       struct o2hb_heartbeat_group *reg = to_o2hb_heartbeat_group(to_config_group(item));
-       struct o2hb_heartbeat_group_attribute *o2hb_heartbeat_group_attr =
-               container_of(attr, struct o2hb_heartbeat_group_attribute, attr);
-       ssize_t ret = 0;
-
-       if (o2hb_heartbeat_group_attr->show)
-               ret = o2hb_heartbeat_group_attr->show(reg, page);
-       return ret;
-}
-
-static ssize_t o2hb_heartbeat_group_store(struct config_item *item,
-                                         struct configfs_attribute *attr,
-                                         const char *page, size_t count)
-{
-       struct o2hb_heartbeat_group *reg = to_o2hb_heartbeat_group(to_config_group(item));
-       struct o2hb_heartbeat_group_attribute *o2hb_heartbeat_group_attr =
-               container_of(attr, struct o2hb_heartbeat_group_attribute, attr);
-       ssize_t ret = -EINVAL;
-
-       if (o2hb_heartbeat_group_attr->store)
-               ret = o2hb_heartbeat_group_attr->store(reg, page, count);
-       return ret;
-}
-
-static ssize_t o2hb_heartbeat_group_threshold_show(struct o2hb_heartbeat_group *group,
-                                                    char *page)
-{
-       return sprintf(page, "%u\n", o2hb_dead_threshold);
-}
-
-static ssize_t o2hb_heartbeat_group_threshold_store(struct o2hb_heartbeat_group *group,
-                                                   const char *page,
-                                                   size_t count)
-{
-       unsigned long tmp;
-       char *p = (char *)page;
-
-       tmp = simple_strtoul(p, &p, 10);
-       if (!p || (*p && (*p != '\n')))
-                return -EINVAL;
-
-       /* this will validate ranges for us. */
-       o2hb_dead_threshold_set((unsigned int) tmp);
-
-       return count;
-}
-
-static
-ssize_t o2hb_heartbeat_group_mode_show(struct o2hb_heartbeat_group *group,
-                                      char *page)
-{
-       return sprintf(page, "%s\n",
-                      o2hb_heartbeat_mode_desc[o2hb_heartbeat_mode]);
-}
-
-static
-ssize_t o2hb_heartbeat_group_mode_store(struct o2hb_heartbeat_group *group,
-                                       const char *page, size_t count)
-{
-       unsigned int i;
-       int ret;
-       size_t len;
-
-       len = (page[count - 1] == '\n') ? count - 1 : count;
-       if (!len)
-               return -EINVAL;
-
-       for (i = 0; i < O2HB_HEARTBEAT_NUM_MODES; ++i) {
-               if (strnicmp(page, o2hb_heartbeat_mode_desc[i], len))
-                       continue;
-
-               ret = o2hb_global_hearbeat_mode_set(i);
-               if (!ret)
-                       printk(KERN_NOTICE "o2hb: Heartbeat mode set to %s\n",
-                              o2hb_heartbeat_mode_desc[i]);
-               return count;
-       }
-
-       return -EINVAL;
-
-}
-
-static struct o2hb_heartbeat_group_attribute o2hb_heartbeat_group_attr_threshold = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "dead_threshold",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2hb_heartbeat_group_threshold_show,
-       .store  = o2hb_heartbeat_group_threshold_store,
-};
-
-static struct o2hb_heartbeat_group_attribute o2hb_heartbeat_group_attr_mode = {
-       .attr   = { .ca_owner = THIS_MODULE,
-               .ca_name = "mode",
-               .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2hb_heartbeat_group_mode_show,
-       .store  = o2hb_heartbeat_group_mode_store,
-};
-
-static struct configfs_attribute *o2hb_heartbeat_group_attrs[] = {
-       &o2hb_heartbeat_group_attr_threshold.attr,
-       &o2hb_heartbeat_group_attr_mode.attr,
-       NULL,
-};
-
-static struct configfs_item_operations o2hb_hearbeat_group_item_ops = {
-       .show_attribute         = o2hb_heartbeat_group_show,
-       .store_attribute        = o2hb_heartbeat_group_store,
-};
-
-static struct configfs_group_operations o2hb_heartbeat_group_group_ops = {
-       .make_item      = o2hb_heartbeat_group_make_item,
-       .drop_item      = o2hb_heartbeat_group_drop_item,
-};
-
-static struct config_item_type o2hb_heartbeat_group_type = {
-       .ct_group_ops   = &o2hb_heartbeat_group_group_ops,
-       .ct_item_ops    = &o2hb_hearbeat_group_item_ops,
-       .ct_attrs       = o2hb_heartbeat_group_attrs,
-       .ct_owner       = THIS_MODULE,
-};
-
-/* this is just here to avoid touching group in heartbeat.h which the
- * entire damn world #includes */
-struct config_group *o2hb_alloc_hb_set(void)
-{
-       struct o2hb_heartbeat_group *hs = NULL;
-       struct config_group *ret = NULL;
-
-       hs = kzalloc(sizeof(struct o2hb_heartbeat_group), GFP_KERNEL);
-       if (hs == NULL)
-               goto out;
-
-       config_group_init_type_name(&hs->hs_group, "heartbeat",
-                                   &o2hb_heartbeat_group_type);
-
-       ret = &hs->hs_group;
-out:
-       if (ret == NULL)
-               kfree(hs);
-       return ret;
-}
-
-void o2hb_free_hb_set(struct config_group *group)
-{
-       struct o2hb_heartbeat_group *hs = to_o2hb_heartbeat_group(group);
-       kfree(hs);
-}
-
-/* hb callback registration and issuing */
-
-static struct o2hb_callback *hbcall_from_type(enum o2hb_callback_type type)
-{
-       if (type == O2HB_NUM_CB)
-               return ERR_PTR(-EINVAL);
-
-       return &o2hb_callbacks[type];
-}
-
-void o2hb_setup_callback(struct o2hb_callback_func *hc,
-                        enum o2hb_callback_type type,
-                        o2hb_cb_func *func,
-                        void *data,
-                        int priority)
-{
-       INIT_LIST_HEAD(&hc->hc_item);
-       hc->hc_func = func;
-       hc->hc_data = data;
-       hc->hc_priority = priority;
-       hc->hc_type = type;
-       hc->hc_magic = O2HB_CB_MAGIC;
-}
-EXPORT_SYMBOL_GPL(o2hb_setup_callback);
-
-/*
- * In local heartbeat mode, region_uuid passed matches the dlm domain name.
- * In global heartbeat mode, region_uuid passed is NULL.
- *
- * In local, we only pin the matching region. In global we pin all the active
- * regions.
- */
-static int o2hb_region_pin(const char *region_uuid)
-{
-       int ret = 0, found = 0;
-       struct o2hb_region *reg;
-       char *uuid;
-
-       assert_spin_locked(&o2hb_live_lock);
-
-       list_for_each_entry(reg, &o2hb_all_regions, hr_all_item) {
-               uuid = config_item_name(&reg->hr_item);
-
-               /* local heartbeat */
-               if (region_uuid) {
-                       if (strcmp(region_uuid, uuid))
-                               continue;
-                       found = 1;
-               }
-
-               if (reg->hr_item_pinned || reg->hr_item_dropped)
-                       goto skip_pin;
-
-               /* Ignore ENOENT only for local hb (userdlm domain) */
-               ret = o2nm_depend_item(&reg->hr_item);
-               if (!ret) {
-                       mlog(ML_CLUSTER, "Pin region %s\n", uuid);
-                       reg->hr_item_pinned = 1;
-               } else {
-                       if (ret == -ENOENT && found)
-                               ret = 0;
-                       else {
-                               mlog(ML_ERROR, "Pin region %s fails with %d\n",
-                                    uuid, ret);
-                               break;
-                       }
-               }
-skip_pin:
-               if (found)
-                       break;
-       }
-
-       return ret;
-}
-
-/*
- * In local heartbeat mode, region_uuid passed matches the dlm domain name.
- * In global heartbeat mode, region_uuid passed is NULL.
- *
- * In local, we only unpin the matching region. In global we unpin all the
- * active regions.
- */
-static void o2hb_region_unpin(const char *region_uuid)
-{
-       struct o2hb_region *reg;
-       char *uuid;
-       int found = 0;
-
-       assert_spin_locked(&o2hb_live_lock);
-
-       list_for_each_entry(reg, &o2hb_all_regions, hr_all_item) {
-               uuid = config_item_name(&reg->hr_item);
-               if (region_uuid) {
-                       if (strcmp(region_uuid, uuid))
-                               continue;
-                       found = 1;
-               }
-
-               if (reg->hr_item_pinned) {
-                       mlog(ML_CLUSTER, "Unpin region %s\n", uuid);
-                       o2nm_undepend_item(&reg->hr_item);
-                       reg->hr_item_pinned = 0;
-               }
-               if (found)
-                       break;
-       }
-}
-
-static int o2hb_region_inc_user(const char *region_uuid)
-{
-       int ret = 0;
-
-       spin_lock(&o2hb_live_lock);
-
-       /* local heartbeat */
-       if (!o2hb_global_heartbeat_active()) {
-           ret = o2hb_region_pin(region_uuid);
-           goto unlock;
-       }
-
-       /*
-        * if global heartbeat active and this is the first dependent user,
-        * pin all regions if quorum region count <= CUT_OFF
-        */
-       o2hb_dependent_users++;
-       if (o2hb_dependent_users > 1)
-               goto unlock;
-
-       if (o2hb_pop_count(&o2hb_quorum_region_bitmap,
-                          O2NM_MAX_REGIONS) <= O2HB_PIN_CUT_OFF)
-               ret = o2hb_region_pin(NULL);
-
-unlock:
-       spin_unlock(&o2hb_live_lock);
-       return ret;
-}
-
-void o2hb_region_dec_user(const char *region_uuid)
-{
-       spin_lock(&o2hb_live_lock);
-
-       /* local heartbeat */
-       if (!o2hb_global_heartbeat_active()) {
-           o2hb_region_unpin(region_uuid);
-           goto unlock;
-       }
-
-       /*
-        * if global heartbeat active and there are no dependent users,
-        * unpin all quorum regions
-        */
-       o2hb_dependent_users--;
-       if (!o2hb_dependent_users)
-               o2hb_region_unpin(NULL);
-
-unlock:
-       spin_unlock(&o2hb_live_lock);
-}
-
-int o2hb_register_callback(const char *region_uuid,
-                          struct o2hb_callback_func *hc)
-{
-       struct o2hb_callback_func *tmp;
-       struct list_head *iter;
-       struct o2hb_callback *hbcall;
-       int ret;
-
-       BUG_ON(hc->hc_magic != O2HB_CB_MAGIC);
-       BUG_ON(!list_empty(&hc->hc_item));
-
-       hbcall = hbcall_from_type(hc->hc_type);
-       if (IS_ERR(hbcall)) {
-               ret = PTR_ERR(hbcall);
-               goto out;
-       }
-
-       if (region_uuid) {
-               ret = o2hb_region_inc_user(region_uuid);
-               if (ret) {
-                       mlog_errno(ret);
-                       goto out;
-               }
-       }
-
-       down_write(&o2hb_callback_sem);
-
-       list_for_each(iter, &hbcall->list) {
-               tmp = list_entry(iter, struct o2hb_callback_func, hc_item);
-               if (hc->hc_priority < tmp->hc_priority) {
-                       list_add_tail(&hc->hc_item, iter);
-                       break;
-               }
-       }
-       if (list_empty(&hc->hc_item))
-               list_add_tail(&hc->hc_item, &hbcall->list);
-
-       up_write(&o2hb_callback_sem);
-       ret = 0;
-out:
-       mlog(ML_CLUSTER, "returning %d on behalf of %p for funcs %p\n",
-            ret, __builtin_return_address(0), hc);
-       return ret;
-}
-EXPORT_SYMBOL_GPL(o2hb_register_callback);
-
-void o2hb_unregister_callback(const char *region_uuid,
-                             struct o2hb_callback_func *hc)
-{
-       BUG_ON(hc->hc_magic != O2HB_CB_MAGIC);
-
-       mlog(ML_CLUSTER, "on behalf of %p for funcs %p\n",
-            __builtin_return_address(0), hc);
-
-       /* XXX Can this happen _with_ a region reference? */
-       if (list_empty(&hc->hc_item))
-               return;
-
-       if (region_uuid)
-               o2hb_region_dec_user(region_uuid);
-
-       down_write(&o2hb_callback_sem);
-
-       list_del_init(&hc->hc_item);
-
-       up_write(&o2hb_callback_sem);
-}
-EXPORT_SYMBOL_GPL(o2hb_unregister_callback);
-
-int o2hb_check_node_heartbeating(u8 node_num)
-{
-       unsigned long testing_map[BITS_TO_LONGS(O2NM_MAX_NODES)];
-
-       o2hb_fill_node_map(testing_map, sizeof(testing_map));
-       if (!test_bit(node_num, testing_map)) {
-               mlog(ML_HEARTBEAT,
-                    "node (%u) does not have heartbeating enabled.\n",
-                    node_num);
-               return 0;
-       }
-
-       return 1;
-}
-EXPORT_SYMBOL_GPL(o2hb_check_node_heartbeating);
-
-int o2hb_check_node_heartbeating_from_callback(u8 node_num)
-{
-       unsigned long testing_map[BITS_TO_LONGS(O2NM_MAX_NODES)];
-
-       o2hb_fill_node_map_from_callback(testing_map, sizeof(testing_map));
-       if (!test_bit(node_num, testing_map)) {
-               mlog(ML_HEARTBEAT,
-                    "node (%u) does not have heartbeating enabled.\n",
-                    node_num);
-               return 0;
-       }
-
-       return 1;
-}
-EXPORT_SYMBOL_GPL(o2hb_check_node_heartbeating_from_callback);
-
-/* Makes sure our local node is configured with a node number, and is
- * heartbeating. */
-int o2hb_check_local_node_heartbeating(void)
-{
-       u8 node_num;
-
-       /* if this node was set then we have networking */
-       node_num = o2nm_this_node();
-       if (node_num == O2NM_MAX_NODES) {
-               mlog(ML_HEARTBEAT, "this node has not been configured.\n");
-               return 0;
-       }
-
-       return o2hb_check_node_heartbeating(node_num);
-}
-EXPORT_SYMBOL_GPL(o2hb_check_local_node_heartbeating);
-
-/*
- * this is just a hack until we get the plumbing which flips file systems
- * read only and drops the hb ref instead of killing the node dead.
- */
-void o2hb_stop_all_regions(void)
-{
-       struct o2hb_region *reg;
-
-       mlog(ML_ERROR, "stopping heartbeat on all active regions.\n");
-
-       spin_lock(&o2hb_live_lock);
-
-       list_for_each_entry(reg, &o2hb_all_regions, hr_all_item)
-               reg->hr_unclean_stop = 1;
-
-       spin_unlock(&o2hb_live_lock);
-}
-EXPORT_SYMBOL_GPL(o2hb_stop_all_regions);
-
-int o2hb_get_all_regions(char *region_uuids, u8 max_regions)
-{
-       struct o2hb_region *reg;
-       int numregs = 0;
-       char *p;
-
-       spin_lock(&o2hb_live_lock);
-
-       p = region_uuids;
-       list_for_each_entry(reg, &o2hb_all_regions, hr_all_item) {
-               mlog(0, "Region: %s\n", config_item_name(&reg->hr_item));
-               if (numregs < max_regions) {
-                       memcpy(p, config_item_name(&reg->hr_item),
-                              O2HB_MAX_REGION_NAME_LEN);
-                       p += O2HB_MAX_REGION_NAME_LEN;
-               }
-               numregs++;
-       }
-
-       spin_unlock(&o2hb_live_lock);
-
-       return numregs;
-}
-EXPORT_SYMBOL_GPL(o2hb_get_all_regions);
-
-int o2hb_global_heartbeat_active(void)
-{
-       return (o2hb_heartbeat_mode == O2HB_HEARTBEAT_GLOBAL);
-}
-EXPORT_SYMBOL(o2hb_global_heartbeat_active);
-
-#ifdef CONFIG_RAMSTER
-void o2hb_manual_set_node_heartbeating(int node_num)
-{
-       if (node_num < O2NM_MAX_NODES)
-               set_bit(node_num, o2hb_live_node_bitmap);
-}
-EXPORT_SYMBOL(o2hb_manual_set_node_heartbeating);
-#endif
diff --git a/drivers/staging/ramster/cluster/heartbeat.h b/drivers/staging/ramster/cluster/heartbeat.h
deleted file mode 100644 (file)
index cf1a164..0000000
+++ /dev/null
@@ -1,92 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * heartbeat.h
- *
- * Function prototypes
- *
- * Copyright (C) 2004 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifndef O2CLUSTER_HEARTBEAT_H
-#define O2CLUSTER_HEARTBEAT_H
-
-#include "ocfs2_heartbeat.h"
-
-#define O2HB_REGION_TIMEOUT_MS         2000
-
-#define O2HB_MAX_REGION_NAME_LEN       32
-
-/* number of changes to be seen as live */
-#define O2HB_LIVE_THRESHOLD       2
-/* number of equal samples to be seen as dead */
-extern unsigned int o2hb_dead_threshold;
-#define O2HB_DEFAULT_DEAD_THRESHOLD       31
-/* Otherwise MAX_WRITE_TIMEOUT will be zero... */
-#define O2HB_MIN_DEAD_THRESHOLD          2
-#define O2HB_MAX_WRITE_TIMEOUT_MS (O2HB_REGION_TIMEOUT_MS * (o2hb_dead_threshold - 1))
-
-#define O2HB_CB_MAGIC          0x51d1e4ec
-
-/* callback stuff */
-enum o2hb_callback_type {
-       O2HB_NODE_DOWN_CB = 0,
-       O2HB_NODE_UP_CB,
-       O2HB_NUM_CB
-};
-
-struct o2nm_node;
-typedef void (o2hb_cb_func)(struct o2nm_node *, int, void *);
-
-struct o2hb_callback_func {
-       u32                     hc_magic;
-       struct list_head        hc_item;
-       o2hb_cb_func            *hc_func;
-       void                    *hc_data;
-       int                     hc_priority;
-       enum o2hb_callback_type hc_type;
-};
-
-struct config_group *o2hb_alloc_hb_set(void);
-void o2hb_free_hb_set(struct config_group *group);
-
-void o2hb_setup_callback(struct o2hb_callback_func *hc,
-                        enum o2hb_callback_type type,
-                        o2hb_cb_func *func,
-                        void *data,
-                        int priority);
-int o2hb_register_callback(const char *region_uuid,
-                          struct o2hb_callback_func *hc);
-void o2hb_unregister_callback(const char *region_uuid,
-                             struct o2hb_callback_func *hc);
-void o2hb_fill_node_map(unsigned long *map,
-                       unsigned bytes);
-void o2hb_exit(void);
-int o2hb_init(void);
-int o2hb_check_node_heartbeating(u8 node_num);
-int o2hb_check_node_heartbeating_from_callback(u8 node_num);
-int o2hb_check_local_node_heartbeating(void);
-void o2hb_stop_all_regions(void);
-int o2hb_get_all_regions(char *region_uuids, u8 numregions);
-int o2hb_global_heartbeat_active(void);
-#ifdef CONFIG_RAMSTER
-void o2hb_manual_set_node_heartbeating(int);
-#endif
-
-#endif /* O2CLUSTER_HEARTBEAT_H */
diff --git a/drivers/staging/ramster/cluster/masklog.c b/drivers/staging/ramster/cluster/masklog.c
deleted file mode 100644 (file)
index 07ac24f..0000000
+++ /dev/null
@@ -1,155 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2004, 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#include <linux/module.h>
-#include <linux/kernel.h>
-#include <linux/proc_fs.h>
-#include <linux/seq_file.h>
-#include <linux/string.h>
-#include <asm/uaccess.h>
-
-#include "masklog.h"
-
-struct mlog_bits mlog_and_bits = MLOG_BITS_RHS(MLOG_INITIAL_AND_MASK);
-EXPORT_SYMBOL_GPL(mlog_and_bits);
-struct mlog_bits mlog_not_bits = MLOG_BITS_RHS(0);
-EXPORT_SYMBOL_GPL(mlog_not_bits);
-
-static ssize_t mlog_mask_show(u64 mask, char *buf)
-{
-       char *state;
-
-       if (__mlog_test_u64(mask, mlog_and_bits))
-               state = "allow";
-       else if (__mlog_test_u64(mask, mlog_not_bits))
-               state = "deny";
-       else
-               state = "off";
-
-       return snprintf(buf, PAGE_SIZE, "%s\n", state);
-}
-
-static ssize_t mlog_mask_store(u64 mask, const char *buf, size_t count)
-{
-       if (!strnicmp(buf, "allow", 5)) {
-               __mlog_set_u64(mask, mlog_and_bits);
-               __mlog_clear_u64(mask, mlog_not_bits);
-       } else if (!strnicmp(buf, "deny", 4)) {
-               __mlog_set_u64(mask, mlog_not_bits);
-               __mlog_clear_u64(mask, mlog_and_bits);
-       } else if (!strnicmp(buf, "off", 3)) {
-               __mlog_clear_u64(mask, mlog_not_bits);
-               __mlog_clear_u64(mask, mlog_and_bits);
-       } else
-               return -EINVAL;
-
-       return count;
-}
-
-struct mlog_attribute {
-       struct attribute attr;
-       u64 mask;
-};
-
-#define to_mlog_attr(_attr) container_of(_attr, struct mlog_attribute, attr)
-
-#define define_mask(_name) {                   \
-       .attr = {                               \
-               .name = #_name,                 \
-               .mode = S_IRUGO | S_IWUSR,      \
-       },                                      \
-       .mask = ML_##_name,                     \
-}
-
-static struct mlog_attribute mlog_attrs[MLOG_MAX_BITS] = {
-       define_mask(TCP),
-       define_mask(MSG),
-       define_mask(SOCKET),
-       define_mask(HEARTBEAT),
-       define_mask(HB_BIO),
-       define_mask(DLMFS),
-       define_mask(DLM),
-       define_mask(DLM_DOMAIN),
-       define_mask(DLM_THREAD),
-       define_mask(DLM_MASTER),
-       define_mask(DLM_RECOVERY),
-       define_mask(DLM_GLUE),
-       define_mask(VOTE),
-       define_mask(CONN),
-       define_mask(QUORUM),
-       define_mask(BASTS),
-       define_mask(CLUSTER),
-       define_mask(ERROR),
-       define_mask(NOTICE),
-       define_mask(KTHREAD),
-};
-
-static struct attribute *mlog_attr_ptrs[MLOG_MAX_BITS] = {NULL, };
-
-static ssize_t mlog_show(struct kobject *obj, struct attribute *attr,
-                        char *buf)
-{
-       struct mlog_attribute *mlog_attr = to_mlog_attr(attr);
-
-       return mlog_mask_show(mlog_attr->mask, buf);
-}
-
-static ssize_t mlog_store(struct kobject *obj, struct attribute *attr,
-                         const char *buf, size_t count)
-{
-       struct mlog_attribute *mlog_attr = to_mlog_attr(attr);
-
-       return mlog_mask_store(mlog_attr->mask, buf, count);
-}
-
-static const struct sysfs_ops mlog_attr_ops = {
-       .show  = mlog_show,
-       .store = mlog_store,
-};
-
-static struct kobj_type mlog_ktype = {
-       .default_attrs = mlog_attr_ptrs,
-       .sysfs_ops     = &mlog_attr_ops,
-};
-
-static struct kset mlog_kset = {
-       .kobj   = {.ktype = &mlog_ktype},
-};
-
-int mlog_sys_init(struct kset *o2cb_kset)
-{
-       int i = 0;
-
-       while (mlog_attrs[i].attr.mode) {
-               mlog_attr_ptrs[i] = &mlog_attrs[i].attr;
-               i++;
-       }
-       mlog_attr_ptrs[i] = NULL;
-
-       kobject_set_name(&mlog_kset.kobj, "logmask");
-       mlog_kset.kobj.kset = o2cb_kset;
-       return kset_register(&mlog_kset);
-}
-
-void mlog_sys_shutdown(void)
-{
-       kset_unregister(&mlog_kset);
-}
diff --git a/drivers/staging/ramster/cluster/masklog.h b/drivers/staging/ramster/cluster/masklog.h
deleted file mode 100644 (file)
index baa2b9e..0000000
+++ /dev/null
@@ -1,219 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#ifndef O2CLUSTER_MASKLOG_H
-#define O2CLUSTER_MASKLOG_H
-
-/*
- * For now this is a trivial wrapper around printk() that gives the critical
- * ability to enable sets of debugging output at run-time.  In the future this
- * will almost certainly be redirected to relayfs so that it can pay a
- * substantially lower heisenberg tax.
- *
- * Callers associate the message with a bitmask and a global bitmask is
- * maintained with help from /proc.  If any of the bits match the message is
- * output.
- *
- * We must have efficient bit tests on i386 and it seems gcc still emits crazy
- * code for the 64bit compare.  It emits very good code for the dual unsigned
- * long tests, though, completely avoiding tests that can never pass if the
- * caller gives a constant bitmask that fills one of the longs with all 0s.  So
- * the desire is to have almost all of the calls decided on by comparing just
- * one of the longs.  This leads to having infrequently given bits that are
- * frequently matched in the high bits.
- *
- * _ERROR and _NOTICE are used for messages that always go to the console and
- * have appropriate KERN_ prefixes.  We wrap these in our function instead of
- * just calling printk() so that this can eventually make its way through
- * relayfs along with the debugging messages.  Everything else gets KERN_DEBUG.
- * The inline tests and macro dance give GCC the opportunity to quite cleverly
- * only emit the appropriage printk() when the caller passes in a constant
- * mask, as is almost always the case.
- *
- * All this bitmask nonsense is managed from the files under
- * /sys/fs/o2cb/logmask/.  Reading the files gives a straightforward
- * indication of which bits are allowed (allow) or denied (off/deny).
- *     ENTRY deny
- *     EXIT deny
- *     TCP off
- *     MSG off
- *     SOCKET off
- *     ERROR allow
- *     NOTICE allow
- *
- * Writing changes the state of a given bit and requires a strictly formatted
- * single write() call:
- *
- *     write(fd, "allow", 5);
- *
- * Echoing allow/deny/off string into the logmask files can flip the bits
- * on or off as expected; here is the bash script for example:
- *
- * log_mask="/sys/fs/o2cb/log_mask"
- * for node in ENTRY EXIT TCP MSG SOCKET ERROR NOTICE; do
- *     echo allow >"$log_mask"/"$node"
- * done
- *
- * The debugfs.ocfs2 tool can also flip the bits with the -l option:
- *
- * debugfs.ocfs2 -l TCP allow
- */
-
-/* for task_struct */
-#include <linux/sched.h>
-
-/* bits that are frequently given and infrequently matched in the low word */
-/* NOTE: If you add a flag, you need to also update masklog.c! */
-#define ML_TCP         0x0000000000000001ULL /* net cluster/tcp.c */
-#define ML_MSG         0x0000000000000002ULL /* net network messages */
-#define ML_SOCKET      0x0000000000000004ULL /* net socket lifetime */
-#define ML_HEARTBEAT   0x0000000000000008ULL /* hb all heartbeat tracking */
-#define ML_HB_BIO      0x0000000000000010ULL /* hb io tracing */
-#define ML_DLMFS       0x0000000000000020ULL /* dlm user dlmfs */
-#define ML_DLM         0x0000000000000040ULL /* dlm general debugging */
-#define ML_DLM_DOMAIN  0x0000000000000080ULL /* dlm domain debugging */
-#define ML_DLM_THREAD  0x0000000000000100ULL /* dlm domain thread */
-#define ML_DLM_MASTER  0x0000000000000200ULL /* dlm master functions */
-#define ML_DLM_RECOVERY        0x0000000000000400ULL /* dlm master functions */
-#define ML_DLM_GLUE    0x0000000000000800ULL /* ocfs2 dlm glue layer */
-#define ML_VOTE                0x0000000000001000ULL /* ocfs2 node messaging  */
-#define ML_CONN                0x0000000000002000ULL /* net connection management */
-#define ML_QUORUM      0x0000000000004000ULL /* net connection quorum */
-#define ML_BASTS       0x0000000000008000ULL /* dlmglue asts and basts */
-#define ML_CLUSTER     0x0000000000010000ULL /* cluster stack */
-
-/* bits that are infrequently given and frequently matched in the high word */
-#define ML_ERROR       0x1000000000000000ULL /* sent to KERN_ERR */
-#define ML_NOTICE      0x2000000000000000ULL /* setn to KERN_NOTICE */
-#define ML_KTHREAD     0x4000000000000000ULL /* kernel thread activity */
-
-#define MLOG_INITIAL_AND_MASK (ML_ERROR|ML_NOTICE)
-#ifndef MLOG_MASK_PREFIX
-#define MLOG_MASK_PREFIX 0
-#endif
-
-/*
- * When logging is disabled, force the bit test to 0 for anything other
- * than errors and notices, allowing gcc to remove the code completely.
- * When enabled, allow all masks.
- */
-#if defined(CONFIG_OCFS2_DEBUG_MASKLOG)
-#define ML_ALLOWED_BITS ~0
-#else
-#define ML_ALLOWED_BITS (ML_ERROR|ML_NOTICE)
-#endif
-
-#define MLOG_MAX_BITS 64
-
-struct mlog_bits {
-       unsigned long words[MLOG_MAX_BITS / BITS_PER_LONG];
-};
-
-extern struct mlog_bits mlog_and_bits, mlog_not_bits;
-
-#if BITS_PER_LONG == 32
-
-#define __mlog_test_u64(mask, bits)                    \
-       ( (u32)(mask & 0xffffffff) & bits.words[0] ||   \
-         ((u64)(mask) >> 32) & bits.words[1] )
-#define __mlog_set_u64(mask, bits) do {                        \
-       bits.words[0] |= (u32)(mask & 0xffffffff);      \
-               bits.words[1] |= (u64)(mask) >> 32;             \
-} while (0)
-#define __mlog_clear_u64(mask, bits) do {              \
-       bits.words[0] &= ~((u32)(mask & 0xffffffff));   \
-               bits.words[1] &= ~((u64)(mask) >> 32);          \
-} while (0)
-#define MLOG_BITS_RHS(mask) {                          \
-       {                                               \
-               [0] = (u32)(mask & 0xffffffff),         \
-               [1] = (u64)(mask) >> 32,                \
-       }                                               \
-}
-
-#else /* 32bit long above, 64bit long below */
-
-#define __mlog_test_u64(mask, bits)    ((mask) & bits.words[0])
-#define __mlog_set_u64(mask, bits) do {                \
-       bits.words[0] |= (mask);                \
-} while (0)
-#define __mlog_clear_u64(mask, bits) do {      \
-       bits.words[0] &= ~(mask);               \
-} while (0)
-#define MLOG_BITS_RHS(mask) { { (mask) } }
-
-#endif
-
-/*
- * smp_processor_id() "helpfully" screams when called outside preemptible
- * regions in current kernels.  sles doesn't have the variants that don't
- * scream.  just do this instead of trying to guess which we're building
- * against.. *sigh*.
- */
-#define __mlog_cpu_guess ({            \
-       unsigned long _cpu = get_cpu(); \
-       put_cpu();                      \
-       _cpu;                           \
-})
-
-/* In the following two macros, the whitespace after the ',' just
- * before ##args is intentional. Otherwise, gcc 2.95 will eat the
- * previous token if args expands to nothing.
- */
-#define __mlog_printk(level, fmt, args...)                             \
-       printk(level "(%s,%u,%lu):%s:%d " fmt, current->comm,           \
-              task_pid_nr(current), __mlog_cpu_guess,                  \
-              __PRETTY_FUNCTION__, __LINE__ , ##args)
-
-#define mlog(mask, fmt, args...) do {                                  \
-       u64 __m = MLOG_MASK_PREFIX | (mask);                            \
-       if ((__m & ML_ALLOWED_BITS) &&                                  \
-           __mlog_test_u64(__m, mlog_and_bits) &&                      \
-           !__mlog_test_u64(__m, mlog_not_bits)) {                     \
-               if (__m & ML_ERROR)                                     \
-                       __mlog_printk(KERN_ERR, "ERROR: "fmt , ##args); \
-               else if (__m & ML_NOTICE)                               \
-                       __mlog_printk(KERN_NOTICE, fmt , ##args);       \
-               else __mlog_printk(KERN_INFO, fmt , ##args);            \
-       }                                                               \
-} while (0)
-
-#define mlog_errno(st) do {                                            \
-       int _st = (st);                                                 \
-       if (_st != -ERESTARTSYS && _st != -EINTR &&                     \
-           _st != AOP_TRUNCATED_PAGE && _st != -ENOSPC)                \
-               mlog(ML_ERROR, "status = %lld\n", (long long)_st);      \
-} while (0)
-
-#define mlog_bug_on_msg(cond, fmt, args...) do {                       \
-       if (cond) {                                                     \
-               mlog(ML_ERROR, "bug expression: " #cond "\n");          \
-               mlog(ML_ERROR, fmt, ##args);                            \
-               BUG();                                                  \
-       }                                                               \
-} while (0)
-
-#include <linux/kobject.h>
-#include <linux/sysfs.h>
-int mlog_sys_init(struct kset *o2cb_subsys);
-void mlog_sys_shutdown(void);
-
-#endif /* O2CLUSTER_MASKLOG_H */
diff --git a/drivers/staging/ramster/cluster/netdebug.c b/drivers/staging/ramster/cluster/netdebug.c
deleted file mode 100644 (file)
index dc45deb..0000000
+++ /dev/null
@@ -1,579 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * netdebug.c
- *
- * debug functionality for o2net
- *
- * Copyright (C) 2005, 2008 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifdef CONFIG_DEBUG_FS
-
-#include <linux/module.h>
-#include <linux/types.h>
-#include <linux/slab.h>
-#include <linux/idr.h>
-#include <linux/kref.h>
-#include <linux/seq_file.h>
-#include <linux/debugfs.h>
-
-#include <linux/uaccess.h>
-
-#include "tcp.h"
-#include "nodemanager.h"
-#define MLOG_MASK_PREFIX ML_TCP
-#include "masklog.h"
-
-#include "tcp_internal.h"
-
-#define O2NET_DEBUG_DIR                "o2net"
-#define SC_DEBUG_NAME          "sock_containers"
-#define NST_DEBUG_NAME         "send_tracking"
-#define STATS_DEBUG_NAME       "stats"
-#define NODES_DEBUG_NAME       "connected_nodes"
-
-#define SHOW_SOCK_CONTAINERS   0
-#define SHOW_SOCK_STATS                1
-
-static struct dentry *o2net_dentry;
-static struct dentry *sc_dentry;
-static struct dentry *nst_dentry;
-static struct dentry *stats_dentry;
-static struct dentry *nodes_dentry;
-
-static DEFINE_SPINLOCK(o2net_debug_lock);
-
-static LIST_HEAD(sock_containers);
-static LIST_HEAD(send_tracking);
-
-void o2net_debug_add_nst(struct o2net_send_tracking *nst)
-{
-       spin_lock(&o2net_debug_lock);
-       list_add(&nst->st_net_debug_item, &send_tracking);
-       spin_unlock(&o2net_debug_lock);
-}
-
-void o2net_debug_del_nst(struct o2net_send_tracking *nst)
-{
-       spin_lock(&o2net_debug_lock);
-       if (!list_empty(&nst->st_net_debug_item))
-               list_del_init(&nst->st_net_debug_item);
-       spin_unlock(&o2net_debug_lock);
-}
-
-static struct o2net_send_tracking
-                       *next_nst(struct o2net_send_tracking *nst_start)
-{
-       struct o2net_send_tracking *nst, *ret = NULL;
-
-       assert_spin_locked(&o2net_debug_lock);
-
-       list_for_each_entry(nst, &nst_start->st_net_debug_item,
-                           st_net_debug_item) {
-               /* discover the head of the list */
-               if (&nst->st_net_debug_item == &send_tracking)
-                       break;
-
-               /* use st_task to detect real nsts in the list */
-               if (nst->st_task != NULL) {
-                       ret = nst;
-                       break;
-               }
-       }
-
-       return ret;
-}
-
-static void *nst_seq_start(struct seq_file *seq, loff_t *pos)
-{
-       struct o2net_send_tracking *nst, *dummy_nst = seq->private;
-
-       spin_lock(&o2net_debug_lock);
-       nst = next_nst(dummy_nst);
-       spin_unlock(&o2net_debug_lock);
-
-       return nst;
-}
-
-static void *nst_seq_next(struct seq_file *seq, void *v, loff_t *pos)
-{
-       struct o2net_send_tracking *nst, *dummy_nst = seq->private;
-
-       spin_lock(&o2net_debug_lock);
-       nst = next_nst(dummy_nst);
-       list_del_init(&dummy_nst->st_net_debug_item);
-       if (nst)
-               list_add(&dummy_nst->st_net_debug_item,
-                        &nst->st_net_debug_item);
-       spin_unlock(&o2net_debug_lock);
-
-       return nst; /* unused, just needs to be null when done */
-}
-
-static int nst_seq_show(struct seq_file *seq, void *v)
-{
-       struct o2net_send_tracking *nst, *dummy_nst = seq->private;
-       ktime_t now;
-       s64 sock, send, status;
-
-       spin_lock(&o2net_debug_lock);
-       nst = next_nst(dummy_nst);
-       if (!nst)
-               goto out;
-
-       now = ktime_get();
-       sock = ktime_to_us(ktime_sub(now, nst->st_sock_time));
-       send = ktime_to_us(ktime_sub(now, nst->st_send_time));
-       status = ktime_to_us(ktime_sub(now, nst->st_status_time));
-
-       /* get_task_comm isn't exported.  oh well. */
-       seq_printf(seq, "%p:\n"
-                  "  pid:          %lu\n"
-                  "  tgid:         %lu\n"
-                  "  process name: %s\n"
-                  "  node:         %u\n"
-                  "  sc:           %p\n"
-                  "  message id:   %d\n"
-                  "  message type: %u\n"
-                  "  message key:  0x%08x\n"
-                  "  sock acquiry: %lld usecs ago\n"
-                  "  send start:   %lld usecs ago\n"
-                  "  wait start:   %lld usecs ago\n",
-                  nst, (unsigned long)task_pid_nr(nst->st_task),
-                  (unsigned long)nst->st_task->tgid,
-                  nst->st_task->comm, nst->st_node,
-                  nst->st_sc, nst->st_id, nst->st_msg_type,
-                  nst->st_msg_key,
-                  (long long)sock,
-                  (long long)send,
-                  (long long)status);
-
-out:
-       spin_unlock(&o2net_debug_lock);
-
-       return 0;
-}
-
-static void nst_seq_stop(struct seq_file *seq, void *v)
-{
-}
-
-static const struct seq_operations nst_seq_ops = {
-       .start = nst_seq_start,
-       .next = nst_seq_next,
-       .stop = nst_seq_stop,
-       .show = nst_seq_show,
-};
-
-static int nst_fop_open(struct inode *inode, struct file *file)
-{
-       struct o2net_send_tracking *dummy_nst;
-       struct seq_file *seq;
-       int ret;
-
-       dummy_nst = kmalloc(sizeof(struct o2net_send_tracking), GFP_KERNEL);
-       if (dummy_nst == NULL) {
-               ret = -ENOMEM;
-               goto out;
-       }
-       dummy_nst->st_task = NULL;
-
-       ret = seq_open(file, &nst_seq_ops);
-       if (ret)
-               goto out;
-
-       seq = file->private_data;
-       seq->private = dummy_nst;
-       o2net_debug_add_nst(dummy_nst);
-
-       dummy_nst = NULL;
-
-out:
-       kfree(dummy_nst);
-       return ret;
-}
-
-static int nst_fop_release(struct inode *inode, struct file *file)
-{
-       struct seq_file *seq = file->private_data;
-       struct o2net_send_tracking *dummy_nst = seq->private;
-
-       o2net_debug_del_nst(dummy_nst);
-       return seq_release_private(inode, file);
-}
-
-static const struct file_operations nst_seq_fops = {
-       .open = nst_fop_open,
-       .read = seq_read,
-       .llseek = seq_lseek,
-       .release = nst_fop_release,
-};
-
-void o2net_debug_add_sc(struct o2net_sock_container *sc)
-{
-       spin_lock(&o2net_debug_lock);
-       list_add(&sc->sc_net_debug_item, &sock_containers);
-       spin_unlock(&o2net_debug_lock);
-}
-
-void o2net_debug_del_sc(struct o2net_sock_container *sc)
-{
-       spin_lock(&o2net_debug_lock);
-       list_del_init(&sc->sc_net_debug_item);
-       spin_unlock(&o2net_debug_lock);
-}
-
-struct o2net_sock_debug {
-       int dbg_ctxt;
-       struct o2net_sock_container *dbg_sock;
-};
-
-static struct o2net_sock_container
-                       *next_sc(struct o2net_sock_container *sc_start)
-{
-       struct o2net_sock_container *sc, *ret = NULL;
-
-       assert_spin_locked(&o2net_debug_lock);
-
-       list_for_each_entry(sc, &sc_start->sc_net_debug_item,
-                           sc_net_debug_item) {
-               /* discover the head of the list miscast as a sc */
-               if (&sc->sc_net_debug_item == &sock_containers)
-                       break;
-
-               /* use sc_page to detect real scs in the list */
-               if (sc->sc_page != NULL) {
-                       ret = sc;
-                       break;
-               }
-       }
-
-       return ret;
-}
-
-static void *sc_seq_start(struct seq_file *seq, loff_t *pos)
-{
-       struct o2net_sock_debug *sd = seq->private;
-       struct o2net_sock_container *sc, *dummy_sc = sd->dbg_sock;
-
-       spin_lock(&o2net_debug_lock);
-       sc = next_sc(dummy_sc);
-       spin_unlock(&o2net_debug_lock);
-
-       return sc;
-}
-
-static void *sc_seq_next(struct seq_file *seq, void *v, loff_t *pos)
-{
-       struct o2net_sock_debug *sd = seq->private;
-       struct o2net_sock_container *sc, *dummy_sc = sd->dbg_sock;
-
-       spin_lock(&o2net_debug_lock);
-       sc = next_sc(dummy_sc);
-       list_del_init(&dummy_sc->sc_net_debug_item);
-       if (sc)
-               list_add(&dummy_sc->sc_net_debug_item, &sc->sc_net_debug_item);
-       spin_unlock(&o2net_debug_lock);
-
-       return sc; /* unused, just needs to be null when done */
-}
-
-#ifdef CONFIG_OCFS2_FS_STATS
-# define sc_send_count(_s)             ((_s)->sc_send_count)
-# define sc_recv_count(_s)             ((_s)->sc_recv_count)
-# define sc_tv_acquiry_total_ns(_s)    (ktime_to_ns((_s)->sc_tv_acquiry_total))
-# define sc_tv_send_total_ns(_s)       (ktime_to_ns((_s)->sc_tv_send_total))
-# define sc_tv_status_total_ns(_s)     (ktime_to_ns((_s)->sc_tv_status_total))
-# define sc_tv_process_total_ns(_s)    (ktime_to_ns((_s)->sc_tv_process_total))
-#else
-# define sc_send_count(_s)             (0U)
-# define sc_recv_count(_s)             (0U)
-# define sc_tv_acquiry_total_ns(_s)    (0LL)
-# define sc_tv_send_total_ns(_s)       (0LL)
-# define sc_tv_status_total_ns(_s)     (0LL)
-# define sc_tv_process_total_ns(_s)    (0LL)
-#endif
-
-/* So that debugfs.ocfs2 can determine which format is being used */
-#define O2NET_STATS_STR_VERSION                1
-static void sc_show_sock_stats(struct seq_file *seq,
-                              struct o2net_sock_container *sc)
-{
-       if (!sc)
-               return;
-
-       seq_printf(seq, "%d,%u,%lu,%lld,%lld,%lld,%lu,%lld\n", O2NET_STATS_STR_VERSION,
-                  sc->sc_node->nd_num, (unsigned long)sc_send_count(sc),
-                  (long long)sc_tv_acquiry_total_ns(sc),
-                  (long long)sc_tv_send_total_ns(sc),
-                  (long long)sc_tv_status_total_ns(sc),
-                  (unsigned long)sc_recv_count(sc),
-                  (long long)sc_tv_process_total_ns(sc));
-}
-
-static void sc_show_sock_container(struct seq_file *seq,
-                                  struct o2net_sock_container *sc)
-{
-       struct inet_sock *inet = NULL;
-       __be32 saddr = 0, daddr = 0;
-       __be16 sport = 0, dport = 0;
-
-       if (!sc)
-               return;
-
-       if (sc->sc_sock) {
-               inet = inet_sk(sc->sc_sock->sk);
-               /* the stack's structs aren't sparse endian clean */
-               saddr = (__force __be32)inet->inet_saddr;
-               daddr = (__force __be32)inet->inet_daddr;
-               sport = (__force __be16)inet->inet_sport;
-               dport = (__force __be16)inet->inet_dport;
-       }
-
-       /* XXX sigh, inet-> doesn't have sparse annotation so any
-        * use of it here generates a warning with -Wbitwise */
-       seq_printf(seq, "%p:\n"
-                  "  krefs:           %d\n"
-                  "  sock:            %pI4:%u -> "
-                                     "%pI4:%u\n"
-                  "  remote node:     %s\n"
-                  "  page off:        %zu\n"
-                  "  handshake ok:    %u\n"
-                  "  timer:           %lld usecs\n"
-                  "  data ready:      %lld usecs\n"
-                  "  advance start:   %lld usecs\n"
-                  "  advance stop:    %lld usecs\n"
-                  "  func start:      %lld usecs\n"
-                  "  func stop:       %lld usecs\n"
-                  "  func key:        0x%08x\n"
-                  "  func type:       %u\n",
-                  sc,
-                  atomic_read(&sc->sc_kref.refcount),
-                  &saddr, inet ? ntohs(sport) : 0,
-                  &daddr, inet ? ntohs(dport) : 0,
-                  sc->sc_node->nd_name,
-                  sc->sc_page_off,
-                  sc->sc_handshake_ok,
-                  (long long)ktime_to_us(sc->sc_tv_timer),
-                  (long long)ktime_to_us(sc->sc_tv_data_ready),
-                  (long long)ktime_to_us(sc->sc_tv_advance_start),
-                  (long long)ktime_to_us(sc->sc_tv_advance_stop),
-                  (long long)ktime_to_us(sc->sc_tv_func_start),
-                  (long long)ktime_to_us(sc->sc_tv_func_stop),
-                  sc->sc_msg_key,
-                  sc->sc_msg_type);
-}
-
-static int sc_seq_show(struct seq_file *seq, void *v)
-{
-       struct o2net_sock_debug *sd = seq->private;
-       struct o2net_sock_container *sc, *dummy_sc = sd->dbg_sock;
-
-       spin_lock(&o2net_debug_lock);
-       sc = next_sc(dummy_sc);
-
-       if (sc) {
-               if (sd->dbg_ctxt == SHOW_SOCK_CONTAINERS)
-                       sc_show_sock_container(seq, sc);
-               else
-                       sc_show_sock_stats(seq, sc);
-       }
-
-       spin_unlock(&o2net_debug_lock);
-
-       return 0;
-}
-
-static void sc_seq_stop(struct seq_file *seq, void *v)
-{
-}
-
-static const struct seq_operations sc_seq_ops = {
-       .start = sc_seq_start,
-       .next = sc_seq_next,
-       .stop = sc_seq_stop,
-       .show = sc_seq_show,
-};
-
-static int sc_common_open(struct file *file, struct o2net_sock_debug *sd)
-{
-       struct o2net_sock_container *dummy_sc;
-       struct seq_file *seq;
-       int ret;
-
-       dummy_sc = kmalloc(sizeof(struct o2net_sock_container), GFP_KERNEL);
-       if (dummy_sc == NULL) {
-               ret = -ENOMEM;
-               goto out;
-       }
-       dummy_sc->sc_page = NULL;
-
-       ret = seq_open(file, &sc_seq_ops);
-       if (ret)
-               goto out;
-
-       seq = file->private_data;
-       seq->private = sd;
-       sd->dbg_sock = dummy_sc;
-       o2net_debug_add_sc(dummy_sc);
-
-       dummy_sc = NULL;
-
-out:
-       kfree(dummy_sc);
-       return ret;
-}
-
-static int sc_fop_release(struct inode *inode, struct file *file)
-{
-       struct seq_file *seq = file->private_data;
-       struct o2net_sock_debug *sd = seq->private;
-       struct o2net_sock_container *dummy_sc = sd->dbg_sock;
-
-       o2net_debug_del_sc(dummy_sc);
-       return seq_release_private(inode, file);
-}
-
-static int stats_fop_open(struct inode *inode, struct file *file)
-{
-       struct o2net_sock_debug *sd;
-
-       sd = kmalloc(sizeof(struct o2net_sock_debug), GFP_KERNEL);
-       if (sd == NULL)
-               return -ENOMEM;
-
-       sd->dbg_ctxt = SHOW_SOCK_STATS;
-       sd->dbg_sock = NULL;
-
-       return sc_common_open(file, sd);
-}
-
-static const struct file_operations stats_seq_fops = {
-       .open = stats_fop_open,
-       .read = seq_read,
-       .llseek = seq_lseek,
-       .release = sc_fop_release,
-};
-
-static int sc_fop_open(struct inode *inode, struct file *file)
-{
-       struct o2net_sock_debug *sd;
-
-       sd = kmalloc(sizeof(struct o2net_sock_debug), GFP_KERNEL);
-       if (sd == NULL)
-               return -ENOMEM;
-
-       sd->dbg_ctxt = SHOW_SOCK_CONTAINERS;
-       sd->dbg_sock = NULL;
-
-       return sc_common_open(file, sd);
-}
-
-static const struct file_operations sc_seq_fops = {
-       .open = sc_fop_open,
-       .read = seq_read,
-       .llseek = seq_lseek,
-       .release = sc_fop_release,
-};
-
-static int o2net_fill_bitmap(char *buf, int len)
-{
-       unsigned long map[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       int i = -1, out = 0;
-
-       o2net_fill_node_map(map, sizeof(map));
-
-       while ((i = find_next_bit(map, O2NM_MAX_NODES, i + 1)) < O2NM_MAX_NODES)
-               out += snprintf(buf + out, PAGE_SIZE - out, "%d ", i);
-       out += snprintf(buf + out, PAGE_SIZE - out, "\n");
-
-       return out;
-}
-
-static int nodes_fop_open(struct inode *inode, struct file *file)
-{
-       char *buf;
-
-       buf = kmalloc(PAGE_SIZE, GFP_KERNEL);
-       if (!buf)
-               return -ENOMEM;
-
-       i_size_write(inode, o2net_fill_bitmap(buf, PAGE_SIZE));
-
-       file->private_data = buf;
-
-       return 0;
-}
-
-static int o2net_debug_release(struct inode *inode, struct file *file)
-{
-       kfree(file->private_data);
-       return 0;
-}
-
-static ssize_t o2net_debug_read(struct file *file, char __user *buf,
-                               size_t nbytes, loff_t *ppos)
-{
-       return simple_read_from_buffer(buf, nbytes, ppos, file->private_data,
-                                      i_size_read(file->f_mapping->host));
-}
-
-static const struct file_operations nodes_fops = {
-       .open           = nodes_fop_open,
-       .release        = o2net_debug_release,
-       .read           = o2net_debug_read,
-       .llseek         = generic_file_llseek,
-};
-
-void o2net_debugfs_exit(void)
-{
-       debugfs_remove(nodes_dentry);
-       debugfs_remove(stats_dentry);
-       debugfs_remove(sc_dentry);
-       debugfs_remove(nst_dentry);
-       debugfs_remove(o2net_dentry);
-}
-
-int o2net_debugfs_init(void)
-{
-       mode_t mode = S_IFREG|S_IRUSR;
-
-       o2net_dentry = debugfs_create_dir(O2NET_DEBUG_DIR, NULL);
-       if (o2net_dentry)
-               nst_dentry = debugfs_create_file(NST_DEBUG_NAME, mode,
-                                       o2net_dentry, NULL, &nst_seq_fops);
-       if (nst_dentry)
-               sc_dentry = debugfs_create_file(SC_DEBUG_NAME, mode,
-                                       o2net_dentry, NULL, &sc_seq_fops);
-       if (sc_dentry)
-               stats_dentry = debugfs_create_file(STATS_DEBUG_NAME, mode,
-                                       o2net_dentry, NULL, &stats_seq_fops);
-       if (stats_dentry)
-               nodes_dentry = debugfs_create_file(NODES_DEBUG_NAME, mode,
-                                       o2net_dentry, NULL, &nodes_fops);
-       if (nodes_dentry)
-               return 0;
-
-       o2net_debugfs_exit();
-       mlog_errno(-ENOMEM);
-       return -ENOMEM;
-}
-
-#endif /* CONFIG_DEBUG_FS */
diff --git a/drivers/staging/ramster/cluster/nodemanager.c b/drivers/staging/ramster/cluster/nodemanager.c
deleted file mode 100644 (file)
index bb24064..0000000
+++ /dev/null
@@ -1,989 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2004, 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#include <linux/slab.h>
-#include <linux/kernel.h>
-#include <linux/module.h>
-#include <linux/configfs.h>
-
-#include "tcp.h"
-#include "nodemanager.h"
-#include "heartbeat.h"
-#include "masklog.h"
-#include "sys.h"
-#include "ver.h"
-
-/* for now we operate under the assertion that there can be only one
- * cluster active at a time.  Changing this will require trickling
- * cluster references throughout where nodes are looked up */
-struct o2nm_cluster *o2nm_single_cluster = NULL;
-
-char *o2nm_fence_method_desc[O2NM_FENCE_METHODS] = {
-               "reset",        /* O2NM_FENCE_RESET */
-               "panic",        /* O2NM_FENCE_PANIC */
-};
-
-struct o2nm_node *o2nm_get_node_by_num(u8 node_num)
-{
-       struct o2nm_node *node = NULL;
-
-       if (node_num >= O2NM_MAX_NODES || o2nm_single_cluster == NULL)
-               goto out;
-
-       read_lock(&o2nm_single_cluster->cl_nodes_lock);
-       node = o2nm_single_cluster->cl_nodes[node_num];
-       if (node)
-               config_item_get(&node->nd_item);
-       read_unlock(&o2nm_single_cluster->cl_nodes_lock);
-out:
-       return node;
-}
-EXPORT_SYMBOL_GPL(o2nm_get_node_by_num);
-
-int o2nm_configured_node_map(unsigned long *map, unsigned bytes)
-{
-       struct o2nm_cluster *cluster = o2nm_single_cluster;
-
-       BUG_ON(bytes < (sizeof(cluster->cl_nodes_bitmap)));
-
-       if (cluster == NULL)
-               return -EINVAL;
-
-       read_lock(&cluster->cl_nodes_lock);
-       memcpy(map, cluster->cl_nodes_bitmap, sizeof(cluster->cl_nodes_bitmap));
-       read_unlock(&cluster->cl_nodes_lock);
-
-       return 0;
-}
-EXPORT_SYMBOL_GPL(o2nm_configured_node_map);
-
-static struct o2nm_node *o2nm_node_ip_tree_lookup(struct o2nm_cluster *cluster,
-                                                 __be32 ip_needle,
-                                                 struct rb_node ***ret_p,
-                                                 struct rb_node **ret_parent)
-{
-       struct rb_node **p = &cluster->cl_node_ip_tree.rb_node;
-       struct rb_node *parent = NULL;
-       struct o2nm_node *node, *ret = NULL;
-
-       while (*p) {
-               int cmp;
-
-               parent = *p;
-               node = rb_entry(parent, struct o2nm_node, nd_ip_node);
-
-               cmp = memcmp(&ip_needle, &node->nd_ipv4_address,
-                               sizeof(ip_needle));
-               if (cmp < 0)
-                       p = &(*p)->rb_left;
-               else if (cmp > 0)
-                       p = &(*p)->rb_right;
-               else {
-                       ret = node;
-                       break;
-               }
-       }
-
-       if (ret_p != NULL)
-               *ret_p = p;
-       if (ret_parent != NULL)
-               *ret_parent = parent;
-
-       return ret;
-}
-
-struct o2nm_node *o2nm_get_node_by_ip(__be32 addr)
-{
-       struct o2nm_node *node = NULL;
-       struct o2nm_cluster *cluster = o2nm_single_cluster;
-
-       if (cluster == NULL)
-               goto out;
-
-       read_lock(&cluster->cl_nodes_lock);
-       node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL);
-       if (node)
-               config_item_get(&node->nd_item);
-       read_unlock(&cluster->cl_nodes_lock);
-
-out:
-       return node;
-}
-EXPORT_SYMBOL_GPL(o2nm_get_node_by_ip);
-
-void o2nm_node_put(struct o2nm_node *node)
-{
-       config_item_put(&node->nd_item);
-}
-EXPORT_SYMBOL_GPL(o2nm_node_put);
-
-void o2nm_node_get(struct o2nm_node *node)
-{
-       config_item_get(&node->nd_item);
-}
-EXPORT_SYMBOL_GPL(o2nm_node_get);
-
-u8 o2nm_this_node(void)
-{
-       u8 node_num = O2NM_MAX_NODES;
-
-       if (o2nm_single_cluster && o2nm_single_cluster->cl_has_local)
-               node_num = o2nm_single_cluster->cl_local_node;
-
-       return node_num;
-}
-EXPORT_SYMBOL_GPL(o2nm_this_node);
-
-/* node configfs bits */
-
-static struct o2nm_cluster *to_o2nm_cluster(struct config_item *item)
-{
-       return item ?
-               container_of(to_config_group(item), struct o2nm_cluster,
-                            cl_group)
-               : NULL;
-}
-
-static struct o2nm_node *to_o2nm_node(struct config_item *item)
-{
-       return item ? container_of(item, struct o2nm_node, nd_item) : NULL;
-}
-
-static void o2nm_node_release(struct config_item *item)
-{
-       struct o2nm_node *node = to_o2nm_node(item);
-       kfree(node);
-}
-
-static ssize_t o2nm_node_num_read(struct o2nm_node *node, char *page)
-{
-       return sprintf(page, "%d\n", node->nd_num);
-}
-
-static struct o2nm_cluster *to_o2nm_cluster_from_node(struct o2nm_node *node)
-{
-       /* through the first node_set .parent
-        * mycluster/nodes/mynode == o2nm_cluster->o2nm_node_group->o2nm_node */
-       return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent);
-}
-
-enum {
-       O2NM_NODE_ATTR_NUM = 0,
-       O2NM_NODE_ATTR_PORT,
-       O2NM_NODE_ATTR_ADDRESS,
-       O2NM_NODE_ATTR_LOCAL,
-};
-
-static ssize_t o2nm_node_num_write(struct o2nm_node *node, const char *page,
-                                  size_t count)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
-       unsigned long tmp;
-       char *p = (char *)page;
-
-       tmp = simple_strtoul(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       if (tmp >= O2NM_MAX_NODES)
-               return -ERANGE;
-
-       /* once we're in the cl_nodes tree networking can look us up by
-        * node number and try to use our address and port attributes
-        * to connect to this node.. make sure that they've been set
-        * before writing the node attribute? */
-       if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
-           !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
-               return -EINVAL; /* XXX */
-
-       write_lock(&cluster->cl_nodes_lock);
-       if (cluster->cl_nodes[tmp])
-               p = NULL;
-       else  {
-               cluster->cl_nodes[tmp] = node;
-               node->nd_num = tmp;
-               set_bit(tmp, cluster->cl_nodes_bitmap);
-       }
-       write_unlock(&cluster->cl_nodes_lock);
-       if (p == NULL)
-               return -EEXIST;
-
-       return count;
-}
-static ssize_t o2nm_node_ipv4_port_read(struct o2nm_node *node, char *page)
-{
-       return sprintf(page, "%u\n", ntohs(node->nd_ipv4_port));
-}
-
-static ssize_t o2nm_node_ipv4_port_write(struct o2nm_node *node,
-                                        const char *page, size_t count)
-{
-       unsigned long tmp;
-       char *p = (char *)page;
-
-       tmp = simple_strtoul(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       if (tmp == 0)
-               return -EINVAL;
-       if (tmp >= (u16)-1)
-               return -ERANGE;
-
-       node->nd_ipv4_port = htons(tmp);
-
-       return count;
-}
-
-static ssize_t o2nm_node_ipv4_address_read(struct o2nm_node *node, char *page)
-{
-       return sprintf(page, "%pI4\n", &node->nd_ipv4_address);
-}
-
-static ssize_t o2nm_node_ipv4_address_write(struct o2nm_node *node,
-                                           const char *page,
-                                           size_t count)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
-       int ret, i;
-       struct rb_node **p, *parent;
-       unsigned int octets[4];
-       __be32 ipv4_addr = 0;
-
-       ret = sscanf(page, "%3u.%3u.%3u.%3u", &octets[3], &octets[2],
-                    &octets[1], &octets[0]);
-       if (ret != 4)
-               return -EINVAL;
-
-       for (i = 0; i < ARRAY_SIZE(octets); i++) {
-               if (octets[i] > 255)
-                       return -ERANGE;
-               be32_add_cpu(&ipv4_addr, octets[i] << (i * 8));
-       }
-
-       ret = 0;
-       write_lock(&cluster->cl_nodes_lock);
-       if (o2nm_node_ip_tree_lookup(cluster, ipv4_addr, &p, &parent))
-               ret = -EEXIST;
-       else {
-               rb_link_node(&node->nd_ip_node, parent, p);
-               rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree);
-       }
-       write_unlock(&cluster->cl_nodes_lock);
-       if (ret)
-               return ret;
-
-       memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr));
-
-       return count;
-}
-
-static ssize_t o2nm_node_local_read(struct o2nm_node *node, char *page)
-{
-       return sprintf(page, "%d\n", node->nd_local);
-}
-
-static ssize_t o2nm_node_local_write(struct o2nm_node *node, const char *page,
-                                    size_t count)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
-       unsigned long tmp;
-       char *p = (char *)page;
-       ssize_t ret;
-
-       tmp = simple_strtoul(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       tmp = !!tmp; /* boolean of whether this node wants to be local */
-
-       /* setting local turns on networking rx for now so we require having
-        * set everything else first */
-       if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
-           !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) ||
-           !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
-               return -EINVAL; /* XXX */
-
-       /* the only failure case is trying to set a new local node
-        * when a different one is already set */
-       if (tmp && tmp == cluster->cl_has_local &&
-           cluster->cl_local_node != node->nd_num)
-               return -EBUSY;
-
-       /* bring up the rx thread if we're setting the new local node. */
-       if (tmp && !cluster->cl_has_local) {
-               ret = o2net_start_listening(node);
-               if (ret)
-                       return ret;
-       }
-
-       if (!tmp && cluster->cl_has_local &&
-           cluster->cl_local_node == node->nd_num) {
-               o2net_stop_listening(node);
-               cluster->cl_local_node = O2NM_INVALID_NODE_NUM;
-       }
-
-       node->nd_local = tmp;
-       if (node->nd_local) {
-               cluster->cl_has_local = tmp;
-               cluster->cl_local_node = node->nd_num;
-       }
-
-       return count;
-}
-
-struct o2nm_node_attribute {
-       struct configfs_attribute attr;
-       ssize_t (*show)(struct o2nm_node *, char *);
-       ssize_t (*store)(struct o2nm_node *, const char *, size_t);
-};
-
-static struct o2nm_node_attribute o2nm_node_attr_num = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "num",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_node_num_read,
-       .store  = o2nm_node_num_write,
-};
-
-static struct o2nm_node_attribute o2nm_node_attr_ipv4_port = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "ipv4_port",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_node_ipv4_port_read,
-       .store  = o2nm_node_ipv4_port_write,
-};
-
-static struct o2nm_node_attribute o2nm_node_attr_ipv4_address = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "ipv4_address",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_node_ipv4_address_read,
-       .store  = o2nm_node_ipv4_address_write,
-};
-
-static struct o2nm_node_attribute o2nm_node_attr_local = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "local",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_node_local_read,
-       .store  = o2nm_node_local_write,
-};
-
-static struct configfs_attribute *o2nm_node_attrs[] = {
-       [O2NM_NODE_ATTR_NUM] = &o2nm_node_attr_num.attr,
-       [O2NM_NODE_ATTR_PORT] = &o2nm_node_attr_ipv4_port.attr,
-       [O2NM_NODE_ATTR_ADDRESS] = &o2nm_node_attr_ipv4_address.attr,
-       [O2NM_NODE_ATTR_LOCAL] = &o2nm_node_attr_local.attr,
-       NULL,
-};
-
-static int o2nm_attr_index(struct configfs_attribute *attr)
-{
-       int i;
-       for (i = 0; i < ARRAY_SIZE(o2nm_node_attrs); i++) {
-               if (attr == o2nm_node_attrs[i])
-                       return i;
-       }
-       BUG();
-       return 0;
-}
-
-static ssize_t o2nm_node_show(struct config_item *item,
-                             struct configfs_attribute *attr,
-                             char *page)
-{
-       struct o2nm_node *node = to_o2nm_node(item);
-       struct o2nm_node_attribute *o2nm_node_attr =
-               container_of(attr, struct o2nm_node_attribute, attr);
-       ssize_t ret = 0;
-
-       if (o2nm_node_attr->show)
-               ret = o2nm_node_attr->show(node, page);
-       return ret;
-}
-
-static ssize_t o2nm_node_store(struct config_item *item,
-                              struct configfs_attribute *attr,
-                              const char *page, size_t count)
-{
-       struct o2nm_node *node = to_o2nm_node(item);
-       struct o2nm_node_attribute *o2nm_node_attr =
-               container_of(attr, struct o2nm_node_attribute, attr);
-       ssize_t ret;
-       int attr_index = o2nm_attr_index(attr);
-
-       if (o2nm_node_attr->store == NULL) {
-               ret = -EINVAL;
-               goto out;
-       }
-
-       if (test_bit(attr_index, &node->nd_set_attributes))
-               return -EBUSY;
-
-       ret = o2nm_node_attr->store(node, page, count);
-       if (ret < count)
-               goto out;
-
-       set_bit(attr_index, &node->nd_set_attributes);
-out:
-       return ret;
-}
-
-static struct configfs_item_operations o2nm_node_item_ops = {
-       .release                = o2nm_node_release,
-       .show_attribute         = o2nm_node_show,
-       .store_attribute        = o2nm_node_store,
-};
-
-static struct config_item_type o2nm_node_type = {
-       .ct_item_ops    = &o2nm_node_item_ops,
-       .ct_attrs       = o2nm_node_attrs,
-       .ct_owner       = THIS_MODULE,
-};
-
-/* node set */
-
-struct o2nm_node_group {
-       struct config_group ns_group;
-       /* some stuff? */
-};
-
-#if 0
-static struct o2nm_node_group *to_o2nm_node_group(struct config_group *group)
-{
-       return group ?
-               container_of(group, struct o2nm_node_group, ns_group)
-               : NULL;
-}
-#endif
-
-struct o2nm_cluster_attribute {
-       struct configfs_attribute attr;
-       ssize_t (*show)(struct o2nm_cluster *, char *);
-       ssize_t (*store)(struct o2nm_cluster *, const char *, size_t);
-};
-
-static ssize_t o2nm_cluster_attr_write(const char *page, ssize_t count,
-                                       unsigned int *val)
-{
-       unsigned long tmp;
-       char *p = (char *)page;
-
-       tmp = simple_strtoul(p, &p, 0);
-       if (!p || (*p && (*p != '\n')))
-               return -EINVAL;
-
-       if (tmp == 0)
-               return -EINVAL;
-       if (tmp >= (u32)-1)
-               return -ERANGE;
-
-       *val = tmp;
-
-       return count;
-}
-
-static ssize_t o2nm_cluster_attr_idle_timeout_ms_read(
-       struct o2nm_cluster *cluster, char *page)
-{
-       return sprintf(page, "%u\n", cluster->cl_idle_timeout_ms);
-}
-
-static ssize_t o2nm_cluster_attr_idle_timeout_ms_write(
-       struct o2nm_cluster *cluster, const char *page, size_t count)
-{
-       ssize_t ret;
-       unsigned int val;
-
-       ret =  o2nm_cluster_attr_write(page, count, &val);
-
-       if (ret > 0) {
-               if (cluster->cl_idle_timeout_ms != val
-                       && o2net_num_connected_peers()) {
-                       mlog(ML_NOTICE,
-                            "o2net: cannot change idle timeout after "
-                            "the first peer has agreed to it."
-                            "  %d connected peers\n",
-                            o2net_num_connected_peers());
-                       ret = -EINVAL;
-               } else if (val <= cluster->cl_keepalive_delay_ms) {
-                       mlog(ML_NOTICE, "o2net: idle timeout must be larger "
-                            "than keepalive delay\n");
-                       ret = -EINVAL;
-               } else {
-                       cluster->cl_idle_timeout_ms = val;
-               }
-       }
-
-       return ret;
-}
-
-static ssize_t o2nm_cluster_attr_keepalive_delay_ms_read(
-       struct o2nm_cluster *cluster, char *page)
-{
-       return sprintf(page, "%u\n", cluster->cl_keepalive_delay_ms);
-}
-
-static ssize_t o2nm_cluster_attr_keepalive_delay_ms_write(
-       struct o2nm_cluster *cluster, const char *page, size_t count)
-{
-       ssize_t ret;
-       unsigned int val;
-
-       ret =  o2nm_cluster_attr_write(page, count, &val);
-
-       if (ret > 0) {
-               if (cluster->cl_keepalive_delay_ms != val
-                   && o2net_num_connected_peers()) {
-                       mlog(ML_NOTICE,
-                            "o2net: cannot change keepalive delay after"
-                            " the first peer has agreed to it."
-                            "  %d connected peers\n",
-                            o2net_num_connected_peers());
-                       ret = -EINVAL;
-               } else if (val >= cluster->cl_idle_timeout_ms) {
-                       mlog(ML_NOTICE, "o2net: keepalive delay must be "
-                            "smaller than idle timeout\n");
-                       ret = -EINVAL;
-               } else {
-                       cluster->cl_keepalive_delay_ms = val;
-               }
-       }
-
-       return ret;
-}
-
-static ssize_t o2nm_cluster_attr_reconnect_delay_ms_read(
-       struct o2nm_cluster *cluster, char *page)
-{
-       return sprintf(page, "%u\n", cluster->cl_reconnect_delay_ms);
-}
-
-static ssize_t o2nm_cluster_attr_reconnect_delay_ms_write(
-       struct o2nm_cluster *cluster, const char *page, size_t count)
-{
-       return o2nm_cluster_attr_write(page, count,
-                                      &cluster->cl_reconnect_delay_ms);
-}
-
-static ssize_t o2nm_cluster_attr_fence_method_read(
-       struct o2nm_cluster *cluster, char *page)
-{
-       ssize_t ret = 0;
-
-       if (cluster)
-               ret = sprintf(page, "%s\n",
-                             o2nm_fence_method_desc[cluster->cl_fence_method]);
-       return ret;
-}
-
-static ssize_t o2nm_cluster_attr_fence_method_write(
-       struct o2nm_cluster *cluster, const char *page, size_t count)
-{
-       unsigned int i;
-
-       if (page[count - 1] != '\n')
-               goto bail;
-
-       for (i = 0; i < O2NM_FENCE_METHODS; ++i) {
-               if (count != strlen(o2nm_fence_method_desc[i]) + 1)
-                       continue;
-               if (strncasecmp(page, o2nm_fence_method_desc[i], count - 1))
-                       continue;
-               if (cluster->cl_fence_method != i) {
-                       printk(KERN_INFO "ocfs2: Changing fence method to %s\n",
-                              o2nm_fence_method_desc[i]);
-                       cluster->cl_fence_method = i;
-               }
-               return count;
-       }
-
-bail:
-       return -EINVAL;
-}
-
-static struct o2nm_cluster_attribute o2nm_cluster_attr_idle_timeout_ms = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "idle_timeout_ms",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_cluster_attr_idle_timeout_ms_read,
-       .store  = o2nm_cluster_attr_idle_timeout_ms_write,
-};
-
-static struct o2nm_cluster_attribute o2nm_cluster_attr_keepalive_delay_ms = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "keepalive_delay_ms",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_cluster_attr_keepalive_delay_ms_read,
-       .store  = o2nm_cluster_attr_keepalive_delay_ms_write,
-};
-
-static struct o2nm_cluster_attribute o2nm_cluster_attr_reconnect_delay_ms = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "reconnect_delay_ms",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_cluster_attr_reconnect_delay_ms_read,
-       .store  = o2nm_cluster_attr_reconnect_delay_ms_write,
-};
-
-static struct o2nm_cluster_attribute o2nm_cluster_attr_fence_method = {
-       .attr   = { .ca_owner = THIS_MODULE,
-                   .ca_name = "fence_method",
-                   .ca_mode = S_IRUGO | S_IWUSR },
-       .show   = o2nm_cluster_attr_fence_method_read,
-       .store  = o2nm_cluster_attr_fence_method_write,
-};
-
-static struct configfs_attribute *o2nm_cluster_attrs[] = {
-       &o2nm_cluster_attr_idle_timeout_ms.attr,
-       &o2nm_cluster_attr_keepalive_delay_ms.attr,
-       &o2nm_cluster_attr_reconnect_delay_ms.attr,
-       &o2nm_cluster_attr_fence_method.attr,
-       NULL,
-};
-static ssize_t o2nm_cluster_show(struct config_item *item,
-                                 struct configfs_attribute *attr,
-                                 char *page)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster(item);
-       struct o2nm_cluster_attribute *o2nm_cluster_attr =
-               container_of(attr, struct o2nm_cluster_attribute, attr);
-       ssize_t ret = 0;
-
-       if (o2nm_cluster_attr->show)
-               ret = o2nm_cluster_attr->show(cluster, page);
-       return ret;
-}
-
-static ssize_t o2nm_cluster_store(struct config_item *item,
-                                  struct configfs_attribute *attr,
-                                  const char *page, size_t count)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster(item);
-       struct o2nm_cluster_attribute *o2nm_cluster_attr =
-               container_of(attr, struct o2nm_cluster_attribute, attr);
-       ssize_t ret;
-
-       if (o2nm_cluster_attr->store == NULL) {
-               ret = -EINVAL;
-               goto out;
-       }
-
-       ret = o2nm_cluster_attr->store(cluster, page, count);
-       if (ret < count)
-               goto out;
-out:
-       return ret;
-}
-
-static struct config_item *o2nm_node_group_make_item(struct config_group *group,
-                                                    const char *name)
-{
-       struct o2nm_node *node = NULL;
-
-       if (strlen(name) > O2NM_MAX_NAME_LEN)
-               return ERR_PTR(-ENAMETOOLONG);
-
-       node = kzalloc(sizeof(struct o2nm_node), GFP_KERNEL);
-       if (node == NULL)
-               return ERR_PTR(-ENOMEM);
-
-       strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */
-       config_item_init_type_name(&node->nd_item, name, &o2nm_node_type);
-       spin_lock_init(&node->nd_lock);
-
-       mlog(ML_CLUSTER, "o2nm: Registering node %s\n", name);
-
-       return &node->nd_item;
-}
-
-static void o2nm_node_group_drop_item(struct config_group *group,
-                                     struct config_item *item)
-{
-       struct o2nm_node *node = to_o2nm_node(item);
-       struct o2nm_cluster *cluster = to_o2nm_cluster(group->cg_item.ci_parent);
-
-       o2net_disconnect_node(node);
-
-       if (cluster->cl_has_local &&
-           (cluster->cl_local_node == node->nd_num)) {
-               cluster->cl_has_local = 0;
-               cluster->cl_local_node = O2NM_INVALID_NODE_NUM;
-               o2net_stop_listening(node);
-       }
-
-       /* XXX call into net to stop this node from trading messages */
-
-       write_lock(&cluster->cl_nodes_lock);
-
-       /* XXX sloppy */
-       if (node->nd_ipv4_address)
-               rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree);
-
-       /* nd_num might be 0 if the node number hasn't been set.. */
-       if (cluster->cl_nodes[node->nd_num] == node) {
-               cluster->cl_nodes[node->nd_num] = NULL;
-               clear_bit(node->nd_num, cluster->cl_nodes_bitmap);
-       }
-       write_unlock(&cluster->cl_nodes_lock);
-
-       mlog(ML_CLUSTER, "o2nm: Unregistered node %s\n",
-            config_item_name(&node->nd_item));
-
-       config_item_put(item);
-}
-
-static struct configfs_group_operations o2nm_node_group_group_ops = {
-       .make_item      = o2nm_node_group_make_item,
-       .drop_item      = o2nm_node_group_drop_item,
-};
-
-static struct config_item_type o2nm_node_group_type = {
-       .ct_group_ops   = &o2nm_node_group_group_ops,
-       .ct_owner       = THIS_MODULE,
-};
-
-/* cluster */
-
-static void o2nm_cluster_release(struct config_item *item)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster(item);
-
-       kfree(cluster->cl_group.default_groups);
-       kfree(cluster);
-}
-
-static struct configfs_item_operations o2nm_cluster_item_ops = {
-       .release        = o2nm_cluster_release,
-       .show_attribute         = o2nm_cluster_show,
-       .store_attribute        = o2nm_cluster_store,
-};
-
-static struct config_item_type o2nm_cluster_type = {
-       .ct_item_ops    = &o2nm_cluster_item_ops,
-       .ct_attrs       = o2nm_cluster_attrs,
-       .ct_owner       = THIS_MODULE,
-};
-
-/* cluster set */
-
-struct o2nm_cluster_group {
-       struct configfs_subsystem cs_subsys;
-       /* some stuff? */
-};
-
-#if 0
-static struct o2nm_cluster_group *to_o2nm_cluster_group(struct config_group *group)
-{
-       return group ?
-               container_of(to_configfs_subsystem(group), struct o2nm_cluster_group, cs_subsys)
-              : NULL;
-}
-#endif
-
-static struct config_group *o2nm_cluster_group_make_group(struct config_group *group,
-                                                         const char *name)
-{
-       struct o2nm_cluster *cluster = NULL;
-       struct o2nm_node_group *ns = NULL;
-       struct config_group *o2hb_group = NULL, *ret = NULL;
-       void *defs = NULL;
-
-       /* this runs under the parent dir's i_mutex; there can be only
-        * one caller in here at a time */
-       if (o2nm_single_cluster)
-               return ERR_PTR(-ENOSPC);
-
-       cluster = kzalloc(sizeof(struct o2nm_cluster), GFP_KERNEL);
-       ns = kzalloc(sizeof(struct o2nm_node_group), GFP_KERNEL);
-       defs = kcalloc(3, sizeof(struct config_group *), GFP_KERNEL);
-       o2hb_group = o2hb_alloc_hb_set();
-       if (cluster == NULL || ns == NULL || o2hb_group == NULL || defs == NULL)
-               goto out;
-
-       config_group_init_type_name(&cluster->cl_group, name,
-                                   &o2nm_cluster_type);
-       config_group_init_type_name(&ns->ns_group, "node",
-                                   &o2nm_node_group_type);
-
-       cluster->cl_group.default_groups = defs;
-       cluster->cl_group.default_groups[0] = &ns->ns_group;
-       cluster->cl_group.default_groups[1] = o2hb_group;
-       cluster->cl_group.default_groups[2] = NULL;
-       rwlock_init(&cluster->cl_nodes_lock);
-       cluster->cl_node_ip_tree = RB_ROOT;
-       cluster->cl_reconnect_delay_ms = O2NET_RECONNECT_DELAY_MS_DEFAULT;
-       cluster->cl_idle_timeout_ms    = O2NET_IDLE_TIMEOUT_MS_DEFAULT;
-       cluster->cl_keepalive_delay_ms = O2NET_KEEPALIVE_DELAY_MS_DEFAULT;
-       cluster->cl_fence_method       = O2NM_FENCE_RESET;
-
-       ret = &cluster->cl_group;
-       o2nm_single_cluster = cluster;
-
-out:
-       if (ret == NULL) {
-               kfree(cluster);
-               kfree(ns);
-               o2hb_free_hb_set(o2hb_group);
-               kfree(defs);
-               ret = ERR_PTR(-ENOMEM);
-       }
-
-       return ret;
-}
-
-static void o2nm_cluster_group_drop_item(struct config_group *group, struct config_item *item)
-{
-       struct o2nm_cluster *cluster = to_o2nm_cluster(item);
-       int i;
-       struct config_item *killme;
-
-       BUG_ON(o2nm_single_cluster != cluster);
-       o2nm_single_cluster = NULL;
-
-       for (i = 0; cluster->cl_group.default_groups[i]; i++) {
-               killme = &cluster->cl_group.default_groups[i]->cg_item;
-               cluster->cl_group.default_groups[i] = NULL;
-               config_item_put(killme);
-       }
-
-       config_item_put(item);
-}
-
-static struct configfs_group_operations o2nm_cluster_group_group_ops = {
-       .make_group     = o2nm_cluster_group_make_group,
-       .drop_item      = o2nm_cluster_group_drop_item,
-};
-
-static struct config_item_type o2nm_cluster_group_type = {
-       .ct_group_ops   = &o2nm_cluster_group_group_ops,
-       .ct_owner       = THIS_MODULE,
-};
-
-static struct o2nm_cluster_group o2nm_cluster_group = {
-       .cs_subsys = {
-               .su_group = {
-                       .cg_item = {
-                               .ci_namebuf = "cluster",
-                               .ci_type = &o2nm_cluster_group_type,
-                       },
-               },
-       },
-};
-
-int o2nm_depend_item(struct config_item *item)
-{
-       return configfs_depend_item(&o2nm_cluster_group.cs_subsys, item);
-}
-
-void o2nm_undepend_item(struct config_item *item)
-{
-       configfs_undepend_item(&o2nm_cluster_group.cs_subsys, item);
-}
-
-int o2nm_depend_this_node(void)
-{
-       int ret = 0;
-       struct o2nm_node *local_node;
-
-       local_node = o2nm_get_node_by_num(o2nm_this_node());
-       if (!local_node) {
-               ret = -EINVAL;
-               goto out;
-       }
-
-       ret = o2nm_depend_item(&local_node->nd_item);
-       o2nm_node_put(local_node);
-
-out:
-       return ret;
-}
-
-void o2nm_undepend_this_node(void)
-{
-       struct o2nm_node *local_node;
-
-       local_node = o2nm_get_node_by_num(o2nm_this_node());
-       BUG_ON(!local_node);
-
-       o2nm_undepend_item(&local_node->nd_item);
-       o2nm_node_put(local_node);
-}
-
-
-static void __exit exit_o2nm(void)
-{
-       /* XXX sync with hb callbacks and shut down hb? */
-       o2net_unregister_hb_callbacks();
-       configfs_unregister_subsystem(&o2nm_cluster_group.cs_subsys);
-       o2cb_sys_shutdown();
-
-       o2net_exit();
-       o2hb_exit();
-}
-
-static int __init init_o2nm(void)
-{
-       int ret = -1;
-
-       cluster_print_version();
-
-       ret = o2hb_init();
-       if (ret)
-               goto out;
-
-       ret = o2net_init();
-       if (ret)
-               goto out_o2hb;
-
-       ret = o2net_register_hb_callbacks();
-       if (ret)
-               goto out_o2net;
-
-       config_group_init(&o2nm_cluster_group.cs_subsys.su_group);
-       mutex_init(&o2nm_cluster_group.cs_subsys.su_mutex);
-       ret = configfs_register_subsystem(&o2nm_cluster_group.cs_subsys);
-       if (ret) {
-               printk(KERN_ERR "nodemanager: Registration returned %d\n", ret);
-               goto out_callbacks;
-       }
-
-       ret = o2cb_sys_init();
-       if (!ret)
-               goto out;
-
-       configfs_unregister_subsystem(&o2nm_cluster_group.cs_subsys);
-out_callbacks:
-       o2net_unregister_hb_callbacks();
-out_o2net:
-       o2net_exit();
-out_o2hb:
-       o2hb_exit();
-out:
-       return ret;
-}
-
-MODULE_AUTHOR("Oracle");
-MODULE_LICENSE("GPL");
-
-module_init(init_o2nm)
-module_exit(exit_o2nm)
diff --git a/drivers/staging/ramster/cluster/nodemanager.h b/drivers/staging/ramster/cluster/nodemanager.h
deleted file mode 100644 (file)
index 09ea2d3..0000000
+++ /dev/null
@@ -1,88 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * nodemanager.h
- *
- * Function prototypes
- *
- * Copyright (C) 2004 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifndef O2CLUSTER_NODEMANAGER_H
-#define O2CLUSTER_NODEMANAGER_H
-
-#include "ocfs2_nodemanager.h"
-
-/* This totally doesn't belong here. */
-#include <linux/configfs.h>
-#include <linux/rbtree.h>
-
-enum o2nm_fence_method {
-       O2NM_FENCE_RESET        = 0,
-       O2NM_FENCE_PANIC,
-       O2NM_FENCE_METHODS,     /* Number of fence methods */
-};
-
-struct o2nm_node {
-       spinlock_t              nd_lock;
-       struct config_item      nd_item;
-       char                    nd_name[O2NM_MAX_NAME_LEN+1]; /* replace? */
-       __u8                    nd_num;
-       /* only one address per node, as attributes, for now. */
-       __be32                  nd_ipv4_address;
-       __be16                  nd_ipv4_port;
-       struct rb_node          nd_ip_node;
-       /* there can be only one local node for now */
-       int                     nd_local;
-
-       unsigned long           nd_set_attributes;
-};
-
-struct o2nm_cluster {
-       struct config_group     cl_group;
-       unsigned                cl_has_local:1;
-       u8                      cl_local_node;
-       rwlock_t                cl_nodes_lock;
-       struct o2nm_node        *cl_nodes[O2NM_MAX_NODES];
-       struct rb_root          cl_node_ip_tree;
-       unsigned int            cl_idle_timeout_ms;
-       unsigned int            cl_keepalive_delay_ms;
-       unsigned int            cl_reconnect_delay_ms;
-       enum o2nm_fence_method  cl_fence_method;
-
-       /* this bitmap is part of a hack for disk bitmap.. will go eventually. - zab */
-       unsigned long   cl_nodes_bitmap[BITS_TO_LONGS(O2NM_MAX_NODES)];
-};
-
-extern struct o2nm_cluster *o2nm_single_cluster;
-
-u8 o2nm_this_node(void);
-
-int o2nm_configured_node_map(unsigned long *map, unsigned bytes);
-struct o2nm_node *o2nm_get_node_by_num(u8 node_num);
-struct o2nm_node *o2nm_get_node_by_ip(__be32 addr);
-void o2nm_node_get(struct o2nm_node *node);
-void o2nm_node_put(struct o2nm_node *node);
-
-int o2nm_depend_item(struct config_item *item);
-void o2nm_undepend_item(struct config_item *item);
-int o2nm_depend_this_node(void);
-void o2nm_undepend_this_node(void);
-
-#endif /* O2CLUSTER_NODEMANAGER_H */
diff --git a/drivers/staging/ramster/cluster/ocfs2_heartbeat.h b/drivers/staging/ramster/cluster/ocfs2_heartbeat.h
deleted file mode 100644 (file)
index 3f4151d..0000000
+++ /dev/null
@@ -1,38 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * ocfs2_heartbeat.h
- *
- * On-disk structures for ocfs2_heartbeat
- *
- * Copyright (C) 2002, 2004 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#ifndef _OCFS2_HEARTBEAT_H
-#define _OCFS2_HEARTBEAT_H
-
-struct o2hb_disk_heartbeat_block {
-       __le64 hb_seq;
-       __u8  hb_node;
-       __u8  hb_pad1[3];
-       __le32 hb_cksum;
-       __le64 hb_generation;
-       __le32 hb_dead_ms;
-};
-
-#endif /* _OCFS2_HEARTBEAT_H */
diff --git a/drivers/staging/ramster/cluster/ocfs2_nodemanager.h b/drivers/staging/ramster/cluster/ocfs2_nodemanager.h
deleted file mode 100644 (file)
index 49b5943..0000000
+++ /dev/null
@@ -1,45 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * ocfs2_nodemanager.h
- *
- * Header describing the interface between userspace and the kernel
- * for the ocfs2_nodemanager module.
- *
- * Copyright (C) 2002, 2004 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifndef _OCFS2_NODEMANAGER_H
-#define _OCFS2_NODEMANAGER_H
-
-#define O2NM_API_VERSION       5
-
-#define O2NM_MAX_NODES         255
-#define O2NM_INVALID_NODE_NUM  255
-
-/* host name, group name, cluster name all 64 bytes */
-#define O2NM_MAX_NAME_LEN        64    // __NEW_UTS_LEN
-
-/*
- * Maximum number of global heartbeat regions allowed.
- * **CAUTION**  Changing this number will break dlm compatibility.
- */
-#define O2NM_MAX_REGIONS       32
-
-#endif /* _OCFS2_NODEMANAGER_H */
diff --git a/drivers/staging/ramster/cluster/quorum.c b/drivers/staging/ramster/cluster/quorum.c
deleted file mode 100644 (file)
index 8f9cea1..0000000
+++ /dev/null
@@ -1,331 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- *
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-/* This quorum hack is only here until we transition to some more rational
- * approach that is driven from userspace.  Honest.  No foolin'.
- *
- * Imagine two nodes lose network connectivity to each other but they're still
- * up and operating in every other way.  Presumably a network timeout indicates
- * that a node is broken and should be recovered.  They can't both recover each
- * other and both carry on without serialising their access to the file system.
- * They need to decide who is authoritative.  Now extend that problem to
- * arbitrary groups of nodes losing connectivity between each other.
- *
- * So we declare that a node which has given up on connecting to a majority
- * of nodes who are still heartbeating will fence itself.
- *
- * There are huge opportunities for races here.  After we give up on a node's
- * connection we need to wait long enough to give heartbeat an opportunity
- * to declare the node as truly dead.  We also need to be careful with the
- * race between when we see a node start heartbeating and when we connect
- * to it.
- *
- * So nodes that are in this transtion put a hold on the quorum decision
- * with a counter.  As they fall out of this transition they drop the count
- * and if they're the last, they fire off the decision.
- */
-#include <linux/kernel.h>
-#include <linux/workqueue.h>
-#include <linux/reboot.h>
-
-#include "heartbeat.h"
-#include "nodemanager.h"
-#define MLOG_MASK_PREFIX ML_QUORUM
-#include "masklog.h"
-#include "quorum.h"
-
-static struct o2quo_state {
-       spinlock_t              qs_lock;
-       struct work_struct      qs_work;
-       int                     qs_pending;
-       int                     qs_heartbeating;
-       unsigned long           qs_hb_bm[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       int                     qs_connected;
-       unsigned long           qs_conn_bm[BITS_TO_LONGS(O2NM_MAX_NODES)];
-       int                     qs_holds;
-       unsigned long           qs_hold_bm[BITS_TO_LONGS(O2NM_MAX_NODES)];
-} o2quo_state;
-
-/* this is horribly heavy-handed.  It should instead flip the file
- * system RO and call some userspace script. */
-static void o2quo_fence_self(void)
-{
-       /* panic spins with interrupts enabled.  with preempt
-        * threads can still schedule, etc, etc */
-       o2hb_stop_all_regions();
-
-       switch (o2nm_single_cluster->cl_fence_method) {
-       case O2NM_FENCE_PANIC:
-               panic("*** ocfs2 is very sorry to be fencing this system by "
-                     "panicing ***\n");
-               break;
-       default:
-               WARN_ON(o2nm_single_cluster->cl_fence_method >=
-                       O2NM_FENCE_METHODS);
-       case O2NM_FENCE_RESET:
-               printk(KERN_ERR "*** ocfs2 is very sorry to be fencing this "
-                      "system by restarting ***\n");
-               emergency_restart();
-               break;
-       };
-}
-
-/* Indicate that a timeout occurred on a hearbeat region write. The
- * other nodes in the cluster may consider us dead at that time so we
- * want to "fence" ourselves so that we don't scribble on the disk
- * after they think they've recovered us. This can't solve all
- * problems related to writeout after recovery but this hack can at
- * least close some of those gaps. When we have real fencing, this can
- * go away as our node would be fenced externally before other nodes
- * begin recovery. */
-void o2quo_disk_timeout(void)
-{
-       o2quo_fence_self();
-}
-
-static void o2quo_make_decision(struct work_struct *work)
-{
-       int quorum;
-       int lowest_hb, lowest_reachable = 0, fence = 0;
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock(&qs->qs_lock);
-
-       lowest_hb = find_first_bit(qs->qs_hb_bm, O2NM_MAX_NODES);
-       if (lowest_hb != O2NM_MAX_NODES)
-               lowest_reachable = test_bit(lowest_hb, qs->qs_conn_bm);
-
-       mlog(0, "heartbeating: %d, connected: %d, "
-            "lowest: %d (%sreachable)\n", qs->qs_heartbeating,
-            qs->qs_connected, lowest_hb, lowest_reachable ? "" : "un");
-
-       if (!test_bit(o2nm_this_node(), qs->qs_hb_bm) ||
-           qs->qs_heartbeating == 1)
-               goto out;
-
-       if (qs->qs_heartbeating & 1) {
-               /* the odd numbered cluster case is straight forward --
-                * if we can't talk to the majority we're hosed */
-               quorum = (qs->qs_heartbeating + 1)/2;
-               if (qs->qs_connected < quorum) {
-                       mlog(ML_ERROR, "fencing this node because it is "
-                            "only connected to %u nodes and %u is needed "
-                            "to make a quorum out of %u heartbeating nodes\n",
-                            qs->qs_connected, quorum,
-                            qs->qs_heartbeating);
-                       fence = 1;
-               }
-       } else {
-               /* the even numbered cluster adds the possibility of each half
-                * of the cluster being able to talk amongst themselves.. in
-                * that case we're hosed if we can't talk to the group that has
-                * the lowest numbered node */
-               quorum = qs->qs_heartbeating / 2;
-               if (qs->qs_connected < quorum) {
-                       mlog(ML_ERROR, "fencing this node because it is "
-                            "only connected to %u nodes and %u is needed "
-                            "to make a quorum out of %u heartbeating nodes\n",
-                            qs->qs_connected, quorum,
-                            qs->qs_heartbeating);
-                       fence = 1;
-               }
-               else if ((qs->qs_connected == quorum) &&
-                        !lowest_reachable) {
-                       mlog(ML_ERROR, "fencing this node because it is "
-                            "connected to a half-quorum of %u out of %u "
-                            "nodes which doesn't include the lowest active "
-                            "node %u\n", quorum, qs->qs_heartbeating,
-                            lowest_hb);
-                       fence = 1;
-               }
-       }
-
-out:
-       spin_unlock(&qs->qs_lock);
-       if (fence)
-               o2quo_fence_self();
-}
-
-static void o2quo_set_hold(struct o2quo_state *qs, u8 node)
-{
-       assert_spin_locked(&qs->qs_lock);
-
-       if (!test_and_set_bit(node, qs->qs_hold_bm)) {
-               qs->qs_holds++;
-               mlog_bug_on_msg(qs->qs_holds == O2NM_MAX_NODES,
-                               "node %u\n", node);
-               mlog(0, "node %u, %d total\n", node, qs->qs_holds);
-       }
-}
-
-static void o2quo_clear_hold(struct o2quo_state *qs, u8 node)
-{
-       assert_spin_locked(&qs->qs_lock);
-
-       if (test_and_clear_bit(node, qs->qs_hold_bm)) {
-               mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1);
-               if (--qs->qs_holds == 0) {
-                       if (qs->qs_pending) {
-                               qs->qs_pending = 0;
-                               schedule_work(&qs->qs_work);
-                       }
-               }
-               mlog_bug_on_msg(qs->qs_holds < 0, "node %u, holds %d\n",
-                               node, qs->qs_holds);
-       }
-}
-
-/* as a node comes up we delay the quorum decision until we know the fate of
- * the connection.  the hold will be droped in conn_up or hb_down.  it might be
- * perpetuated by con_err until hb_down.  if we already have a conn, we might
- * be dropping a hold that conn_up got. */
-void o2quo_hb_up(u8 node)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock(&qs->qs_lock);
-
-       qs->qs_heartbeating++;
-       mlog_bug_on_msg(qs->qs_heartbeating == O2NM_MAX_NODES,
-                       "node %u\n", node);
-       mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node);
-       set_bit(node, qs->qs_hb_bm);
-
-       mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
-
-       if (!test_bit(node, qs->qs_conn_bm))
-               o2quo_set_hold(qs, node);
-       else
-               o2quo_clear_hold(qs, node);
-
-       spin_unlock(&qs->qs_lock);
-}
-
-/* hb going down releases any holds we might have had due to this node from
- * conn_up, conn_err, or hb_up */
-void o2quo_hb_down(u8 node)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock(&qs->qs_lock);
-
-       qs->qs_heartbeating--;
-       mlog_bug_on_msg(qs->qs_heartbeating < 0,
-                       "node %u, %d heartbeating\n",
-                       node, qs->qs_heartbeating);
-       mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node);
-       clear_bit(node, qs->qs_hb_bm);
-
-       mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
-
-       o2quo_clear_hold(qs, node);
-
-       spin_unlock(&qs->qs_lock);
-}
-
-/* this tells us that we've decided that the node is still heartbeating
- * even though we've lost it's conn.  it must only be called after conn_err
- * and indicates that we must now make a quorum decision in the future,
- * though we might be doing so after waiting for holds to drain.  Here
- * we'll be dropping the hold from conn_err. */
-void o2quo_hb_still_up(u8 node)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock(&qs->qs_lock);
-
-       mlog(0, "node %u\n", node);
-
-       qs->qs_pending = 1;
-       o2quo_clear_hold(qs, node);
-
-       spin_unlock(&qs->qs_lock);
-}
-
-/* This is analogous to hb_up.  as a node's connection comes up we delay the
- * quorum decision until we see it heartbeating.  the hold will be droped in
- * hb_up or hb_down.  it might be perpetuated by con_err until hb_down.  if
- * it's already heartbeating we we might be dropping a hold that conn_up got.
- * */
-void o2quo_conn_up(u8 node)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock(&qs->qs_lock);
-
-       qs->qs_connected++;
-       mlog_bug_on_msg(qs->qs_connected == O2NM_MAX_NODES,
-                       "node %u\n", node);
-       mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node);
-       set_bit(node, qs->qs_conn_bm);
-
-       mlog(0, "node %u, %d total\n", node, qs->qs_connected);
-
-       if (!test_bit(node, qs->qs_hb_bm))
-               o2quo_set_hold(qs, node);
-       else
-               o2quo_clear_hold(qs, node);
-
-       spin_unlock(&qs->qs_lock);
-}
-
-/* we've decided that we won't ever be connecting to the node again.  if it's
- * still heartbeating we grab a hold that will delay decisions until either the
- * node stops heartbeating from hb_down or the caller decides that the node is
- * still up and calls still_up */
-void o2quo_conn_err(u8 node)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock(&qs->qs_lock);
-
-       if (test_bit(node, qs->qs_conn_bm)) {
-               qs->qs_connected--;
-               mlog_bug_on_msg(qs->qs_connected < 0,
-                               "node %u, connected %d\n",
-                               node, qs->qs_connected);
-
-               clear_bit(node, qs->qs_conn_bm);
-       }
-
-       mlog(0, "node %u, %d total\n", node, qs->qs_connected);
-
-       if (test_bit(node, qs->qs_hb_bm))
-               o2quo_set_hold(qs, node);
-
-       spin_unlock(&qs->qs_lock);
-}
-
-void o2quo_init(void)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       spin_lock_init(&qs->qs_lock);
-       INIT_WORK(&qs->qs_work, o2quo_make_decision);
-}
-
-void o2quo_exit(void)
-{
-       struct o2quo_state *qs = &o2quo_state;
-
-       flush_work_sync(&qs->qs_work);
-}
diff --git a/drivers/staging/ramster/cluster/quorum.h b/drivers/staging/ramster/cluster/quorum.h
deleted file mode 100644 (file)
index 6649cc6..0000000
+++ /dev/null
@@ -1,36 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifndef O2CLUSTER_QUORUM_H
-#define O2CLUSTER_QUORUM_H
-
-void o2quo_init(void);
-void o2quo_exit(void);
-
-void o2quo_hb_up(u8 node);
-void o2quo_hb_down(u8 node);
-void o2quo_hb_still_up(u8 node);
-void o2quo_conn_up(u8 node);
-void o2quo_conn_err(u8 node);
-void o2quo_disk_timeout(void);
-
-#endif /* O2CLUSTER_QUORUM_H */
diff --git a/drivers/staging/ramster/cluster/sys.c b/drivers/staging/ramster/cluster/sys.c
deleted file mode 100644 (file)
index a4b0773..0000000
+++ /dev/null
@@ -1,82 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * sys.c
- *
- * OCFS2 cluster sysfs interface
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation,
- * version 2 of the License.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#include <linux/kernel.h>
-#include <linux/module.h>
-#include <linux/kobject.h>
-#include <linux/sysfs.h>
-#include <linux/fs.h>
-
-#include "ocfs2_nodemanager.h"
-#include "masklog.h"
-#include "sys.h"
-
-
-static ssize_t version_show(struct kobject *kobj, struct kobj_attribute *attr,
-                           char *buf)
-{
-       return snprintf(buf, PAGE_SIZE, "%u\n", O2NM_API_VERSION);
-}
-static struct kobj_attribute attr_version =
-       __ATTR(interface_revision, S_IFREG | S_IRUGO, version_show, NULL);
-
-static struct attribute *o2cb_attrs[] = {
-       &attr_version.attr,
-       NULL,
-};
-
-static struct attribute_group o2cb_attr_group = {
-       .attrs = o2cb_attrs,
-};
-
-static struct kset *o2cb_kset;
-
-void o2cb_sys_shutdown(void)
-{
-       mlog_sys_shutdown();
-       kset_unregister(o2cb_kset);
-}
-
-int o2cb_sys_init(void)
-{
-       int ret;
-
-       o2cb_kset = kset_create_and_add("o2cb", NULL, fs_kobj);
-       if (!o2cb_kset)
-               return -ENOMEM;
-
-       ret = sysfs_create_group(&o2cb_kset->kobj, &o2cb_attr_group);
-       if (ret)
-               goto error;
-
-       ret = mlog_sys_init(o2cb_kset);
-       if (ret)
-               goto error;
-       return 0;
-error:
-       kset_unregister(o2cb_kset);
-       return ret;
-}
diff --git a/drivers/staging/ramster/cluster/sys.h b/drivers/staging/ramster/cluster/sys.h
deleted file mode 100644 (file)
index d66b8ab..0000000
+++ /dev/null
@@ -1,33 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * sys.h
- *
- * Function prototypes for o2cb sysfs interface
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation,
- * version 2 of the License.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifndef O2CLUSTER_SYS_H
-#define O2CLUSTER_SYS_H
-
-void o2cb_sys_shutdown(void);
-int o2cb_sys_init(void);
-
-#endif /* O2CLUSTER_SYS_H */
diff --git a/drivers/staging/ramster/cluster/tcp.c b/drivers/staging/ramster/cluster/tcp.c
deleted file mode 100644 (file)
index e9bd00e..0000000
+++ /dev/null
@@ -1,2262 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- *
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2004 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- * ----
- *
- * Callers for this were originally written against a very simple synchronus
- * API.  This implementation reflects those simple callers.  Some day I'm sure
- * we'll need to move to a more robust posting/callback mechanism.
- *
- * Transmit calls pass in kernel virtual addresses and block copying this into
- * the socket's tx buffers via a usual blocking sendmsg.  They'll block waiting
- * for a failed socket to timeout.  TX callers can also pass in a poniter to an
- * 'int' which gets filled with an errno off the wire in response to the
- * message they send.
- *
- * Handlers for unsolicited messages are registered.  Each socket has a page
- * that incoming data is copied into.  First the header, then the data.
- * Handlers are called from only one thread with a reference to this per-socket
- * page.  This page is destroyed after the handler call, so it can't be
- * referenced beyond the call.  Handlers may block but are discouraged from
- * doing so.
- *
- * Any framing errors (bad magic, large payload lengths) close a connection.
- *
- * Our sock_container holds the state we associate with a socket.  It's current
- * framing state is held there as well as the refcounting we do around when it
- * is safe to tear down the socket.  The socket is only finally torn down from
- * the container when the container loses all of its references -- so as long
- * as you hold a ref on the container you can trust that the socket is valid
- * for use with kernel socket APIs.
- *
- * Connections are initiated between a pair of nodes when the node with the
- * higher node number gets a heartbeat callback which indicates that the lower
- * numbered node has started heartbeating.  The lower numbered node is passive
- * and only accepts the connection if the higher numbered node is heartbeating.
- */
-
-#include <linux/kernel.h>
-#include <linux/jiffies.h>
-#include <linux/slab.h>
-#include <linux/idr.h>
-#include <linux/kref.h>
-#include <linux/net.h>
-#include <linux/export.h>
-#include <net/tcp.h>
-
-#include <asm/uaccess.h>
-
-#include "heartbeat.h"
-#include "tcp.h"
-#include "nodemanager.h"
-#define MLOG_MASK_PREFIX ML_TCP
-#include "masklog.h"
-#include "quorum.h"
-
-#include "tcp_internal.h"
-
-#define SC_NODEF_FMT "node %s (num %u) at %pI4:%u"
-#define SC_NODEF_ARGS(sc) sc->sc_node->nd_name, sc->sc_node->nd_num,   \
-                         &sc->sc_node->nd_ipv4_address,                \
-                         ntohs(sc->sc_node->nd_ipv4_port)
-
-/*
- * In the following two log macros, the whitespace after the ',' just
- * before ##args is intentional. Otherwise, gcc 2.95 will eat the
- * previous token if args expands to nothing.
- */
-#define msglog(hdr, fmt, args...) do {                                 \
-       typeof(hdr) __hdr = (hdr);                                      \
-       mlog(ML_MSG, "[mag %u len %u typ %u stat %d sys_stat %d "       \
-            "key %08x num %u] " fmt,                                   \
-            be16_to_cpu(__hdr->magic), be16_to_cpu(__hdr->data_len),   \
-            be16_to_cpu(__hdr->msg_type), be32_to_cpu(__hdr->status),  \
-            be32_to_cpu(__hdr->sys_status), be32_to_cpu(__hdr->key),   \
-            be32_to_cpu(__hdr->msg_num) ,  ##args);                    \
-} while (0)
-
-#define sclog(sc, fmt, args...) do {                                   \
-       typeof(sc) __sc = (sc);                                         \
-       mlog(ML_SOCKET, "[sc %p refs %d sock %p node %u page %p "       \
-            "pg_off %zu] " fmt, __sc,                                  \
-            atomic_read(&__sc->sc_kref.refcount), __sc->sc_sock,       \
-           __sc->sc_node->nd_num, __sc->sc_page, __sc->sc_page_off ,   \
-           ##args);                                                    \
-} while (0)
-
-static DEFINE_RWLOCK(o2net_handler_lock);
-static struct rb_root o2net_handler_tree = RB_ROOT;
-
-static struct o2net_node o2net_nodes[O2NM_MAX_NODES];
-
-/* XXX someday we'll need better accounting */
-static struct socket *o2net_listen_sock = NULL;
-
-/*
- * listen work is only queued by the listening socket callbacks on the
- * o2net_wq.  teardown detaches the callbacks before destroying the workqueue.
- * quorum work is queued as sock containers are shutdown.. stop_listening
- * tears down all the node's sock containers, preventing future shutdowns
- * and queued quroum work, before canceling delayed quorum work and
- * destroying the work queue.
- */
-static struct workqueue_struct *o2net_wq;
-static struct work_struct o2net_listen_work;
-
-static struct o2hb_callback_func o2net_hb_up, o2net_hb_down;
-#define O2NET_HB_PRI 0x1
-
-static struct o2net_handshake *o2net_hand;
-static struct o2net_msg *o2net_keep_req, *o2net_keep_resp;
-
-static int o2net_sys_err_translations[O2NET_ERR_MAX] =
-               {[O2NET_ERR_NONE]       = 0,
-                [O2NET_ERR_NO_HNDLR]   = -ENOPROTOOPT,
-                [O2NET_ERR_OVERFLOW]   = -EOVERFLOW,
-                [O2NET_ERR_DIED]       = -EHOSTDOWN,};
-
-/* can't quite avoid *all* internal declarations :/ */
-static void o2net_sc_connect_completed(struct work_struct *work);
-static void o2net_rx_until_empty(struct work_struct *work);
-static void o2net_shutdown_sc(struct work_struct *work);
-static void o2net_listen_data_ready(struct sock *sk, int bytes);
-static void o2net_sc_send_keep_req(struct work_struct *work);
-static void o2net_idle_timer(unsigned long data);
-static void o2net_sc_postpone_idle(struct o2net_sock_container *sc);
-static void o2net_sc_reset_idle_timer(struct o2net_sock_container *sc);
-
-#ifdef CONFIG_DEBUG_FS
-static void o2net_init_nst(struct o2net_send_tracking *nst, u32 msgtype,
-                          u32 msgkey, struct task_struct *task, u8 node)
-{
-       INIT_LIST_HEAD(&nst->st_net_debug_item);
-       nst->st_task = task;
-       nst->st_msg_type = msgtype;
-       nst->st_msg_key = msgkey;
-       nst->st_node = node;
-}
-
-static inline void o2net_set_nst_sock_time(struct o2net_send_tracking *nst)
-{
-       nst->st_sock_time = ktime_get();
-}
-
-static inline void o2net_set_nst_send_time(struct o2net_send_tracking *nst)
-{
-       nst->st_send_time = ktime_get();
-}
-
-static inline void o2net_set_nst_status_time(struct o2net_send_tracking *nst)
-{
-       nst->st_status_time = ktime_get();
-}
-
-static inline void o2net_set_nst_sock_container(struct o2net_send_tracking *nst,
-                                               struct o2net_sock_container *sc)
-{
-       nst->st_sc = sc;
-}
-
-static inline void o2net_set_nst_msg_id(struct o2net_send_tracking *nst,
-                                       u32 msg_id)
-{
-       nst->st_id = msg_id;
-}
-
-static inline void o2net_set_sock_timer(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_timer = ktime_get();
-}
-
-static inline void o2net_set_data_ready_time(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_data_ready = ktime_get();
-}
-
-static inline void o2net_set_advance_start_time(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_advance_start = ktime_get();
-}
-
-static inline void o2net_set_advance_stop_time(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_advance_stop = ktime_get();
-}
-
-static inline void o2net_set_func_start_time(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_func_start = ktime_get();
-}
-
-static inline void o2net_set_func_stop_time(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_func_stop = ktime_get();
-}
-
-#else  /* CONFIG_DEBUG_FS */
-# define o2net_init_nst(a, b, c, d, e)
-# define o2net_set_nst_sock_time(a)
-# define o2net_set_nst_send_time(a)
-# define o2net_set_nst_status_time(a)
-# define o2net_set_nst_sock_container(a, b)
-# define o2net_set_nst_msg_id(a, b)
-# define o2net_set_sock_timer(a)
-# define o2net_set_data_ready_time(a)
-# define o2net_set_advance_start_time(a)
-# define o2net_set_advance_stop_time(a)
-# define o2net_set_func_start_time(a)
-# define o2net_set_func_stop_time(a)
-#endif /* CONFIG_DEBUG_FS */
-
-#ifdef CONFIG_OCFS2_FS_STATS
-static ktime_t o2net_get_func_run_time(struct o2net_sock_container *sc)
-{
-       return ktime_sub(sc->sc_tv_func_stop, sc->sc_tv_func_start);
-}
-
-static void o2net_update_send_stats(struct o2net_send_tracking *nst,
-                                   struct o2net_sock_container *sc)
-{
-       sc->sc_tv_status_total = ktime_add(sc->sc_tv_status_total,
-                                          ktime_sub(ktime_get(),
-                                                    nst->st_status_time));
-       sc->sc_tv_send_total = ktime_add(sc->sc_tv_send_total,
-                                        ktime_sub(nst->st_status_time,
-                                                  nst->st_send_time));
-       sc->sc_tv_acquiry_total = ktime_add(sc->sc_tv_acquiry_total,
-                                           ktime_sub(nst->st_send_time,
-                                                     nst->st_sock_time));
-       sc->sc_send_count++;
-}
-
-static void o2net_update_recv_stats(struct o2net_sock_container *sc)
-{
-       sc->sc_tv_process_total = ktime_add(sc->sc_tv_process_total,
-                                           o2net_get_func_run_time(sc));
-       sc->sc_recv_count++;
-}
-
-#else
-
-# define o2net_update_send_stats(a, b)
-
-# define o2net_update_recv_stats(sc)
-
-#endif /* CONFIG_OCFS2_FS_STATS */
-
-static inline int o2net_reconnect_delay(void)
-{
-       return o2nm_single_cluster->cl_reconnect_delay_ms;
-}
-
-static inline int o2net_keepalive_delay(void)
-{
-       return o2nm_single_cluster->cl_keepalive_delay_ms;
-}
-
-static inline int o2net_idle_timeout(void)
-{
-       return o2nm_single_cluster->cl_idle_timeout_ms;
-}
-
-static inline int o2net_sys_err_to_errno(enum o2net_system_error err)
-{
-       int trans;
-       BUG_ON(err >= O2NET_ERR_MAX);
-       trans = o2net_sys_err_translations[err];
-
-       /* Just in case we mess up the translation table above */
-       BUG_ON(err != O2NET_ERR_NONE && trans == 0);
-       return trans;
-}
-
-#ifdef CONFIG_RAMSTER
-struct o2net_node *o2net_nn_from_num(u8 node_num)
-#else
-static struct o2net_node * o2net_nn_from_num(u8 node_num)
-#endif
-{
-       BUG_ON(node_num >= ARRAY_SIZE(o2net_nodes));
-       return &o2net_nodes[node_num];
-}
-
-static u8 o2net_num_from_nn(struct o2net_node *nn)
-{
-       BUG_ON(nn == NULL);
-       return nn - o2net_nodes;
-}
-
-/* ------------------------------------------------------------ */
-
-static int o2net_prep_nsw(struct o2net_node *nn, struct o2net_status_wait *nsw)
-{
-       int ret = 0;
-
-       do {
-               if (!idr_pre_get(&nn->nn_status_idr, GFP_ATOMIC)) {
-                       ret = -EAGAIN;
-                       break;
-               }
-               spin_lock(&nn->nn_lock);
-               ret = idr_get_new(&nn->nn_status_idr, nsw, &nsw->ns_id);
-               if (ret == 0)
-                       list_add_tail(&nsw->ns_node_item,
-                                     &nn->nn_status_list);
-               spin_unlock(&nn->nn_lock);
-       } while (ret == -EAGAIN);
-
-       if (ret == 0)  {
-               init_waitqueue_head(&nsw->ns_wq);
-               nsw->ns_sys_status = O2NET_ERR_NONE;
-               nsw->ns_status = 0;
-       }
-
-       return ret;
-}
-
-static void o2net_complete_nsw_locked(struct o2net_node *nn,
-                                     struct o2net_status_wait *nsw,
-                                     enum o2net_system_error sys_status,
-                                     s32 status)
-{
-       assert_spin_locked(&nn->nn_lock);
-
-       if (!list_empty(&nsw->ns_node_item)) {
-               list_del_init(&nsw->ns_node_item);
-               nsw->ns_sys_status = sys_status;
-               nsw->ns_status = status;
-               idr_remove(&nn->nn_status_idr, nsw->ns_id);
-               wake_up(&nsw->ns_wq);
-       }
-}
-
-static void o2net_complete_nsw(struct o2net_node *nn,
-                              struct o2net_status_wait *nsw,
-                              u64 id, enum o2net_system_error sys_status,
-                              s32 status)
-{
-       spin_lock(&nn->nn_lock);
-       if (nsw == NULL) {
-               if (id > INT_MAX)
-                       goto out;
-
-               nsw = idr_find(&nn->nn_status_idr, id);
-               if (nsw == NULL)
-                       goto out;
-       }
-
-       o2net_complete_nsw_locked(nn, nsw, sys_status, status);
-
-out:
-       spin_unlock(&nn->nn_lock);
-       return;
-}
-
-static void o2net_complete_nodes_nsw(struct o2net_node *nn)
-{
-       struct o2net_status_wait *nsw, *tmp;
-       unsigned int num_kills = 0;
-
-       assert_spin_locked(&nn->nn_lock);
-
-       list_for_each_entry_safe(nsw, tmp, &nn->nn_status_list, ns_node_item) {
-               o2net_complete_nsw_locked(nn, nsw, O2NET_ERR_DIED, 0);
-               num_kills++;
-       }
-
-       mlog(0, "completed %d messages for node %u\n", num_kills,
-            o2net_num_from_nn(nn));
-}
-
-static int o2net_nsw_completed(struct o2net_node *nn,
-                              struct o2net_status_wait *nsw)
-{
-       int completed;
-       spin_lock(&nn->nn_lock);
-       completed = list_empty(&nsw->ns_node_item);
-       spin_unlock(&nn->nn_lock);
-       return completed;
-}
-
-/* ------------------------------------------------------------ */
-
-static void sc_kref_release(struct kref *kref)
-{
-       struct o2net_sock_container *sc = container_of(kref,
-                                       struct o2net_sock_container, sc_kref);
-       BUG_ON(timer_pending(&sc->sc_idle_timeout));
-
-       sclog(sc, "releasing\n");
-
-       if (sc->sc_sock) {
-               sock_release(sc->sc_sock);
-               sc->sc_sock = NULL;
-       }
-
-       o2nm_undepend_item(&sc->sc_node->nd_item);
-       o2nm_node_put(sc->sc_node);
-       sc->sc_node = NULL;
-
-       o2net_debug_del_sc(sc);
-       kfree(sc);
-}
-
-static void sc_put(struct o2net_sock_container *sc)
-{
-       sclog(sc, "put\n");
-       kref_put(&sc->sc_kref, sc_kref_release);
-}
-static void sc_get(struct o2net_sock_container *sc)
-{
-       sclog(sc, "get\n");
-       kref_get(&sc->sc_kref);
-}
-static struct o2net_sock_container *sc_alloc(struct o2nm_node *node)
-{
-       struct o2net_sock_container *sc, *ret = NULL;
-       struct page *page = NULL;
-       int status = 0;
-
-       page = alloc_page(GFP_NOFS);
-       sc = kzalloc(sizeof(*sc), GFP_NOFS);
-       if (sc == NULL || page == NULL)
-               goto out;
-
-       kref_init(&sc->sc_kref);
-       o2nm_node_get(node);
-       sc->sc_node = node;
-
-       /* pin the node item of the remote node */
-       status = o2nm_depend_item(&node->nd_item);
-       if (status) {
-               mlog_errno(status);
-               o2nm_node_put(node);
-               goto out;
-       }
-       INIT_WORK(&sc->sc_connect_work, o2net_sc_connect_completed);
-       INIT_WORK(&sc->sc_rx_work, o2net_rx_until_empty);
-       INIT_WORK(&sc->sc_shutdown_work, o2net_shutdown_sc);
-       INIT_DELAYED_WORK(&sc->sc_keepalive_work, o2net_sc_send_keep_req);
-
-       init_timer(&sc->sc_idle_timeout);
-       sc->sc_idle_timeout.function = o2net_idle_timer;
-       sc->sc_idle_timeout.data = (unsigned long)sc;
-
-       sclog(sc, "alloced\n");
-
-       ret = sc;
-       sc->sc_page = page;
-       o2net_debug_add_sc(sc);
-       sc = NULL;
-       page = NULL;
-
-out:
-       if (page)
-               __free_page(page);
-       kfree(sc);
-
-       return ret;
-}
-
-/* ------------------------------------------------------------ */
-
-static void o2net_sc_queue_work(struct o2net_sock_container *sc,
-                               struct work_struct *work)
-{
-       sc_get(sc);
-       if (!queue_work(o2net_wq, work))
-               sc_put(sc);
-}
-static void o2net_sc_queue_delayed_work(struct o2net_sock_container *sc,
-                                       struct delayed_work *work,
-                                       int delay)
-{
-       sc_get(sc);
-       if (!queue_delayed_work(o2net_wq, work, delay))
-               sc_put(sc);
-}
-static void o2net_sc_cancel_delayed_work(struct o2net_sock_container *sc,
-                                        struct delayed_work *work)
-{
-       if (cancel_delayed_work(work))
-               sc_put(sc);
-}
-
-static atomic_t o2net_connected_peers = ATOMIC_INIT(0);
-
-int o2net_num_connected_peers(void)
-{
-       return atomic_read(&o2net_connected_peers);
-}
-
-static void o2net_set_nn_state(struct o2net_node *nn,
-                              struct o2net_sock_container *sc,
-                              unsigned valid, int err)
-{
-       int was_valid = nn->nn_sc_valid;
-       int was_err = nn->nn_persistent_error;
-       struct o2net_sock_container *old_sc = nn->nn_sc;
-
-       assert_spin_locked(&nn->nn_lock);
-
-       if (old_sc && !sc)
-               atomic_dec(&o2net_connected_peers);
-       else if (!old_sc && sc)
-               atomic_inc(&o2net_connected_peers);
-
-       /* the node num comparison and single connect/accept path should stop
-        * an non-null sc from being overwritten with another */
-       BUG_ON(sc && nn->nn_sc && nn->nn_sc != sc);
-       mlog_bug_on_msg(err && valid, "err %d valid %u\n", err, valid);
-       mlog_bug_on_msg(valid && !sc, "valid %u sc %p\n", valid, sc);
-
-       if (was_valid && !valid && err == 0)
-               err = -ENOTCONN;
-
-       mlog(ML_CONN, "node %u sc: %p -> %p, valid %u -> %u, err %d -> %d\n",
-            o2net_num_from_nn(nn), nn->nn_sc, sc, nn->nn_sc_valid, valid,
-            nn->nn_persistent_error, err);
-
-       nn->nn_sc = sc;
-       nn->nn_sc_valid = valid ? 1 : 0;
-       nn->nn_persistent_error = err;
-
-       /* mirrors o2net_tx_can_proceed() */
-       if (nn->nn_persistent_error || nn->nn_sc_valid)
-               wake_up(&nn->nn_sc_wq);
-
-       if (!was_err && nn->nn_persistent_error) {
-               o2quo_conn_err(o2net_num_from_nn(nn));
-               queue_delayed_work(o2net_wq, &nn->nn_still_up,
-                                  msecs_to_jiffies(O2NET_QUORUM_DELAY_MS));
-       }
-
-       if (was_valid && !valid) {
-               printk(KERN_NOTICE "o2net: No longer connected to "
-                      SC_NODEF_FMT "\n", SC_NODEF_ARGS(old_sc));
-               o2net_complete_nodes_nsw(nn);
-       }
-
-       if (!was_valid && valid) {
-               o2quo_conn_up(o2net_num_from_nn(nn));
-               cancel_delayed_work(&nn->nn_connect_expired);
-               printk(KERN_NOTICE "o2net: %s " SC_NODEF_FMT "\n",
-                      o2nm_this_node() > sc->sc_node->nd_num ?
-                      "Connected to" : "Accepted connection from",
-                      SC_NODEF_ARGS(sc));
-       }
-
-       /* trigger the connecting worker func as long as we're not valid,
-        * it will back off if it shouldn't connect.  This can be called
-        * from node config teardown and so needs to be careful about
-        * the work queue actually being up. */
-       if (!valid && o2net_wq) {
-               unsigned long delay;
-               /* delay if we're within a RECONNECT_DELAY of the
-                * last attempt */
-               delay = (nn->nn_last_connect_attempt +
-                        msecs_to_jiffies(o2net_reconnect_delay()))
-                       - jiffies;
-               if (delay > msecs_to_jiffies(o2net_reconnect_delay()))
-                       delay = 0;
-               mlog(ML_CONN, "queueing conn attempt in %lu jiffies\n", delay);
-               queue_delayed_work(o2net_wq, &nn->nn_connect_work, delay);
-
-               /*
-                * Delay the expired work after idle timeout.
-                *
-                * We might have lots of failed connection attempts that run
-                * through here but we only cancel the connect_expired work when
-                * a connection attempt succeeds.  So only the first enqueue of
-                * the connect_expired work will do anything.  The rest will see
-                * that it's already queued and do nothing.
-                */
-               delay += msecs_to_jiffies(o2net_idle_timeout());
-               queue_delayed_work(o2net_wq, &nn->nn_connect_expired, delay);
-       }
-
-       /* keep track of the nn's sc ref for the caller */
-       if ((old_sc == NULL) && sc)
-               sc_get(sc);
-       if (old_sc && (old_sc != sc)) {
-               o2net_sc_queue_work(old_sc, &old_sc->sc_shutdown_work);
-               sc_put(old_sc);
-       }
-}
-
-/* see o2net_register_callbacks() */
-static void o2net_data_ready(struct sock *sk, int bytes)
-{
-       void (*ready)(struct sock *sk, int bytes);
-
-       read_lock(&sk->sk_callback_lock);
-       if (sk->sk_user_data) {
-               struct o2net_sock_container *sc = sk->sk_user_data;
-               sclog(sc, "data_ready hit\n");
-               o2net_set_data_ready_time(sc);
-               o2net_sc_queue_work(sc, &sc->sc_rx_work);
-               ready = sc->sc_data_ready;
-       } else {
-               ready = sk->sk_data_ready;
-       }
-       read_unlock(&sk->sk_callback_lock);
-
-       ready(sk, bytes);
-}
-
-/* see o2net_register_callbacks() */
-static void o2net_state_change(struct sock *sk)
-{
-       void (*state_change)(struct sock *sk);
-       struct o2net_sock_container *sc;
-
-       read_lock(&sk->sk_callback_lock);
-       sc = sk->sk_user_data;
-       if (sc == NULL) {
-               state_change = sk->sk_state_change;
-               goto out;
-       }
-
-       sclog(sc, "state_change to %d\n", sk->sk_state);
-
-       state_change = sc->sc_state_change;
-
-       switch(sk->sk_state) {
-               /* ignore connecting sockets as they make progress */
-               case TCP_SYN_SENT:
-               case TCP_SYN_RECV:
-                       break;
-               case TCP_ESTABLISHED:
-                       o2net_sc_queue_work(sc, &sc->sc_connect_work);
-                       break;
-               default:
-                       printk(KERN_INFO "o2net: Connection to " SC_NODEF_FMT
-                             " shutdown, state %d\n",
-                             SC_NODEF_ARGS(sc), sk->sk_state);
-                       o2net_sc_queue_work(sc, &sc->sc_shutdown_work);
-                       break;
-       }
-out:
-       read_unlock(&sk->sk_callback_lock);
-       state_change(sk);
-}
-
-/*
- * we register callbacks so we can queue work on events before calling
- * the original callbacks.  our callbacks our careful to test user_data
- * to discover when they've reaced with o2net_unregister_callbacks().
- */
-static void o2net_register_callbacks(struct sock *sk,
-                                    struct o2net_sock_container *sc)
-{
-       write_lock_bh(&sk->sk_callback_lock);
-
-       /* accepted sockets inherit the old listen socket data ready */
-       if (sk->sk_data_ready == o2net_listen_data_ready) {
-               sk->sk_data_ready = sk->sk_user_data;
-               sk->sk_user_data = NULL;
-       }
-
-       BUG_ON(sk->sk_user_data != NULL);
-       sk->sk_user_data = sc;
-       sc_get(sc);
-
-       sc->sc_data_ready = sk->sk_data_ready;
-       sc->sc_state_change = sk->sk_state_change;
-       sk->sk_data_ready = o2net_data_ready;
-       sk->sk_state_change = o2net_state_change;
-
-       mutex_init(&sc->sc_send_lock);
-
-       write_unlock_bh(&sk->sk_callback_lock);
-}
-
-static int o2net_unregister_callbacks(struct sock *sk,
-                                  struct o2net_sock_container *sc)
-{
-       int ret = 0;
-
-       write_lock_bh(&sk->sk_callback_lock);
-       if (sk->sk_user_data == sc) {
-               ret = 1;
-               sk->sk_user_data = NULL;
-               sk->sk_data_ready = sc->sc_data_ready;
-               sk->sk_state_change = sc->sc_state_change;
-       }
-       write_unlock_bh(&sk->sk_callback_lock);
-
-       return ret;
-}
-
-/*
- * this is a little helper that is called by callers who have seen a problem
- * with an sc and want to detach it from the nn if someone already hasn't beat
- * them to it.  if an error is given then the shutdown will be persistent
- * and pending transmits will be canceled.
- */
-static void o2net_ensure_shutdown(struct o2net_node *nn,
-                                  struct o2net_sock_container *sc,
-                                  int err)
-{
-       spin_lock(&nn->nn_lock);
-       if (nn->nn_sc == sc)
-               o2net_set_nn_state(nn, NULL, 0, err);
-       spin_unlock(&nn->nn_lock);
-}
-
-/*
- * This work queue function performs the blocking parts of socket shutdown.  A
- * few paths lead here.  set_nn_state will trigger this callback if it sees an
- * sc detached from the nn.  state_change will also trigger this callback
- * directly when it sees errors.  In that case we need to call set_nn_state
- * ourselves as state_change couldn't get the nn_lock and call set_nn_state
- * itself.
- */
-static void o2net_shutdown_sc(struct work_struct *work)
-{
-       struct o2net_sock_container *sc =
-               container_of(work, struct o2net_sock_container,
-                            sc_shutdown_work);
-       struct o2net_node *nn = o2net_nn_from_num(sc->sc_node->nd_num);
-
-       sclog(sc, "shutting down\n");
-
-       /* drop the callbacks ref and call shutdown only once */
-       if (o2net_unregister_callbacks(sc->sc_sock->sk, sc)) {
-               /* we shouldn't flush as we're in the thread, the
-                * races with pending sc work structs are harmless */
-               del_timer_sync(&sc->sc_idle_timeout);
-               o2net_sc_cancel_delayed_work(sc, &sc->sc_keepalive_work);
-               sc_put(sc);
-               kernel_sock_shutdown(sc->sc_sock, SHUT_RDWR);
-       }
-
-       /* not fatal so failed connects before the other guy has our
-        * heartbeat can be retried */
-       o2net_ensure_shutdown(nn, sc, 0);
-       sc_put(sc);
-}
-
-/* ------------------------------------------------------------ */
-
-static int o2net_handler_cmp(struct o2net_msg_handler *nmh, u32 msg_type,
-                            u32 key)
-{
-       int ret = memcmp(&nmh->nh_key, &key, sizeof(key));
-
-       if (ret == 0)
-               ret = memcmp(&nmh->nh_msg_type, &msg_type, sizeof(msg_type));
-
-       return ret;
-}
-
-static struct o2net_msg_handler *
-o2net_handler_tree_lookup(u32 msg_type, u32 key, struct rb_node ***ret_p,
-                         struct rb_node **ret_parent)
-{
-        struct rb_node **p = &o2net_handler_tree.rb_node;
-        struct rb_node *parent = NULL;
-       struct o2net_msg_handler *nmh, *ret = NULL;
-       int cmp;
-
-        while (*p) {
-                parent = *p;
-                nmh = rb_entry(parent, struct o2net_msg_handler, nh_node);
-               cmp = o2net_handler_cmp(nmh, msg_type, key);
-
-                if (cmp < 0)
-                        p = &(*p)->rb_left;
-                else if (cmp > 0)
-                        p = &(*p)->rb_right;
-                else {
-                       ret = nmh;
-                        break;
-               }
-        }
-
-        if (ret_p != NULL)
-                *ret_p = p;
-        if (ret_parent != NULL)
-                *ret_parent = parent;
-
-        return ret;
-}
-
-static void o2net_handler_kref_release(struct kref *kref)
-{
-       struct o2net_msg_handler *nmh;
-       nmh = container_of(kref, struct o2net_msg_handler, nh_kref);
-
-       kfree(nmh);
-}
-
-static void o2net_handler_put(struct o2net_msg_handler *nmh)
-{
-       kref_put(&nmh->nh_kref, o2net_handler_kref_release);
-}
-
-/* max_len is protection for the handler func.  incoming messages won't
- * be given to the handler if their payload is longer than the max. */
-int o2net_register_handler(u32 msg_type, u32 key, u32 max_len,
-                          o2net_msg_handler_func *func, void *data,
-                          o2net_post_msg_handler_func *post_func,
-                          struct list_head *unreg_list)
-{
-       struct o2net_msg_handler *nmh = NULL;
-       struct rb_node **p, *parent;
-       int ret = 0;
-
-       if (max_len > O2NET_MAX_PAYLOAD_BYTES) {
-               mlog(0, "max_len for message handler out of range: %u\n",
-                       max_len);
-               ret = -EINVAL;
-               goto out;
-       }
-
-       if (!msg_type) {
-               mlog(0, "no message type provided: %u, %p\n", msg_type, func);
-               ret = -EINVAL;
-               goto out;
-
-       }
-       if (!func) {
-               mlog(0, "no message handler provided: %u, %p\n",
-                      msg_type, func);
-               ret = -EINVAL;
-               goto out;
-       }
-
-               nmh = kzalloc(sizeof(struct o2net_msg_handler), GFP_NOFS);
-       if (nmh == NULL) {
-               ret = -ENOMEM;
-               goto out;
-       }
-
-       nmh->nh_func = func;
-       nmh->nh_func_data = data;
-       nmh->nh_post_func = post_func;
-       nmh->nh_msg_type = msg_type;
-       nmh->nh_max_len = max_len;
-       nmh->nh_key = key;
-       /* the tree and list get this ref.. they're both removed in
-        * unregister when this ref is dropped */
-       kref_init(&nmh->nh_kref);
-       INIT_LIST_HEAD(&nmh->nh_unregister_item);
-
-       write_lock(&o2net_handler_lock);
-       if (o2net_handler_tree_lookup(msg_type, key, &p, &parent))
-               ret = -EEXIST;
-       else {
-               rb_link_node(&nmh->nh_node, parent, p);
-               rb_insert_color(&nmh->nh_node, &o2net_handler_tree);
-               list_add_tail(&nmh->nh_unregister_item, unreg_list);
-
-               mlog(ML_TCP, "registered handler func %p type %u key %08x\n",
-                    func, msg_type, key);
-               /* we've had some trouble with handlers seemingly vanishing. */
-               mlog_bug_on_msg(o2net_handler_tree_lookup(msg_type, key, &p,
-                                                         &parent) == NULL,
-                               "couldn't find handler we *just* registerd "
-                               "for type %u key %08x\n", msg_type, key);
-       }
-       write_unlock(&o2net_handler_lock);
-       if (ret)
-               goto out;
-
-out:
-       if (ret)
-               kfree(nmh);
-
-       return ret;
-}
-EXPORT_SYMBOL_GPL(o2net_register_handler);
-
-void o2net_unregister_handler_list(struct list_head *list)
-{
-       struct o2net_msg_handler *nmh, *n;
-
-       write_lock(&o2net_handler_lock);
-       list_for_each_entry_safe(nmh, n, list, nh_unregister_item) {
-               mlog(ML_TCP, "unregistering handler func %p type %u key %08x\n",
-                    nmh->nh_func, nmh->nh_msg_type, nmh->nh_key);
-               rb_erase(&nmh->nh_node, &o2net_handler_tree);
-               list_del_init(&nmh->nh_unregister_item);
-               kref_put(&nmh->nh_kref, o2net_handler_kref_release);
-       }
-       write_unlock(&o2net_handler_lock);
-}
-EXPORT_SYMBOL_GPL(o2net_unregister_handler_list);
-
-static struct o2net_msg_handler *o2net_handler_get(u32 msg_type, u32 key)
-{
-       struct o2net_msg_handler *nmh;
-
-       read_lock(&o2net_handler_lock);
-       nmh = o2net_handler_tree_lookup(msg_type, key, NULL, NULL);
-       if (nmh)
-               kref_get(&nmh->nh_kref);
-       read_unlock(&o2net_handler_lock);
-
-       return nmh;
-}
-
-/* ------------------------------------------------------------ */
-
-static int o2net_recv_tcp_msg(struct socket *sock, void *data, size_t len)
-{
-       int ret;
-       mm_segment_t oldfs;
-       struct kvec vec = {
-               .iov_len = len,
-               .iov_base = data,
-       };
-       struct msghdr msg = {
-               .msg_iovlen = 1,
-               .msg_iov = (struct iovec *)&vec,
-                       .msg_flags = MSG_DONTWAIT,
-       };
-
-       oldfs = get_fs();
-       set_fs(get_ds());
-       ret = sock_recvmsg(sock, &msg, len, msg.msg_flags);
-       set_fs(oldfs);
-
-       return ret;
-}
-
-static int o2net_send_tcp_msg(struct socket *sock, struct kvec *vec,
-                             size_t veclen, size_t total)
-{
-       int ret;
-       mm_segment_t oldfs;
-       struct msghdr msg = {
-               .msg_iov = (struct iovec *)vec,
-               .msg_iovlen = veclen,
-       };
-
-       if (sock == NULL) {
-               ret = -EINVAL;
-               goto out;
-       }
-
-       oldfs = get_fs();
-       set_fs(get_ds());
-       ret = sock_sendmsg(sock, &msg, total);
-       set_fs(oldfs);
-       if (ret != total) {
-               mlog(ML_ERROR, "sendmsg returned %d instead of %zu\n", ret,
-                    total);
-               if (ret >= 0)
-                       ret = -EPIPE; /* should be smarter, I bet */
-               goto out;
-       }
-
-       ret = 0;
-out:
-       if (ret < 0)
-               mlog(0, "returning error: %d\n", ret);
-       return ret;
-}
-
-static void o2net_sendpage(struct o2net_sock_container *sc,
-                          void *kmalloced_virt,
-                          size_t size)
-{
-       struct o2net_node *nn = o2net_nn_from_num(sc->sc_node->nd_num);
-       ssize_t ret;
-
-       while (1) {
-               mutex_lock(&sc->sc_send_lock);
-               ret = sc->sc_sock->ops->sendpage(sc->sc_sock,
-                                                virt_to_page(kmalloced_virt),
-                                                (long)kmalloced_virt & ~PAGE_MASK,
-                                                size, MSG_DONTWAIT);
-               mutex_unlock(&sc->sc_send_lock);
-               if (ret == size)
-                       break;
-               if (ret == (ssize_t)-EAGAIN) {
-                       mlog(0, "sendpage of size %zu to " SC_NODEF_FMT
-                            " returned EAGAIN\n", size, SC_NODEF_ARGS(sc));
-                       cond_resched();
-                       continue;
-               }
-               mlog(ML_ERROR, "sendpage of size %zu to " SC_NODEF_FMT
-                    " failed with %zd\n", size, SC_NODEF_ARGS(sc), ret);
-               o2net_ensure_shutdown(nn, sc, 0);
-               break;
-       }
-}
-
-static void o2net_init_msg(struct o2net_msg *msg, u16 data_len, u16 msg_type, u32 key)
-{
-       memset(msg, 0, sizeof(struct o2net_msg));
-       msg->magic = cpu_to_be16(O2NET_MSG_MAGIC);
-       msg->data_len = cpu_to_be16(data_len);
-       msg->msg_type = cpu_to_be16(msg_type);
-       msg->sys_status = cpu_to_be32(O2NET_ERR_NONE);
-       msg->status = 0;
-       msg->key = cpu_to_be32(key);
-}
-
-static int o2net_tx_can_proceed(struct o2net_node *nn,
-                               struct o2net_sock_container **sc_ret,
-                               int *error)
-{
-       int ret = 0;
-
-       spin_lock(&nn->nn_lock);
-       if (nn->nn_persistent_error) {
-               ret = 1;
-               *sc_ret = NULL;
-               *error = nn->nn_persistent_error;
-       } else if (nn->nn_sc_valid) {
-               kref_get(&nn->nn_sc->sc_kref);
-
-               ret = 1;
-               *sc_ret = nn->nn_sc;
-               *error = 0;
-       }
-       spin_unlock(&nn->nn_lock);
-
-       return ret;
-}
-
-/* Get a map of all nodes to which this node is currently connected to */
-void o2net_fill_node_map(unsigned long *map, unsigned bytes)
-{
-       struct o2net_sock_container *sc;
-       int node, ret;
-
-       BUG_ON(bytes < (BITS_TO_LONGS(O2NM_MAX_NODES) * sizeof(unsigned long)));
-
-       memset(map, 0, bytes);
-       for (node = 0; node < O2NM_MAX_NODES; ++node) {
-               o2net_tx_can_proceed(o2net_nn_from_num(node), &sc, &ret);
-               if (!ret) {
-                       set_bit(node, map);
-                       sc_put(sc);
-               }
-       }
-}
-EXPORT_SYMBOL_GPL(o2net_fill_node_map);
-
-int o2net_send_message_vec(u32 msg_type, u32 key, struct kvec *caller_vec,
-                          size_t caller_veclen, u8 target_node, int *status)
-{
-       int ret = 0;
-       struct o2net_msg *msg = NULL;
-       size_t veclen, caller_bytes = 0;
-       struct kvec *vec = NULL;
-       struct o2net_sock_container *sc = NULL;
-       struct o2net_node *nn = o2net_nn_from_num(target_node);
-       struct o2net_status_wait nsw = {
-               .ns_node_item = LIST_HEAD_INIT(nsw.ns_node_item),
-       };
-       struct o2net_send_tracking nst;
-
-#ifdef CONFIG_RAMSTER
-       /* this may be a general bug fix */
-       init_waitqueue_head(&nsw.ns_wq);
-#endif
-
-       o2net_init_nst(&nst, msg_type, key, current, target_node);
-
-       if (o2net_wq == NULL) {
-               mlog(0, "attempt to tx without o2netd running\n");
-               ret = -ESRCH;
-               goto out;
-       }
-
-       if (caller_veclen == 0) {
-               mlog(0, "bad kvec array length\n");
-               ret = -EINVAL;
-               goto out;
-       }
-
-       caller_bytes = iov_length((struct iovec *)caller_vec, caller_veclen);
-       if (caller_bytes > O2NET_MAX_PAYLOAD_BYTES) {
-               mlog(0, "total payload len %zu too large\n", caller_bytes);
-               ret = -EINVAL;
-               goto out;
-       }
-
-       if (target_node == o2nm_this_node()) {
-               ret = -ELOOP;
-               goto out;
-       }
-
-       o2net_debug_add_nst(&nst);
-
-       o2net_set_nst_sock_time(&nst);
-
-       wait_event(nn->nn_sc_wq, o2net_tx_can_proceed(nn, &sc, &ret));
-       if (ret)
-               goto out;
-
-       o2net_set_nst_sock_container(&nst, sc);
-
-       veclen = caller_veclen + 1;
-       vec = kmalloc(sizeof(struct kvec) * veclen, GFP_ATOMIC);
-       if (vec == NULL) {
-               mlog(0, "failed to %zu element kvec!\n", veclen);
-               ret = -ENOMEM;
-               goto out;
-       }
-
-       msg = kmalloc(sizeof(struct o2net_msg), GFP_ATOMIC);
-       if (!msg) {
-               mlog(0, "failed to allocate a o2net_msg!\n");
-               ret = -ENOMEM;
-               goto out;
-       }
-
-       o2net_init_msg(msg, caller_bytes, msg_type, key);
-
-       vec[0].iov_len = sizeof(struct o2net_msg);
-       vec[0].iov_base = msg;
-       memcpy(&vec[1], caller_vec, caller_veclen * sizeof(struct kvec));
-
-       ret = o2net_prep_nsw(nn, &nsw);
-       if (ret)
-               goto out;
-
-       msg->msg_num = cpu_to_be32(nsw.ns_id);
-       o2net_set_nst_msg_id(&nst, nsw.ns_id);
-
-       o2net_set_nst_send_time(&nst);
-
-       /* finally, convert the message header to network byte-order
-        * and send */
-       mutex_lock(&sc->sc_send_lock);
-       ret = o2net_send_tcp_msg(sc->sc_sock, vec, veclen,
-                                sizeof(struct o2net_msg) + caller_bytes);
-       mutex_unlock(&sc->sc_send_lock);
-       msglog(msg, "sending returned %d\n", ret);
-       if (ret < 0) {
-               mlog(0, "error returned from o2net_send_tcp_msg=%d\n", ret);
-               goto out;
-       }
-
-       /* wait on other node's handler */
-       o2net_set_nst_status_time(&nst);
-       wait_event(nsw.ns_wq, o2net_nsw_completed(nn, &nsw));
-
-       o2net_update_send_stats(&nst, sc);
-
-       /* Note that we avoid overwriting the callers status return
-        * variable if a system error was reported on the other
-        * side. Callers beware. */
-       ret = o2net_sys_err_to_errno(nsw.ns_sys_status);
-       if (status && !ret)
-               *status = nsw.ns_status;
-
-       mlog(0, "woken, returning system status %d, user status %d\n",
-            ret, nsw.ns_status);
-out:
-       o2net_debug_del_nst(&nst); /* must be before dropping sc and node */
-       if (sc)
-               sc_put(sc);
-       if (vec)
-               kfree(vec);
-       if (msg)
-               kfree(msg);
-       o2net_complete_nsw(nn, &nsw, 0, 0, 0);
-       return ret;
-}
-EXPORT_SYMBOL_GPL(o2net_send_message_vec);
-
-int o2net_send_message(u32 msg_type, u32 key, void *data, u32 len,
-                      u8 target_node, int *status)
-{
-       struct kvec vec = {
-               .iov_base = data,
-               .iov_len = len,
-       };
-       return o2net_send_message_vec(msg_type, key, &vec, 1,
-                                     target_node, status);
-}
-EXPORT_SYMBOL_GPL(o2net_send_message);
-
-static int o2net_send_status_magic(struct socket *sock, struct o2net_msg *hdr,
-                                  enum o2net_system_error syserr, int err)
-{
-       struct kvec vec = {
-               .iov_base = hdr,
-               .iov_len = sizeof(struct o2net_msg),
-       };
-
-       BUG_ON(syserr >= O2NET_ERR_MAX);
-
-       /* leave other fields intact from the incoming message, msg_num
-        * in particular */
-       hdr->sys_status = cpu_to_be32(syserr);
-       hdr->status = cpu_to_be32(err);
-       hdr->magic = cpu_to_be16(O2NET_MSG_STATUS_MAGIC);  // twiddle the magic
-       hdr->data_len = 0;
-
-       msglog(hdr, "about to send status magic %d\n", err);
-       /* hdr has been in host byteorder this whole time */
-       return o2net_send_tcp_msg(sock, &vec, 1, sizeof(struct o2net_msg));
-}
-
-#ifdef CONFIG_RAMSTER
-/*
- * "data magic" is a long version of "status magic" where the message
- * payload actually contains data to be passed in reply to certain messages
- */
-static int o2net_send_data_magic(struct o2net_sock_container *sc,
-                         struct o2net_msg *hdr,
-                         void *data, size_t data_len,
-                         enum o2net_system_error syserr, int err)
-{
-       struct kvec vec[2];
-       int ret;
-
-       vec[0].iov_base = hdr;
-       vec[0].iov_len = sizeof(struct o2net_msg);
-       vec[1].iov_base = data;
-       vec[1].iov_len = data_len;
-
-       BUG_ON(syserr >= O2NET_ERR_MAX);
-
-       /* leave other fields intact from the incoming message, msg_num
-        * in particular */
-       hdr->sys_status = cpu_to_be32(syserr);
-       hdr->status = cpu_to_be32(err);
-       hdr->magic = cpu_to_be16(O2NET_MSG_DATA_MAGIC);  /* twiddle magic */
-       hdr->data_len = cpu_to_be16(data_len);
-
-       msglog(hdr, "about to send data magic %d\n", err);
-       /* hdr has been in host byteorder this whole time */
-       ret = o2net_send_tcp_msg(sc->sc_sock, vec, 2,
-                       sizeof(struct o2net_msg) + data_len);
-       return ret;
-}
-
-/*
- * called by a message handler to convert an otherwise normal reply
- * message into a "data magic" message
- */
-void o2net_force_data_magic(struct o2net_msg *hdr, u16 msgtype, u32 msgkey)
-{
-       hdr->magic = cpu_to_be16(O2NET_MSG_DATA_MAGIC);
-       hdr->msg_type = cpu_to_be16(msgtype);
-       hdr->key = cpu_to_be32(msgkey);
-}
-#endif
-
-/* this returns -errno if the header was unknown or too large, etc.
- * after this is called the buffer us reused for the next message */
-static int o2net_process_message(struct o2net_sock_container *sc,
-                                struct o2net_msg *hdr)
-{
-       struct o2net_node *nn = o2net_nn_from_num(sc->sc_node->nd_num);
-       int ret = 0, handler_status;
-       enum  o2net_system_error syserr;
-       struct o2net_msg_handler *nmh = NULL;
-       void *ret_data = NULL;
-#ifdef CONFIG_RAMSTER
-       int data_magic = 0;
-#endif
-
-       msglog(hdr, "processing message\n");
-
-       o2net_sc_postpone_idle(sc);
-
-       switch(be16_to_cpu(hdr->magic)) {
-               case O2NET_MSG_STATUS_MAGIC:
-                       /* special type for returning message status */
-                       o2net_complete_nsw(nn, NULL,
-                                          be32_to_cpu(hdr->msg_num),
-                                          be32_to_cpu(hdr->sys_status),
-                                          be32_to_cpu(hdr->status));
-                       goto out;
-               case O2NET_MSG_KEEP_REQ_MAGIC:
-                       o2net_sendpage(sc, o2net_keep_resp,
-                                      sizeof(*o2net_keep_resp));
-                       goto out;
-               case O2NET_MSG_KEEP_RESP_MAGIC:
-                       goto out;
-               case O2NET_MSG_MAGIC:
-                       break;
-#ifdef CONFIG_RAMSTER
-               case O2NET_MSG_DATA_MAGIC:
-                       /*
-                        * unlike a normal status magic, a data magic DOES
-                        * (MUST) have a handler, so the control flow is
-                        * a little funky here as a result
-                        */
-                       data_magic = 1;
-                       break;
-#endif
-               default:
-                       msglog(hdr, "bad magic\n");
-                       ret = -EINVAL;
-                       goto out;
-                       break;
-       }
-
-       /* find a handler for it */
-       handler_status = 0;
-       nmh = o2net_handler_get(be16_to_cpu(hdr->msg_type),
-                               be32_to_cpu(hdr->key));
-       if (!nmh) {
-               mlog(ML_TCP, "couldn't find handler for type %u key %08x\n",
-                    be16_to_cpu(hdr->msg_type), be32_to_cpu(hdr->key));
-               syserr = O2NET_ERR_NO_HNDLR;
-               goto out_respond;
-       }
-
-       syserr = O2NET_ERR_NONE;
-
-       if (be16_to_cpu(hdr->data_len) > nmh->nh_max_len)
-               syserr = O2NET_ERR_OVERFLOW;
-
-       if (syserr != O2NET_ERR_NONE)
-               goto out_respond;
-
-       o2net_set_func_start_time(sc);
-       sc->sc_msg_key = be32_to_cpu(hdr->key);
-       sc->sc_msg_type = be16_to_cpu(hdr->msg_type);
-       handler_status = (nmh->nh_func)(hdr, sizeof(struct o2net_msg) +
-                                            be16_to_cpu(hdr->data_len),
-                                       nmh->nh_func_data, &ret_data);
-#ifdef CONFIG_RAMSTER
-       if (data_magic) {
-               /*
-                * handler handled data sent in reply to request
-                * so complete the transaction
-                */
-               o2net_complete_nsw(nn, NULL, be32_to_cpu(hdr->msg_num),
-                       be32_to_cpu(hdr->sys_status), handler_status);
-               goto out;
-       }
-       /*
-        * handler changed magic to DATA_MAGIC to reply to request for data,
-        * implies ret_data points to data to return and handler_status
-        * is the number of bytes of data
-        */
-       if (be16_to_cpu(hdr->magic) == O2NET_MSG_DATA_MAGIC) {
-               ret = o2net_send_data_magic(sc, hdr,
-                                               ret_data, handler_status,
-                                               syserr, 0);
-               hdr = NULL;
-               mlog(0, "sending data reply %d, syserr %d returned %d\n",
-                       handler_status, syserr, ret);
-               o2net_set_func_stop_time(sc);
-
-               o2net_update_recv_stats(sc);
-               goto out;
-       }
-#endif
-       o2net_set_func_stop_time(sc);
-
-       o2net_update_recv_stats(sc);
-
-out_respond:
-       /* this destroys the hdr, so don't use it after this */
-       mutex_lock(&sc->sc_send_lock);
-       ret = o2net_send_status_magic(sc->sc_sock, hdr, syserr,
-                                     handler_status);
-       mutex_unlock(&sc->sc_send_lock);
-       hdr = NULL;
-       mlog(0, "sending handler status %d, syserr %d returned %d\n",
-            handler_status, syserr, ret);
-
-       if (nmh) {
-               BUG_ON(ret_data != NULL && nmh->nh_post_func == NULL);
-               if (nmh->nh_post_func)
-                       (nmh->nh_post_func)(handler_status, nmh->nh_func_data,
-                                           ret_data);
-       }
-
-out:
-       if (nmh)
-               o2net_handler_put(nmh);
-       return ret;
-}
-
-static int o2net_check_handshake(struct o2net_sock_container *sc)
-{
-       struct o2net_handshake *hand = page_address(sc->sc_page);
-       struct o2net_node *nn = o2net_nn_from_num(sc->sc_node->nd_num);
-
-       if (hand->protocol_version != cpu_to_be64(O2NET_PROTOCOL_VERSION)) {
-               printk(KERN_NOTICE "o2net: " SC_NODEF_FMT " Advertised net "
-                      "protocol version %llu but %llu is required. "
-                      "Disconnecting.\n", SC_NODEF_ARGS(sc),
-                      (unsigned long long)be64_to_cpu(hand->protocol_version),
-                      O2NET_PROTOCOL_VERSION);
-
-               /* don't bother reconnecting if its the wrong version. */
-               o2net_ensure_shutdown(nn, sc, -ENOTCONN);
-               return -1;
-       }
-
-       /*
-        * Ensure timeouts are consistent with other nodes, otherwise
-        * we can end up with one node thinking that the other must be down,
-        * but isn't. This can ultimately cause corruption.
-        */
-       if (be32_to_cpu(hand->o2net_idle_timeout_ms) !=
-                               o2net_idle_timeout()) {
-               printk(KERN_NOTICE "o2net: " SC_NODEF_FMT " uses a network "
-                      "idle timeout of %u ms, but we use %u ms locally. "
-                      "Disconnecting.\n", SC_NODEF_ARGS(sc),
-                      be32_to_cpu(hand->o2net_idle_timeout_ms),
-                      o2net_idle_timeout());
-               o2net_ensure_shutdown(nn, sc, -ENOTCONN);
-               return -1;
-       }
-
-       if (be32_to_cpu(hand->o2net_keepalive_delay_ms) !=
-                       o2net_keepalive_delay()) {
-               printk(KERN_NOTICE "o2net: " SC_NODEF_FMT " uses a keepalive "
-                      "delay of %u ms, but we use %u ms locally. "
-                      "Disconnecting.\n", SC_NODEF_ARGS(sc),
-                      be32_to_cpu(hand->o2net_keepalive_delay_ms),
-                      o2net_keepalive_delay());
-               o2net_ensure_shutdown(nn, sc, -ENOTCONN);
-               return -1;
-       }
-
-       if (be32_to_cpu(hand->o2hb_heartbeat_timeout_ms) !=
-                       O2HB_MAX_WRITE_TIMEOUT_MS) {
-               printk(KERN_NOTICE "o2net: " SC_NODEF_FMT " uses a heartbeat "
-                      "timeout of %u ms, but we use %u ms locally. "
-                      "Disconnecting.\n", SC_NODEF_ARGS(sc),
-                      be32_to_cpu(hand->o2hb_heartbeat_timeout_ms),
-                      O2HB_MAX_WRITE_TIMEOUT_MS);
-               o2net_ensure_shutdown(nn, sc, -ENOTCONN);
-               return -1;
-       }
-
-       sc->sc_handshake_ok = 1;
-
-       spin_lock(&nn->nn_lock);
-       /* set valid and queue the idle timers only if it hasn't been
-        * shut down already */
-       if (nn->nn_sc == sc) {
-               o2net_sc_reset_idle_timer(sc);
-               atomic_set(&nn->nn_timeout, 0);
-               o2net_set_nn_state(nn, sc, 1, 0);
-       }
-       spin_unlock(&nn->nn_lock);
-
-       /* shift everything up as though it wasn't there */
-       sc->sc_page_off -= sizeof(struct o2net_handshake);
-       if (sc->sc_page_off)
-               memmove(hand, hand + 1, sc->sc_page_off);
-
-       return 0;
-}
-
-/* this demuxes the queued rx bytes into header or payload bits and calls
- * handlers as each full message is read off the socket.  it returns -error,
- * == 0 eof, or > 0 for progress made.*/
-static int o2net_advance_rx(struct o2net_sock_container *sc)
-{
-       struct o2net_msg *hdr;
-       int ret = 0;
-       void *data;
-       size_t datalen;
-
-       sclog(sc, "receiving\n");
-       o2net_set_advance_start_time(sc);
-
-       if (unlikely(sc->sc_handshake_ok == 0)) {
-               if(sc->sc_page_off < sizeof(struct o2net_handshake)) {
-                       data = page_address(sc->sc_page) + sc->sc_page_off;
-                       datalen = sizeof(struct o2net_handshake) - sc->sc_page_off;
-                       ret = o2net_recv_tcp_msg(sc->sc_sock, data, datalen);
-                       if (ret > 0)
-                               sc->sc_page_off += ret;
-               }
-
-               if (sc->sc_page_off == sizeof(struct o2net_handshake)) {
-                       o2net_check_handshake(sc);
-                       if (unlikely(sc->sc_handshake_ok == 0))
-                               ret = -EPROTO;
-               }
-               goto out;
-       }
-
-       /* do we need more header? */
-       if (sc->sc_page_off < sizeof(struct o2net_msg)) {
-               data = page_address(sc->sc_page) + sc->sc_page_off;
-               datalen = sizeof(struct o2net_msg) - sc->sc_page_off;
-               ret = o2net_recv_tcp_msg(sc->sc_sock, data, datalen);
-               if (ret > 0) {
-                       sc->sc_page_off += ret;
-                       /* only swab incoming here.. we can
-                        * only get here once as we cross from
-                        * being under to over */
-                       if (sc->sc_page_off == sizeof(struct o2net_msg)) {
-                               hdr = page_address(sc->sc_page);
-                               if (be16_to_cpu(hdr->data_len) >
-                                   O2NET_MAX_PAYLOAD_BYTES)
-                                       ret = -EOVERFLOW;
-                       }
-               }
-               if (ret <= 0)
-                       goto out;
-       }
-
-       if (sc->sc_page_off < sizeof(struct o2net_msg)) {
-               /* oof, still don't have a header */
-               goto out;
-       }
-
-       /* this was swabbed above when we first read it */
-       hdr = page_address(sc->sc_page);
-
-       msglog(hdr, "at page_off %zu\n", sc->sc_page_off);
-
-       /* do we need more payload? */
-       if (sc->sc_page_off - sizeof(struct o2net_msg) < be16_to_cpu(hdr->data_len)) {
-               /* need more payload */
-               data = page_address(sc->sc_page) + sc->sc_page_off;
-               datalen = (sizeof(struct o2net_msg) + be16_to_cpu(hdr->data_len)) -
-                         sc->sc_page_off;
-               ret = o2net_recv_tcp_msg(sc->sc_sock, data, datalen);
-               if (ret > 0)
-                       sc->sc_page_off += ret;
-               if (ret <= 0)
-                       goto out;
-       }
-
-       if (sc->sc_page_off - sizeof(struct o2net_msg) == be16_to_cpu(hdr->data_len)) {
-               /* we can only get here once, the first time we read
-                * the payload.. so set ret to progress if the handler
-                * works out. after calling this the message is toast */
-               ret = o2net_process_message(sc, hdr);
-               if (ret == 0)
-                       ret = 1;
-               sc->sc_page_off = 0;
-       }
-
-out:
-       sclog(sc, "ret = %d\n", ret);
-       o2net_set_advance_stop_time(sc);
-       return ret;
-}
-
-/* this work func is triggerd by data ready.  it reads until it can read no
- * more.  it interprets 0, eof, as fatal.  if data_ready hits while we're doing
- * our work the work struct will be marked and we'll be called again. */
-static void o2net_rx_until_empty(struct work_struct *work)
-{
-       struct o2net_sock_container *sc =
-               container_of(work, struct o2net_sock_container, sc_rx_work);
-       int ret;
-
-       do {
-               ret = o2net_advance_rx(sc);
-       } while (ret > 0);
-
-       if (ret <= 0 && ret != -EAGAIN) {
-               struct o2net_node *nn = o2net_nn_from_num(sc->sc_node->nd_num);
-               sclog(sc, "saw error %d, closing\n", ret);
-               /* not permanent so read failed handshake can retry */
-               o2net_ensure_shutdown(nn, sc, 0);
-       }
-
-       sc_put(sc);
-}
-
-static int o2net_set_nodelay(struct socket *sock)
-{
-       int ret, val = 1;
-       mm_segment_t oldfs;
-
-       oldfs = get_fs();
-       set_fs(KERNEL_DS);
-
-       /*
-        * Dear unsuspecting programmer,
-        *
-        * Don't use sock_setsockopt() for SOL_TCP.  It doesn't check its level
-        * argument and assumes SOL_SOCKET so, say, your TCP_NODELAY will
-        * silently turn into SO_DEBUG.
-        *
-        * Yours,
-        * Keeper of hilariously fragile interfaces.
-        */
-       ret = sock->ops->setsockopt(sock, SOL_TCP, TCP_NODELAY,
-                                   (char __user *)&val, sizeof(val));
-
-       set_fs(oldfs);
-       return ret;
-}
-
-static void o2net_initialize_handshake(void)
-{
-       o2net_hand->o2hb_heartbeat_timeout_ms = cpu_to_be32(
-               O2HB_MAX_WRITE_TIMEOUT_MS);
-       o2net_hand->o2net_idle_timeout_ms = cpu_to_be32(o2net_idle_timeout());
-       o2net_hand->o2net_keepalive_delay_ms = cpu_to_be32(
-               o2net_keepalive_delay());
-       o2net_hand->o2net_reconnect_delay_ms = cpu_to_be32(
-               o2net_reconnect_delay());
-}
-
-/* ------------------------------------------------------------ */
-
-/* called when a connect completes and after a sock is accepted.  the
- * rx path will see the response and mark the sc valid */
-static void o2net_sc_connect_completed(struct work_struct *work)
-{
-       struct o2net_sock_container *sc =
-               container_of(work, struct o2net_sock_container,
-                            sc_connect_work);
-
-       mlog(ML_MSG, "sc sending handshake with ver %llu id %llx\n",
-              (unsigned long long)O2NET_PROTOCOL_VERSION,
-             (unsigned long long)be64_to_cpu(o2net_hand->connector_id));
-
-       o2net_initialize_handshake();
-       o2net_sendpage(sc, o2net_hand, sizeof(*o2net_hand));
-       sc_put(sc);
-}
-
-/* this is called as a work_struct func. */
-static void o2net_sc_send_keep_req(struct work_struct *work)
-{
-       struct o2net_sock_container *sc =
-               container_of(work, struct o2net_sock_container,
-                            sc_keepalive_work.work);
-
-       o2net_sendpage(sc, o2net_keep_req, sizeof(*o2net_keep_req));
-       sc_put(sc);
-}
-
-/* socket shutdown does a del_timer_sync against this as it tears down.
- * we can't start this timer until we've got to the point in sc buildup
- * where shutdown is going to be involved */
-static void o2net_idle_timer(unsigned long data)
-{
-       struct o2net_sock_container *sc = (struct o2net_sock_container *)data;
-#ifndef CONFIG_RAMSTER
-       struct o2net_node *nn = o2net_nn_from_num(sc->sc_node->nd_num);
-#endif
-#ifdef CONFIG_DEBUG_FS
-       unsigned long msecs = ktime_to_ms(ktime_get()) -
-               ktime_to_ms(sc->sc_tv_timer);
-#else
-       unsigned long msecs = o2net_idle_timeout();
-#endif
-
-       printk(KERN_NOTICE "o2net: Connection to " SC_NODEF_FMT " has been "
-              "idle for %lu.%lu secs, shutting it down.\n", SC_NODEF_ARGS(sc),
-              msecs / 1000, msecs % 1000);
-
-       /*
-        * Initialize the nn_timeout so that the next connection attempt
-        * will continue in o2net_start_connect.
-        */
-#ifdef CONFIG_RAMSTER
-       /* Avoid spurious shutdowns... not sure if this is still necessary */
-       pr_err("o2net_idle_timer, skipping shutdown work\n");
-#else
-       atomic_set(&nn->nn_timeout, 1);
-
-       o2net_sc_queue_work(sc, &sc->sc_shutdown_work);
-#endif
-}
-
-static void o2net_sc_reset_idle_timer(struct o2net_sock_container *sc)
-{
-       o2net_sc_cancel_delayed_work(sc, &sc->sc_keepalive_work);
-       o2net_sc_queue_delayed_work(sc, &sc->sc_keepalive_work,
-                     msecs_to_jiffies(o2net_keepalive_delay()));
-       o2net_set_sock_timer(sc);
-       mod_timer(&sc->sc_idle_timeout,
-              jiffies + msecs_to_jiffies(o2net_idle_timeout()));
-}
-
-static void o2net_sc_postpone_idle(struct o2net_sock_container *sc)
-{
-       /* Only push out an existing timer */
-       if (timer_pending(&sc->sc_idle_timeout))
-               o2net_sc_reset_idle_timer(sc);
-}
-
-/* this work func is kicked whenever a path sets the nn state which doesn't
- * have valid set.  This includes seeing hb come up, losing a connection,
- * having a connect attempt fail, etc. This centralizes the logic which decides
- * if a connect attempt should be made or if we should give up and all future
- * transmit attempts should fail */
-static void o2net_start_connect(struct work_struct *work)
-{
-       struct o2net_node *nn =
-               container_of(work, struct o2net_node, nn_connect_work.work);
-       struct o2net_sock_container *sc = NULL;
-       struct o2nm_node *node = NULL, *mynode = NULL;
-       struct socket *sock = NULL;
-       struct sockaddr_in myaddr = {0, }, remoteaddr = {0, };
-       int ret = 0, stop;
-       unsigned int timeout;
-
-       /* if we're greater we initiate tx, otherwise we accept */
-       if (o2nm_this_node() <= o2net_num_from_nn(nn))
-               goto out;
-
-       /* watch for racing with tearing a node down */
-       node = o2nm_get_node_by_num(o2net_num_from_nn(nn));
-       if (node == NULL) {
-               ret = 0;
-               goto out;
-       }
-
-       mynode = o2nm_get_node_by_num(o2nm_this_node());
-       if (mynode == NULL) {
-               ret = 0;
-               goto out;
-       }
-
-       spin_lock(&nn->nn_lock);
-       /*
-        * see if we already have one pending or have given up.
-        * For nn_timeout, it is set when we close the connection
-        * because of the idle time out. So it means that we have
-        * at least connected to that node successfully once,
-        * now try to connect to it again.
-        */
-       timeout = atomic_read(&nn->nn_timeout);
-       stop = (nn->nn_sc ||
-               (nn->nn_persistent_error &&
-               (nn->nn_persistent_error != -ENOTCONN || timeout == 0)));
-       spin_unlock(&nn->nn_lock);
-       if (stop)
-               goto out;
-
-       nn->nn_last_connect_attempt = jiffies;
-
-       sc = sc_alloc(node);
-       if (sc == NULL) {
-               mlog(0, "couldn't allocate sc\n");
-               ret = -ENOMEM;
-               goto out;
-       }
-
-       ret = sock_create(PF_INET, SOCK_STREAM, IPPROTO_TCP, &sock);
-       if (ret < 0) {
-               mlog(0, "can't create socket: %d\n", ret);
-               goto out;
-       }
-       sc->sc_sock = sock; /* freed by sc_kref_release */
-
-       sock->sk->sk_allocation = GFP_ATOMIC;
-
-       myaddr.sin_family = AF_INET;
-       myaddr.sin_addr.s_addr = mynode->nd_ipv4_address;
-       myaddr.sin_port = htons(0); /* any port */
-
-       ret = sock->ops->bind(sock, (struct sockaddr *)&myaddr,
-                             sizeof(myaddr));
-       if (ret) {
-               mlog(ML_ERROR, "bind failed with %d at address %pI4\n",
-                    ret, &mynode->nd_ipv4_address);
-               goto out;
-       }
-
-       ret = o2net_set_nodelay(sc->sc_sock);
-       if (ret) {
-               mlog(ML_ERROR, "setting TCP_NODELAY failed with %d\n", ret);
-               goto out;
-       }
-
-       o2net_register_callbacks(sc->sc_sock->sk, sc);
-
-       spin_lock(&nn->nn_lock);
-       /* handshake completion will set nn->nn_sc_valid */
-       o2net_set_nn_state(nn, sc, 0, 0);
-       spin_unlock(&nn->nn_lock);
-
-       remoteaddr.sin_family = AF_INET;
-       remoteaddr.sin_addr.s_addr = node->nd_ipv4_address;
-       remoteaddr.sin_port = node->nd_ipv4_port;
-
-       ret = sc->sc_sock->ops->connect(sc->sc_sock,
-                                       (struct sockaddr *)&remoteaddr,
-                                       sizeof(remoteaddr),
-                                       O_NONBLOCK);
-       if (ret == -EINPROGRESS)
-               ret = 0;
-
-out:
-       if (ret) {
-               printk(KERN_NOTICE "o2net: Connect attempt to " SC_NODEF_FMT
-                      " failed with errno %d\n", SC_NODEF_ARGS(sc), ret);
-               /* 0 err so that another will be queued and attempted
-                * from set_nn_state */
-               if (sc)
-                       o2net_ensure_shutdown(nn, sc, 0);
-       }
-       if (sc)
-               sc_put(sc);
-       if (node)
-               o2nm_node_put(node);
-       if (mynode)
-               o2nm_node_put(mynode);
-
-       return;
-}
-
-static void o2net_connect_expired(struct work_struct *work)
-{
-       struct o2net_node *nn =
-               container_of(work, struct o2net_node, nn_connect_expired.work);
-
-       spin_lock(&nn->nn_lock);
-       if (!nn->nn_sc_valid) {
-               printk(KERN_NOTICE "o2net: No connection established with "
-                      "node %u after %u.%u seconds, giving up.\n",
-                    o2net_num_from_nn(nn),
-                    o2net_idle_timeout() / 1000,
-                    o2net_idle_timeout() % 1000);
-
-               o2net_set_nn_state(nn, NULL, 0, -ENOTCONN);
-       }
-       spin_unlock(&nn->nn_lock);
-}
-
-static void o2net_still_up(struct work_struct *work)
-{
-       struct o2net_node *nn =
-               container_of(work, struct o2net_node, nn_still_up.work);
-
-       o2quo_hb_still_up(o2net_num_from_nn(nn));
-}
-
-/* ------------------------------------------------------------ */
-
-void o2net_disconnect_node(struct o2nm_node *node)
-{
-       struct o2net_node *nn = o2net_nn_from_num(node->nd_num);
-
-       /* don't reconnect until it's heartbeating again */
-       spin_lock(&nn->nn_lock);
-       atomic_set(&nn->nn_timeout, 0);
-       o2net_set_nn_state(nn, NULL, 0, -ENOTCONN);
-       spin_unlock(&nn->nn_lock);
-
-       if (o2net_wq) {
-               cancel_delayed_work(&nn->nn_connect_expired);
-               cancel_delayed_work(&nn->nn_connect_work);
-               cancel_delayed_work(&nn->nn_still_up);
-               flush_workqueue(o2net_wq);
-       }
-}
-
-static void o2net_hb_node_down_cb(struct o2nm_node *node, int node_num,
-                                 void *data)
-{
-       o2quo_hb_down(node_num);
-
-       if (!node)
-               return;
-
-       if (node_num != o2nm_this_node())
-               o2net_disconnect_node(node);
-
-       BUG_ON(atomic_read(&o2net_connected_peers) < 0);
-}
-
-static void o2net_hb_node_up_cb(struct o2nm_node *node, int node_num,
-                               void *data)
-{
-       struct o2net_node *nn = o2net_nn_from_num(node_num);
-
-       o2quo_hb_up(node_num);
-
-       BUG_ON(!node);
-
-       /* ensure an immediate connect attempt */
-       nn->nn_last_connect_attempt = jiffies -
-               (msecs_to_jiffies(o2net_reconnect_delay()) + 1);
-
-       if (node_num != o2nm_this_node()) {
-               /* believe it or not, accept and node hearbeating testing
-                * can succeed for this node before we got here.. so
-                * only use set_nn_state to clear the persistent error
-                * if that hasn't already happened */
-               spin_lock(&nn->nn_lock);
-               atomic_set(&nn->nn_timeout, 0);
-               if (nn->nn_persistent_error)
-                       o2net_set_nn_state(nn, NULL, 0, 0);
-               spin_unlock(&nn->nn_lock);
-       }
-}
-
-void o2net_unregister_hb_callbacks(void)
-{
-       o2hb_unregister_callback(NULL, &o2net_hb_up);
-       o2hb_unregister_callback(NULL, &o2net_hb_down);
-}
-
-int o2net_register_hb_callbacks(void)
-{
-       int ret;
-
-       o2hb_setup_callback(&o2net_hb_down, O2HB_NODE_DOWN_CB,
-                           o2net_hb_node_down_cb, NULL, O2NET_HB_PRI);
-       o2hb_setup_callback(&o2net_hb_up, O2HB_NODE_UP_CB,
-                           o2net_hb_node_up_cb, NULL, O2NET_HB_PRI);
-
-       ret = o2hb_register_callback(NULL, &o2net_hb_up);
-       if (ret == 0)
-               ret = o2hb_register_callback(NULL, &o2net_hb_down);
-
-       if (ret)
-               o2net_unregister_hb_callbacks();
-
-       return ret;
-}
-
-/* ------------------------------------------------------------ */
-
-static int o2net_accept_one(struct socket *sock)
-{
-       int ret, slen;
-       struct sockaddr_in sin;
-       struct socket *new_sock = NULL;
-       struct o2nm_node *node = NULL;
-       struct o2nm_node *local_node = NULL;
-       struct o2net_sock_container *sc = NULL;
-       struct o2net_node *nn;
-
-       BUG_ON(sock == NULL);
-       ret = sock_create_lite(sock->sk->sk_family, sock->sk->sk_type,
-                              sock->sk->sk_protocol, &new_sock);
-       if (ret)
-               goto out;
-
-       new_sock->type = sock->type;
-       new_sock->ops = sock->ops;
-       ret = sock->ops->accept(sock, new_sock, O_NONBLOCK);
-       if (ret < 0)
-               goto out;
-
-       new_sock->sk->sk_allocation = GFP_ATOMIC;
-
-       ret = o2net_set_nodelay(new_sock);
-       if (ret) {
-               mlog(ML_ERROR, "setting TCP_NODELAY failed with %d\n", ret);
-               goto out;
-       }
-
-       slen = sizeof(sin);
-       ret = new_sock->ops->getname(new_sock, (struct sockaddr *) &sin,
-                                      &slen, 1);
-       if (ret < 0)
-               goto out;
-
-       node = o2nm_get_node_by_ip(sin.sin_addr.s_addr);
-       if (node == NULL) {
-               printk(KERN_NOTICE "o2net: Attempt to connect from unknown "
-                      "node at %pI4:%d\n", &sin.sin_addr.s_addr,
-                      ntohs(sin.sin_port));
-               ret = -EINVAL;
-               goto out;
-       }
-
-       if (o2nm_this_node() >= node->nd_num) {
-               local_node = o2nm_get_node_by_num(o2nm_this_node());
-               printk(KERN_NOTICE "o2net: Unexpected connect attempt seen "
-                      "at node '%s' (%u, %pI4:%d) from node '%s' (%u, "
-                      "%pI4:%d)\n", local_node->nd_name, local_node->nd_num,
-                      &(local_node->nd_ipv4_address),
-                      ntohs(local_node->nd_ipv4_port), node->nd_name,
-                      node->nd_num, &sin.sin_addr.s_addr, ntohs(sin.sin_port));
-               ret = -EINVAL;
-               goto out;
-       }
-
-       /* this happens all the time when the other node sees our heartbeat
-        * and tries to connect before we see their heartbeat */
-       if (!o2hb_check_node_heartbeating_from_callback(node->nd_num)) {
-               mlog(ML_CONN, "attempt to connect from node '%s' at "
-                    "%pI4:%d but it isn't heartbeating\n",
-                    node->nd_name, &sin.sin_addr.s_addr,
-                    ntohs(sin.sin_port));
-               ret = -EINVAL;
-               goto out;
-       }
-
-       nn = o2net_nn_from_num(node->nd_num);
-
-       spin_lock(&nn->nn_lock);
-       if (nn->nn_sc)
-               ret = -EBUSY;
-       else
-               ret = 0;
-       spin_unlock(&nn->nn_lock);
-       if (ret) {
-               printk(KERN_NOTICE "o2net: Attempt to connect from node '%s' "
-                      "at %pI4:%d but it already has an open connection\n",
-                      node->nd_name, &sin.sin_addr.s_addr,
-                      ntohs(sin.sin_port));
-               goto out;
-       }
-
-       sc = sc_alloc(node);
-       if (sc == NULL) {
-               ret = -ENOMEM;
-               goto out;
-       }
-
-       sc->sc_sock = new_sock;
-       new_sock = NULL;
-
-       spin_lock(&nn->nn_lock);
-       atomic_set(&nn->nn_timeout, 0);
-       o2net_set_nn_state(nn, sc, 0, 0);
-       spin_unlock(&nn->nn_lock);
-
-       o2net_register_callbacks(sc->sc_sock->sk, sc);
-       o2net_sc_queue_work(sc, &sc->sc_rx_work);
-
-       o2net_initialize_handshake();
-       o2net_sendpage(sc, o2net_hand, sizeof(*o2net_hand));
-
-out:
-       if (new_sock)
-               sock_release(new_sock);
-       if (node)
-               o2nm_node_put(node);
-       if (local_node)
-               o2nm_node_put(local_node);
-       if (sc)
-               sc_put(sc);
-       return ret;
-}
-
-static void o2net_accept_many(struct work_struct *work)
-{
-       struct socket *sock = o2net_listen_sock;
-       while (o2net_accept_one(sock) == 0)
-               cond_resched();
-}
-
-static void o2net_listen_data_ready(struct sock *sk, int bytes)
-{
-       void (*ready)(struct sock *sk, int bytes);
-
-       read_lock(&sk->sk_callback_lock);
-       ready = sk->sk_user_data;
-       if (ready == NULL) { /* check for teardown race */
-               ready = sk->sk_data_ready;
-               goto out;
-       }
-
-       /* ->sk_data_ready is also called for a newly established child socket
-        * before it has been accepted and the acceptor has set up their
-        * data_ready.. we only want to queue listen work for our listening
-        * socket */
-       if (sk->sk_state == TCP_LISTEN) {
-               mlog(ML_TCP, "bytes: %d\n", bytes);
-               queue_work(o2net_wq, &o2net_listen_work);
-       }
-
-out:
-       read_unlock(&sk->sk_callback_lock);
-       ready(sk, bytes);
-}
-
-static int o2net_open_listening_sock(__be32 addr, __be16 port)
-{
-       struct socket *sock = NULL;
-       int ret;
-       struct sockaddr_in sin = {
-               .sin_family = PF_INET,
-               .sin_addr = { .s_addr = addr },
-               .sin_port = port,
-       };
-
-       ret = sock_create(PF_INET, SOCK_STREAM, IPPROTO_TCP, &sock);
-       if (ret < 0) {
-               printk(KERN_ERR "o2net: Error %d while creating socket\n", ret);
-               goto out;
-       }
-
-       sock->sk->sk_allocation = GFP_ATOMIC;
-
-       write_lock_bh(&sock->sk->sk_callback_lock);
-       sock->sk->sk_user_data = sock->sk->sk_data_ready;
-       sock->sk->sk_data_ready = o2net_listen_data_ready;
-       write_unlock_bh(&sock->sk->sk_callback_lock);
-
-       o2net_listen_sock = sock;
-       INIT_WORK(&o2net_listen_work, o2net_accept_many);
-
-       sock->sk->sk_reuse = 1;
-       ret = sock->ops->bind(sock, (struct sockaddr *)&sin, sizeof(sin));
-       if (ret < 0) {
-               printk(KERN_ERR "o2net: Error %d while binding socket at "
-                      "%pI4:%u\n", ret, &addr, ntohs(port)); 
-               goto out;
-       }
-
-       ret = sock->ops->listen(sock, 64);
-       if (ret < 0)
-               printk(KERN_ERR "o2net: Error %d while listening on %pI4:%u\n",
-                      ret, &addr, ntohs(port));
-
-out:
-       if (ret) {
-               o2net_listen_sock = NULL;
-               if (sock)
-                       sock_release(sock);
-       }
-       return ret;
-}
-
-/*
- * called from node manager when we should bring up our network listening
- * socket.  node manager handles all the serialization to only call this
- * once and to match it with o2net_stop_listening().  note,
- * o2nm_this_node() doesn't work yet as we're being called while it
- * is being set up.
- */
-int o2net_start_listening(struct o2nm_node *node)
-{
-       int ret = 0;
-
-       BUG_ON(o2net_wq != NULL);
-       BUG_ON(o2net_listen_sock != NULL);
-
-       mlog(ML_KTHREAD, "starting o2net thread...\n");
-       o2net_wq = create_singlethread_workqueue("o2net");
-       if (o2net_wq == NULL) {
-               mlog(ML_ERROR, "unable to launch o2net thread\n");
-               return -ENOMEM; /* ? */
-       }
-
-       ret = o2net_open_listening_sock(node->nd_ipv4_address,
-                                       node->nd_ipv4_port);
-       if (ret) {
-               destroy_workqueue(o2net_wq);
-               o2net_wq = NULL;
-       } else
-               o2quo_conn_up(node->nd_num);
-
-       return ret;
-}
-
-/* again, o2nm_this_node() doesn't work here as we're involved in
- * tearing it down */
-void o2net_stop_listening(struct o2nm_node *node)
-{
-       struct socket *sock = o2net_listen_sock;
-       size_t i;
-
-       BUG_ON(o2net_wq == NULL);
-       BUG_ON(o2net_listen_sock == NULL);
-
-       /* stop the listening socket from generating work */
-       write_lock_bh(&sock->sk->sk_callback_lock);
-       sock->sk->sk_data_ready = sock->sk->sk_user_data;
-       sock->sk->sk_user_data = NULL;
-       write_unlock_bh(&sock->sk->sk_callback_lock);
-
-       for (i = 0; i < ARRAY_SIZE(o2net_nodes); i++) {
-               struct o2nm_node *node = o2nm_get_node_by_num(i);
-               if (node) {
-                       o2net_disconnect_node(node);
-                       o2nm_node_put(node);
-               }
-       }
-
-       /* finish all work and tear down the work queue */
-       mlog(ML_KTHREAD, "waiting for o2net thread to exit....\n");
-       destroy_workqueue(o2net_wq);
-       o2net_wq = NULL;
-
-       sock_release(o2net_listen_sock);
-       o2net_listen_sock = NULL;
-
-       o2quo_conn_err(node->nd_num);
-}
-
-#ifdef CONFIG_RAMSTER
-void o2net_hb_node_up_manual(int node_num)
-{
-       struct o2nm_node dummy;
-       o2hb_manual_set_node_heartbeating(node_num);
-       o2net_hb_node_up_cb(&dummy, node_num, NULL);
-}
-#endif
-
-/* ------------------------------------------------------------ */
-
-int o2net_init(void)
-{
-       unsigned long i;
-
-       o2quo_init();
-
-       if (o2net_debugfs_init())
-               return -ENOMEM;
-
-       o2net_hand = kzalloc(sizeof(struct o2net_handshake), GFP_KERNEL);
-       o2net_keep_req = kzalloc(sizeof(struct o2net_msg), GFP_KERNEL);
-       o2net_keep_resp = kzalloc(sizeof(struct o2net_msg), GFP_KERNEL);
-       if (!o2net_hand || !o2net_keep_req || !o2net_keep_resp) {
-               kfree(o2net_hand);
-               kfree(o2net_keep_req);
-               kfree(o2net_keep_resp);
-               return -ENOMEM;
-       }
-
-       o2net_hand->protocol_version = cpu_to_be64(O2NET_PROTOCOL_VERSION);
-       o2net_hand->connector_id = cpu_to_be64(1);
-
-       o2net_keep_req->magic = cpu_to_be16(O2NET_MSG_KEEP_REQ_MAGIC);
-       o2net_keep_resp->magic = cpu_to_be16(O2NET_MSG_KEEP_RESP_MAGIC);
-
-       for (i = 0; i < ARRAY_SIZE(o2net_nodes); i++) {
-               struct o2net_node *nn = o2net_nn_from_num(i);
-
-               atomic_set(&nn->nn_timeout, 0);
-               spin_lock_init(&nn->nn_lock);
-               INIT_DELAYED_WORK(&nn->nn_connect_work, o2net_start_connect);
-               INIT_DELAYED_WORK(&nn->nn_connect_expired,
-                                 o2net_connect_expired);
-               INIT_DELAYED_WORK(&nn->nn_still_up, o2net_still_up);
-               /* until we see hb from a node we'll return einval */
-               nn->nn_persistent_error = -ENOTCONN;
-               init_waitqueue_head(&nn->nn_sc_wq);
-               idr_init(&nn->nn_status_idr);
-               INIT_LIST_HEAD(&nn->nn_status_list);
-       }
-
-       return 0;
-}
-
-void o2net_exit(void)
-{
-       o2quo_exit();
-       kfree(o2net_hand);
-       kfree(o2net_keep_req);
-       kfree(o2net_keep_resp);
-       o2net_debugfs_exit();
-}
diff --git a/drivers/staging/ramster/cluster/tcp.h b/drivers/staging/ramster/cluster/tcp.h
deleted file mode 100644 (file)
index 95a991f..0000000
+++ /dev/null
@@ -1,160 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * tcp.h
- *
- * Function prototypes
- *
- * Copyright (C) 2004 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- *
- */
-
-#ifndef O2CLUSTER_TCP_H
-#define O2CLUSTER_TCP_H
-
-#include <linux/socket.h>
-#ifdef __KERNEL__
-#include <net/sock.h>
-#include <linux/tcp.h>
-#else
-#include <sys/socket.h>
-#endif
-#include <linux/inet.h>
-#include <linux/in.h>
-
-struct o2net_msg
-{
-       __be16 magic;
-       __be16 data_len;
-       __be16 msg_type;
-       __be16 pad1;
-       __be32 sys_status;
-       __be32 status;
-       __be32 key;
-       __be32 msg_num;
-       __u8  buf[0];
-};
-
-typedef int (o2net_msg_handler_func)(struct o2net_msg *msg, u32 len, void *data,
-                                    void **ret_data);
-typedef void (o2net_post_msg_handler_func)(int status, void *data,
-                                          void *ret_data);
-
-#define O2NET_MAX_PAYLOAD_BYTES  (4096 - sizeof(struct o2net_msg))
-
-/* same as hb delay, we're waiting for another node to recognize our hb */
-#define O2NET_RECONNECT_DELAY_MS_DEFAULT       2000
-
-#define O2NET_KEEPALIVE_DELAY_MS_DEFAULT       2000
-#define O2NET_IDLE_TIMEOUT_MS_DEFAULT          30000
-
-
-/* TODO: figure this out.... */
-static inline int o2net_link_down(int err, struct socket *sock)
-{
-       if (sock) {
-               if (sock->sk->sk_state != TCP_ESTABLISHED &&
-                   sock->sk->sk_state != TCP_CLOSE_WAIT)
-                       return 1;
-       }
-
-       if (err >= 0)
-               return 0;
-       switch (err) {
-               /* ????????????????????????? */
-               case -ERESTARTSYS:
-               case -EBADF:
-               /* When the server has died, an ICMP port unreachable
-                * message prompts ECONNREFUSED. */
-               case -ECONNREFUSED:
-               case -ENOTCONN:
-               case -ECONNRESET:
-               case -EPIPE:
-                       return 1;
-       }
-       return 0;
-}
-
-enum {
-       O2NET_DRIVER_UNINITED,
-       O2NET_DRIVER_READY,
-};
-
-int o2net_send_message(u32 msg_type, u32 key, void *data, u32 len,
-                      u8 target_node, int *status);
-int o2net_send_message_vec(u32 msg_type, u32 key, struct kvec *vec,
-                          size_t veclen, u8 target_node, int *status);
-
-int o2net_register_handler(u32 msg_type, u32 key, u32 max_len,
-                          o2net_msg_handler_func *func, void *data,
-                          o2net_post_msg_handler_func *post_func,
-                          struct list_head *unreg_list);
-void o2net_unregister_handler_list(struct list_head *list);
-
-void o2net_fill_node_map(unsigned long *map, unsigned bytes);
-
-#ifdef CONFIG_RAMSTER
-void o2net_force_data_magic(struct o2net_msg *, u16, u32);
-void o2net_hb_node_up_manual(int);
-struct o2net_node *o2net_nn_from_num(u8);
-#endif
-
-struct o2nm_node;
-int o2net_register_hb_callbacks(void);
-void o2net_unregister_hb_callbacks(void);
-int o2net_start_listening(struct o2nm_node *node);
-void o2net_stop_listening(struct o2nm_node *node);
-void o2net_disconnect_node(struct o2nm_node *node);
-int o2net_num_connected_peers(void);
-
-int o2net_init(void);
-void o2net_exit(void);
-
-struct o2net_send_tracking;
-struct o2net_sock_container;
-
-#ifdef CONFIG_DEBUG_FS
-int o2net_debugfs_init(void);
-void o2net_debugfs_exit(void);
-void o2net_debug_add_nst(struct o2net_send_tracking *nst);
-void o2net_debug_del_nst(struct o2net_send_tracking *nst);
-void o2net_debug_add_sc(struct o2net_sock_container *sc);
-void o2net_debug_del_sc(struct o2net_sock_container *sc);
-#else
-static inline int o2net_debugfs_init(void)
-{
-       return 0;
-}
-static inline void o2net_debugfs_exit(void)
-{
-}
-static inline void o2net_debug_add_nst(struct o2net_send_tracking *nst)
-{
-}
-static inline void o2net_debug_del_nst(struct o2net_send_tracking *nst)
-{
-}
-static inline void o2net_debug_add_sc(struct o2net_sock_container *sc)
-{
-}
-static inline void o2net_debug_del_sc(struct o2net_sock_container *sc)
-{
-}
-#endif /* CONFIG_DEBUG_FS */
-
-#endif /* O2CLUSTER_TCP_H */
diff --git a/drivers/staging/ramster/cluster/tcp_internal.h b/drivers/staging/ramster/cluster/tcp_internal.h
deleted file mode 100644 (file)
index ce17135..0000000
+++ /dev/null
@@ -1,249 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#ifndef O2CLUSTER_TCP_INTERNAL_H
-#define O2CLUSTER_TCP_INTERNAL_H
-
-#define O2NET_MSG_MAGIC           ((u16)0xfa55)
-#define O2NET_MSG_STATUS_MAGIC    ((u16)0xfa56)
-#define O2NET_MSG_KEEP_REQ_MAGIC  ((u16)0xfa57)
-#define O2NET_MSG_KEEP_RESP_MAGIC ((u16)0xfa58)
-#ifdef CONFIG_RAMSTER
-/*
- * "data magic" is a long version of "status magic" where the message
- * payload actually contains data to be passed in reply to certain messages
- */
-#define O2NET_MSG_DATA_MAGIC      ((u16)0xfa59)
-#endif
-
-/* we're delaying our quorum decision so that heartbeat will have timed
- * out truly dead nodes by the time we come around to making decisions
- * on their number */
-#define O2NET_QUORUM_DELAY_MS  ((o2hb_dead_threshold + 2) * O2HB_REGION_TIMEOUT_MS)
-
-/*
- * This version number represents quite a lot, unfortunately.  It not
- * only represents the raw network message protocol on the wire but also
- * locking semantics of the file system using the protocol.  It should
- * be somewhere else, I'm sure, but right now it isn't.
- *
- * With version 11, we separate out the filesystem locking portion.  The
- * filesystem now has a major.minor version it negotiates.  Version 11
- * introduces this negotiation to the o2dlm protocol, and as such the
- * version here in tcp_internal.h should not need to be bumped for
- * filesystem locking changes.
- *
- * New in version 11
- *     - Negotiation of filesystem locking in the dlm join.
- *
- * New in version 10:
- *     - Meta/data locks combined
- *
- * New in version 9:
- *     - All votes removed
- *
- * New in version 8:
- *     - Replace delete inode votes with a cluster lock
- *
- * New in version 7:
- *     - DLM join domain includes the live nodemap
- *
- * New in version 6:
- *     - DLM lockres remote refcount fixes.
- *
- * New in version 5:
- *     - Network timeout checking protocol
- *
- * New in version 4:
- *     - Remove i_generation from lock names for better stat performance.
- *
- * New in version 3:
- *     - Replace dentry votes with a cluster lock
- *
- * New in version 2:
- *     - full 64 bit i_size in the metadata lock lvbs
- *     - introduction of "rw" lock and pushing meta/data locking down
- */
-#define O2NET_PROTOCOL_VERSION 11ULL
-struct o2net_handshake {
-       __be64  protocol_version;
-       __be64  connector_id;
-       __be32  o2hb_heartbeat_timeout_ms;
-       __be32  o2net_idle_timeout_ms;
-       __be32  o2net_keepalive_delay_ms;
-       __be32  o2net_reconnect_delay_ms;
-};
-
-struct o2net_node {
-       /* this is never called from int/bh */
-       spinlock_t                      nn_lock;
-
-       /* set the moment an sc is allocated and a connect is started */
-       struct o2net_sock_container     *nn_sc;
-       /* _valid is only set after the handshake passes and tx can happen */
-       unsigned                        nn_sc_valid:1;
-       /* if this is set tx just returns it */
-       int                             nn_persistent_error;
-       /* It is only set to 1 after the idle time out. */
-       atomic_t                        nn_timeout;
-
-       /* threads waiting for an sc to arrive wait on the wq for generation
-        * to increase.  it is increased when a connecting socket succeeds
-        * or fails or when an accepted socket is attached. */
-       wait_queue_head_t               nn_sc_wq;
-
-       struct idr                      nn_status_idr;
-       struct list_head                nn_status_list;
-
-       /* connects are attempted from when heartbeat comes up until either hb
-        * goes down, the node is unconfigured, no connect attempts succeed
-        * before O2NET_CONN_IDLE_DELAY, or a connect succeeds.  connect_work
-        * is queued from set_nn_state both from hb up and from itself if a
-        * connect attempt fails and so can be self-arming.  shutdown is
-        * careful to first mark the nn such that no connects will be attempted
-        * before canceling delayed connect work and flushing the queue. */
-       struct delayed_work             nn_connect_work;
-       unsigned long                   nn_last_connect_attempt;
-
-       /* this is queued as nodes come up and is canceled when a connection is
-        * established.  this expiring gives up on the node and errors out
-        * transmits */
-       struct delayed_work             nn_connect_expired;
-
-       /* after we give up on a socket we wait a while before deciding
-        * that it is still heartbeating and that we should do some
-        * quorum work */
-       struct delayed_work             nn_still_up;
-};
-
-struct o2net_sock_container {
-       struct kref             sc_kref;
-       /* the next two are valid for the life time of the sc */
-       struct socket           *sc_sock;
-       struct o2nm_node        *sc_node;
-
-       /* all of these sc work structs hold refs on the sc while they are
-        * queued.  they should not be able to ref a freed sc.  the teardown
-        * race is with o2net_wq destruction in o2net_stop_listening() */
-
-       /* rx and connect work are generated from socket callbacks.  sc
-        * shutdown removes the callbacks and then flushes the work queue */
-       struct work_struct      sc_rx_work;
-       struct work_struct      sc_connect_work;
-       /* shutdown work is triggered in two ways.  the simple way is
-        * for a code path calls ensure_shutdown which gets a lock, removes
-        * the sc from the nn, and queues the work.  in this case the
-        * work is single-shot.  the work is also queued from a sock
-        * callback, though, and in this case the work will find the sc
-        * still on the nn and will call ensure_shutdown itself.. this
-        * ends up triggering the shutdown work again, though nothing
-        * will be done in that second iteration.  so work queue teardown
-        * has to be careful to remove the sc from the nn before waiting
-        * on the work queue so that the shutdown work doesn't remove the
-        * sc and rearm itself.
-        */
-       struct work_struct      sc_shutdown_work;
-
-       struct timer_list       sc_idle_timeout;
-       struct delayed_work     sc_keepalive_work;
-
-       unsigned                sc_handshake_ok:1;
-
-       struct page             *sc_page;
-       size_t                  sc_page_off;
-
-       /* original handlers for the sockets */
-       void                    (*sc_state_change)(struct sock *sk);
-       void                    (*sc_data_ready)(struct sock *sk, int bytes);
-
-       u32                     sc_msg_key;
-       u16                     sc_msg_type;
-
-#ifdef CONFIG_DEBUG_FS
-       struct list_head        sc_net_debug_item;
-       ktime_t                 sc_tv_timer;
-       ktime_t                 sc_tv_data_ready;
-       ktime_t                 sc_tv_advance_start;
-       ktime_t                 sc_tv_advance_stop;
-       ktime_t                 sc_tv_func_start;
-       ktime_t                 sc_tv_func_stop;
-#endif
-#ifdef CONFIG_OCFS2_FS_STATS
-       ktime_t                 sc_tv_acquiry_total;
-       ktime_t                 sc_tv_send_total;
-       ktime_t                 sc_tv_status_total;
-       u32                     sc_send_count;
-       u32                     sc_recv_count;
-       ktime_t                 sc_tv_process_total;
-#endif
-       struct mutex            sc_send_lock;
-};
-
-struct o2net_msg_handler {
-       struct rb_node          nh_node;
-       u32                     nh_max_len;
-       u32                     nh_msg_type;
-       u32                     nh_key;
-       o2net_msg_handler_func  *nh_func;
-       o2net_msg_handler_func  *nh_func_data;
-       o2net_post_msg_handler_func
-                               *nh_post_func;
-       struct kref             nh_kref;
-       struct list_head        nh_unregister_item;
-};
-
-enum o2net_system_error {
-       O2NET_ERR_NONE = 0,
-       O2NET_ERR_NO_HNDLR,
-       O2NET_ERR_OVERFLOW,
-       O2NET_ERR_DIED,
-       O2NET_ERR_MAX
-};
-
-struct o2net_status_wait {
-       enum o2net_system_error ns_sys_status;
-       s32                     ns_status;
-       int                     ns_id;
-       wait_queue_head_t       ns_wq;
-       struct list_head        ns_node_item;
-};
-
-#ifdef CONFIG_DEBUG_FS
-/* just for state dumps */
-struct o2net_send_tracking {
-       struct list_head                st_net_debug_item;
-       struct task_struct              *st_task;
-       struct o2net_sock_container     *st_sc;
-       u32                             st_id;
-       u32                             st_msg_type;
-       u32                             st_msg_key;
-       u8                              st_node;
-       ktime_t                         st_sock_time;
-       ktime_t                         st_send_time;
-       ktime_t                         st_status_time;
-};
-#else
-struct o2net_send_tracking {
-       u32     dummy;
-};
-#endif /* CONFIG_DEBUG_FS */
-
-#endif /* O2CLUSTER_TCP_INTERNAL_H */
diff --git a/drivers/staging/ramster/cluster/ver.c b/drivers/staging/ramster/cluster/ver.c
deleted file mode 100644 (file)
index a56eee6..0000000
+++ /dev/null
@@ -1,42 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * ver.c
- *
- * version string
- *
- * Copyright (C) 2002, 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#include <linux/module.h>
-#include <linux/kernel.h>
-
-#include "ver.h"
-
-#define CLUSTER_BUILD_VERSION "1.5.0"
-
-#define VERSION_STR "OCFS2 Node Manager " CLUSTER_BUILD_VERSION
-
-void cluster_print_version(void)
-{
-       printk(KERN_INFO "%s\n", VERSION_STR);
-}
-
-MODULE_DESCRIPTION(VERSION_STR);
-
-MODULE_VERSION(CLUSTER_BUILD_VERSION);
diff --git a/drivers/staging/ramster/cluster/ver.h b/drivers/staging/ramster/cluster/ver.h
deleted file mode 100644 (file)
index 32554c3..0000000
+++ /dev/null
@@ -1,31 +0,0 @@
-/* -*- mode: c; c-basic-offset: 8; -*-
- * vim: noexpandtab sw=8 ts=8 sts=0:
- *
- * ver.h
- *
- * Function prototypes
- *
- * Copyright (C) 2005 Oracle.  All rights reserved.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public
- * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public
- * License along with this program; if not, write to the
- * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
- * Boston, MA 021110-1307, USA.
- */
-
-#ifndef O2CLUSTER_VER_H
-#define O2CLUSTER_VER_H
-
-void cluster_print_version(void);
-
-#endif /* O2CLUSTER_VER_H */
diff --git a/drivers/staging/ramster/ramster.h b/drivers/staging/ramster/ramster.h
deleted file mode 100644 (file)
index 3293512..0000000
+++ /dev/null
@@ -1,117 +0,0 @@
-/*
- * ramster.h
- *
- * Peer-to-peer transcendent memory
- *
- * Copyright (c) 2009-2012, Dan Magenheimer, Oracle Corp.
- */
-
-#ifndef _RAMSTER_H_
-#define _RAMSTER_H_
-
-/*
- * format of remote pampd:
- *   bit 0 == intransit
- *   bit 1 == is_remote... if this bit is set, then
- *   bit 2-9 == remotenode
- *   bit 10-22 == size
- *   bit 23-30 == cksum
- */
-#define FAKE_PAMPD_INTRANSIT_BITS      1
-#define FAKE_PAMPD_ISREMOTE_BITS       1
-#define FAKE_PAMPD_REMOTENODE_BITS     8
-#define FAKE_PAMPD_REMOTESIZE_BITS     13
-#define FAKE_PAMPD_CHECKSUM_BITS       8
-
-#define FAKE_PAMPD_INTRANSIT_SHIFT     0
-#define FAKE_PAMPD_ISREMOTE_SHIFT      (FAKE_PAMPD_INTRANSIT_SHIFT + \
-                                        FAKE_PAMPD_INTRANSIT_BITS)
-#define FAKE_PAMPD_REMOTENODE_SHIFT    (FAKE_PAMPD_ISREMOTE_SHIFT + \
-                                        FAKE_PAMPD_ISREMOTE_BITS)
-#define FAKE_PAMPD_REMOTESIZE_SHIFT    (FAKE_PAMPD_REMOTENODE_SHIFT + \
-                                        FAKE_PAMPD_REMOTENODE_BITS)
-#define FAKE_PAMPD_CHECKSUM_SHIFT      (FAKE_PAMPD_REMOTESIZE_SHIFT + \
-                                        FAKE_PAMPD_REMOTESIZE_BITS)
-
-#define FAKE_PAMPD_MASK(x)             ((1UL << (x)) - 1)
-
-static inline void *pampd_make_remote(int remotenode, size_t size,
-                                       unsigned char cksum)
-{
-       unsigned long fake_pampd = 0;
-       fake_pampd |= 1UL << FAKE_PAMPD_ISREMOTE_SHIFT;
-       fake_pampd |= ((unsigned long)remotenode &
-                       FAKE_PAMPD_MASK(FAKE_PAMPD_REMOTENODE_BITS)) <<
-                               FAKE_PAMPD_REMOTENODE_SHIFT;
-       fake_pampd |= ((unsigned long)size &
-                       FAKE_PAMPD_MASK(FAKE_PAMPD_REMOTESIZE_BITS)) <<
-                               FAKE_PAMPD_REMOTESIZE_SHIFT;
-       fake_pampd |= ((unsigned long)cksum &
-                       FAKE_PAMPD_MASK(FAKE_PAMPD_CHECKSUM_BITS)) <<
-                               FAKE_PAMPD_CHECKSUM_SHIFT;
-       return (void *)fake_pampd;
-}
-
-static inline unsigned int pampd_remote_node(void *pampd)
-{
-       unsigned long fake_pampd = (unsigned long)pampd;
-       return (fake_pampd >> FAKE_PAMPD_REMOTENODE_SHIFT) &
-               FAKE_PAMPD_MASK(FAKE_PAMPD_REMOTENODE_BITS);
-}
-
-static inline unsigned int pampd_remote_size(void *pampd)
-{
-       unsigned long fake_pampd = (unsigned long)pampd;
-       return (fake_pampd >> FAKE_PAMPD_REMOTESIZE_SHIFT) &
-               FAKE_PAMPD_MASK(FAKE_PAMPD_REMOTESIZE_BITS);
-}
-
-static inline unsigned char pampd_remote_cksum(void *pampd)
-{
-       unsigned long fake_pampd = (unsigned long)pampd;
-       return (fake_pampd >> FAKE_PAMPD_CHECKSUM_SHIFT) &
-               FAKE_PAMPD_MASK(FAKE_PAMPD_CHECKSUM_BITS);
-}
-
-static inline bool pampd_is_remote(void *pampd)
-{
-       unsigned long fake_pampd = (unsigned long)pampd;
-       return (fake_pampd >> FAKE_PAMPD_ISREMOTE_SHIFT) &
-               FAKE_PAMPD_MASK(FAKE_PAMPD_ISREMOTE_BITS);
-}
-
-static inline bool pampd_is_intransit(void *pampd)
-{
-       unsigned long fake_pampd = (unsigned long)pampd;
-       return (fake_pampd >> FAKE_PAMPD_INTRANSIT_SHIFT) &
-               FAKE_PAMPD_MASK(FAKE_PAMPD_INTRANSIT_BITS);
-}
-
-/* note that it is a BUG for intransit to be set without isremote also set */
-static inline void *pampd_mark_intransit(void *pampd)
-{
-       unsigned long fake_pampd = (unsigned long)pampd;
-
-       fake_pampd |= 1UL << FAKE_PAMPD_ISREMOTE_SHIFT;
-       fake_pampd |= 1UL << FAKE_PAMPD_INTRANSIT_SHIFT;
-       return (void *)fake_pampd;
-}
-
-static inline void *pampd_mask_intransit_and_remote(void *marked_pampd)
-{
-       unsigned long pampd = (unsigned long)marked_pampd;
-
-       pampd &= ~(1UL << FAKE_PAMPD_INTRANSIT_SHIFT);
-       pampd &= ~(1UL << FAKE_PAMPD_ISREMOTE_SHIFT);
-       return (void *)pampd;
-}
-
-extern int ramster_remote_async_get(struct tmem_xhandle *,
-                               bool, int, size_t, uint8_t, void *extra);
-extern int ramster_remote_put(struct tmem_xhandle *, char *, size_t,
-                               bool, int *);
-extern int ramster_remote_flush(struct tmem_xhandle *, int);
-extern int ramster_remote_flush_object(struct tmem_xhandle *, int);
-extern int ramster_o2net_register_handlers(void);
-
-#endif /* _TMEM_H */
diff --git a/drivers/staging/ramster/ramster_o2net.c b/drivers/staging/ramster/ramster_o2net.c
deleted file mode 100644 (file)
index c5a47cc..0000000
+++ /dev/null
@@ -1,402 +0,0 @@
-/*
- * ramster_o2net.c
- *
- * Copyright (c) 2011, Dan Magenheimer, Oracle Corp.
- *
- * Ramster_o2net provides an interface between zcache and o2net.
- *
- * FIXME: support more than two nodes
- */
-
-#include <linux/list.h>
-#include "cluster/tcp.h"
-#include "cluster/nodemanager.h"
-#include "tmem.h"
-#include "zcache.h"
-#include "ramster.h"
-
-#define RAMSTER_TESTING
-
-#define RMSTR_KEY      0x77347734
-
-enum {
-       RMSTR_TMEM_PUT_EPH = 100,
-       RMSTR_TMEM_PUT_PERS,
-       RMSTR_TMEM_ASYNC_GET_REQUEST,
-       RMSTR_TMEM_ASYNC_GET_AND_FREE_REQUEST,
-       RMSTR_TMEM_ASYNC_GET_REPLY,
-       RMSTR_TMEM_FLUSH,
-       RMSTR_TMEM_FLOBJ,
-       RMSTR_TMEM_DESTROY_POOL,
-};
-
-#define RMSTR_O2NET_MAX_LEN \
-               (O2NET_MAX_PAYLOAD_BYTES - sizeof(struct tmem_xhandle))
-
-#include "cluster/tcp_internal.h"
-
-static struct o2nm_node *ramster_choose_node(int *nodenum,
-                                               struct tmem_xhandle *xh)
-{
-       struct o2nm_node *node = NULL;
-       int i;
-
-/* FIXME reproducibly pick a node based on xh that is NOT this node */
-       i = o2nm_this_node();
-       i = !i;         /* FIXME ONLY FOR TWO NODES */
-       node = o2nm_get_node_by_num(i);
-               /* WARNING: THIS DOES NOT CHECK TO ENSURE CONNECTED */
-       if (node != NULL)
-               *nodenum = i;
-       return node;
-}
-
-static void ramster_put_node(struct o2nm_node *node)
-{
-       o2nm_node_put(node);
-}
-
-/* FIXME following buffer should be per-cpu, protected by preempt_disable */
-static char ramster_async_get_buf[O2NET_MAX_PAYLOAD_BYTES];
-
-static int ramster_remote_async_get_request_handler(struct o2net_msg *msg,
-                               u32 len, void *data, void **ret_data)
-{
-       char *pdata;
-       struct tmem_xhandle xh;
-       int found;
-       size_t size = RMSTR_O2NET_MAX_LEN;
-       u16 msgtype = be16_to_cpu(msg->msg_type);
-       bool get_and_free = (msgtype == RMSTR_TMEM_ASYNC_GET_AND_FREE_REQUEST);
-       unsigned long flags;
-
-       xh = *(struct tmem_xhandle *)msg->buf;
-       if (xh.xh_data_size > RMSTR_O2NET_MAX_LEN)
-               BUG();
-       pdata = ramster_async_get_buf;
-       *(struct tmem_xhandle *)pdata = xh;
-       pdata += sizeof(struct tmem_xhandle);
-       local_irq_save(flags);
-       found = zcache_get(xh.client_id, xh.pool_id, &xh.oid, xh.index,
-                               pdata, &size, 1, get_and_free ? 1 : -1);
-       local_irq_restore(flags);
-       if (found < 0) {
-               /* a zero size indicates the get failed */
-               size = 0;
-       }
-       if (size > RMSTR_O2NET_MAX_LEN)
-               BUG();
-       *ret_data = pdata - sizeof(struct tmem_xhandle);
-       /* now make caller (o2net_process_message) handle specially */
-       o2net_force_data_magic(msg, RMSTR_TMEM_ASYNC_GET_REPLY, RMSTR_KEY);
-       return size + sizeof(struct tmem_xhandle);
-}
-
-static int ramster_remote_async_get_reply_handler(struct o2net_msg *msg,
-                               u32 len, void *data, void **ret_data)
-{
-       char *in = (char *)msg->buf;
-       int datalen = len - sizeof(struct o2net_msg);
-       int ret = -1;
-       struct tmem_xhandle *xh = (struct tmem_xhandle *)in;
-
-       in += sizeof(struct tmem_xhandle);
-       datalen -= sizeof(struct tmem_xhandle);
-       BUG_ON(datalen < 0 || datalen > PAGE_SIZE);
-       ret = zcache_localify(xh->pool_id, &xh->oid, xh->index,
-                               in, datalen, xh->extra);
-#ifdef RAMSTER_TESTING
-       if (ret == -EEXIST)
-               pr_err("TESTING ArrgREP, aborted overwrite on racy put\n");
-#endif
-       return ret;
-}
-
-int ramster_remote_put_handler(struct o2net_msg *msg,
-                               u32 len, void *data, void **ret_data)
-{
-       struct tmem_xhandle *xh;
-       char *p = (char *)msg->buf;
-       int datalen = len - sizeof(struct o2net_msg) -
-                               sizeof(struct tmem_xhandle);
-       u16 msgtype = be16_to_cpu(msg->msg_type);
-       bool ephemeral = (msgtype == RMSTR_TMEM_PUT_EPH);
-       unsigned long flags;
-       int ret;
-
-       xh = (struct tmem_xhandle *)p;
-       p += sizeof(struct tmem_xhandle);
-       zcache_autocreate_pool(xh->client_id, xh->pool_id, ephemeral);
-       local_irq_save(flags);
-       ret = zcache_put(xh->client_id, xh->pool_id, &xh->oid, xh->index,
-                               p, datalen, 1, ephemeral ? 1 : -1);
-       local_irq_restore(flags);
-       return ret;
-}
-
-int ramster_remote_flush_handler(struct o2net_msg *msg,
-                               u32 len, void *data, void **ret_data)
-{
-       struct tmem_xhandle *xh;
-       char *p = (char *)msg->buf;
-
-       xh = (struct tmem_xhandle *)p;
-       p += sizeof(struct tmem_xhandle);
-       (void)zcache_flush(xh->client_id, xh->pool_id, &xh->oid, xh->index);
-       return 0;
-}
-
-int ramster_remote_flobj_handler(struct o2net_msg *msg,
-                               u32 len, void *data, void **ret_data)
-{
-       struct tmem_xhandle *xh;
-       char *p = (char *)msg->buf;
-
-       xh = (struct tmem_xhandle *)p;
-       p += sizeof(struct tmem_xhandle);
-       (void)zcache_flush_object(xh->client_id, xh->pool_id, &xh->oid);
-       return 0;
-}
-
-int ramster_remote_async_get(struct tmem_xhandle *xh, bool free, int remotenode,
-                               size_t expect_size, uint8_t expect_cksum,
-                               void *extra)
-{
-       int ret = -1, status;
-       struct o2nm_node *node = NULL;
-       struct kvec vec[1];
-       size_t veclen = 1;
-       u32 msg_type;
-
-       node = o2nm_get_node_by_num(remotenode);
-       if (node == NULL)
-               goto out;
-       xh->client_id = o2nm_this_node(); /* which node is getting */
-       xh->xh_data_cksum = expect_cksum;
-       xh->xh_data_size = expect_size;
-       xh->extra = extra;
-       vec[0].iov_len = sizeof(*xh);
-       vec[0].iov_base = xh;
-       if (free)
-               msg_type = RMSTR_TMEM_ASYNC_GET_AND_FREE_REQUEST;
-       else
-               msg_type = RMSTR_TMEM_ASYNC_GET_REQUEST;
-       ret = o2net_send_message_vec(msg_type, RMSTR_KEY,
-                                       vec, veclen, remotenode, &status);
-       ramster_put_node(node);
-       if (ret < 0) {
-               /* FIXME handle bad message possibilities here? */
-               pr_err("UNTESTED ret<0 in ramster_remote_async_get\n");
-       }
-       ret = status;
-out:
-       return ret;
-}
-
-#ifdef RAMSTER_TESTING
-/* leave me here to see if it catches a weird crash */
-static void ramster_check_irq_counts(void)
-{
-       static int last_hardirq_cnt, last_softirq_cnt, last_preempt_cnt;
-       int cur_hardirq_cnt, cur_softirq_cnt, cur_preempt_cnt;
-
-       cur_hardirq_cnt = hardirq_count() >> HARDIRQ_SHIFT;
-       if (cur_hardirq_cnt > last_hardirq_cnt) {
-               last_hardirq_cnt = cur_hardirq_cnt;
-               if (!(last_hardirq_cnt&(last_hardirq_cnt-1)))
-                       pr_err("RAMSTER TESTING RRP hardirq_count=%d\n",
-                               last_hardirq_cnt);
-       }
-       cur_softirq_cnt = softirq_count() >> SOFTIRQ_SHIFT;
-       if (cur_softirq_cnt > last_softirq_cnt) {
-               last_softirq_cnt = cur_softirq_cnt;
-               if (!(last_softirq_cnt&(last_softirq_cnt-1)))
-                       pr_err("RAMSTER TESTING RRP softirq_count=%d\n",
-                               last_softirq_cnt);
-       }
-       cur_preempt_cnt = preempt_count() & PREEMPT_MASK;
-       if (cur_preempt_cnt > last_preempt_cnt) {
-               last_preempt_cnt = cur_preempt_cnt;
-               if (!(last_preempt_cnt&(last_preempt_cnt-1)))
-                       pr_err("RAMSTER TESTING RRP preempt_count=%d\n",
-                               last_preempt_cnt);
-       }
-}
-#endif
-
-int ramster_remote_put(struct tmem_xhandle *xh, char *data, size_t size,
-                               bool ephemeral, int *remotenode)
-{
-       int nodenum, ret = -1, status;
-       struct o2nm_node *node = NULL;
-       struct kvec vec[2];
-       size_t veclen = 2;
-       u32 msg_type;
-#ifdef RAMSTER_TESTING
-       struct o2net_node *nn;
-#endif
-
-       BUG_ON(size > RMSTR_O2NET_MAX_LEN);
-       xh->client_id = o2nm_this_node(); /* which node is putting */
-       vec[0].iov_len = sizeof(*xh);
-       vec[0].iov_base = xh;
-       vec[1].iov_len = size;
-       vec[1].iov_base = data;
-       node = ramster_choose_node(&nodenum, xh);
-       if (!node)
-               goto out;
-
-#ifdef RAMSTER_TESTING
-       nn = o2net_nn_from_num(nodenum);
-       WARN_ON_ONCE(nn->nn_persistent_error || !nn->nn_sc_valid);
-#endif
-
-       if (ephemeral)
-               msg_type = RMSTR_TMEM_PUT_EPH;
-       else
-               msg_type = RMSTR_TMEM_PUT_PERS;
-#ifdef RAMSTER_TESTING
-       /* leave me here to see if it catches a weird crash */
-       ramster_check_irq_counts();
-#endif
-
-       ret = o2net_send_message_vec(msg_type, RMSTR_KEY,
-                                               vec, veclen, nodenum, &status);
-#ifdef RAMSTER_TESTING
-       if (ret != 0) {
-               static unsigned long cnt;
-               cnt++;
-               if (!(cnt&(cnt-1)))
-                       pr_err("ramster_remote_put: message failed, "
-                               "ret=%d, cnt=%lu\n", ret, cnt);
-               ret = -1;
-       }
-#endif
-       if (ret < 0)
-               ret = -1;
-       else {
-               ret = status;
-               *remotenode = nodenum;
-       }
-
-       ramster_put_node(node);
-out:
-       return ret;
-}
-
-int ramster_remote_flush(struct tmem_xhandle *xh, int remotenode)
-{
-       int ret = -1, status;
-       struct o2nm_node *node = NULL;
-       struct kvec vec[1];
-       size_t veclen = 1;
-
-       node = o2nm_get_node_by_num(remotenode);
-       BUG_ON(node == NULL);
-       xh->client_id = o2nm_this_node(); /* which node is flushing */
-       vec[0].iov_len = sizeof(*xh);
-       vec[0].iov_base = xh;
-       BUG_ON(irqs_disabled());
-       BUG_ON(in_softirq());
-       ret = o2net_send_message_vec(RMSTR_TMEM_FLUSH, RMSTR_KEY,
-                                       vec, veclen, remotenode, &status);
-       ramster_put_node(node);
-       return ret;
-}
-
-int ramster_remote_flush_object(struct tmem_xhandle *xh, int remotenode)
-{
-       int ret = -1, status;
-       struct o2nm_node *node = NULL;
-       struct kvec vec[1];
-       size_t veclen = 1;
-
-       node = o2nm_get_node_by_num(remotenode);
-       BUG_ON(node == NULL);
-       xh->client_id = o2nm_this_node(); /* which node is flobjing */
-       vec[0].iov_len = sizeof(*xh);
-       vec[0].iov_base = xh;
-       ret = o2net_send_message_vec(RMSTR_TMEM_FLOBJ, RMSTR_KEY,
-                                       vec, veclen, remotenode, &status);
-       ramster_put_node(node);
-       return ret;
-}
-
-/*
- * Handler registration
- */
-
-static LIST_HEAD(ramster_o2net_unreg_list);
-
-static void ramster_o2net_unregister_handlers(void)
-{
-       o2net_unregister_handler_list(&ramster_o2net_unreg_list);
-}
-
-int ramster_o2net_register_handlers(void)
-{
-       int status;
-
-       status = o2net_register_handler(RMSTR_TMEM_PUT_EPH, RMSTR_KEY,
-                               RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_put_handler,
-                               NULL, NULL, &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       status = o2net_register_handler(RMSTR_TMEM_PUT_PERS, RMSTR_KEY,
-                               RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_put_handler,
-                               NULL, NULL, &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       status = o2net_register_handler(RMSTR_TMEM_ASYNC_GET_REQUEST, RMSTR_KEY,
-                               RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_async_get_request_handler,
-                               NULL, NULL,
-                               &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       status = o2net_register_handler(RMSTR_TMEM_ASYNC_GET_AND_FREE_REQUEST,
-                               RMSTR_KEY, RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_async_get_request_handler,
-                               NULL, NULL,
-                               &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       status = o2net_register_handler(RMSTR_TMEM_ASYNC_GET_REPLY, RMSTR_KEY,
-                               RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_async_get_reply_handler,
-                               NULL, NULL,
-                               &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       status = o2net_register_handler(RMSTR_TMEM_FLUSH, RMSTR_KEY,
-                               RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_flush_handler,
-                               NULL, NULL,
-                               &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       status = o2net_register_handler(RMSTR_TMEM_FLOBJ, RMSTR_KEY,
-                               RMSTR_O2NET_MAX_LEN,
-                               ramster_remote_flobj_handler,
-                               NULL, NULL,
-                               &ramster_o2net_unreg_list);
-       if (status)
-               goto bail;
-
-       pr_info("ramster_o2net: handlers registered\n");
-
-bail:
-       if (status) {
-               ramster_o2net_unregister_handlers();
-               pr_err("ramster_o2net: couldn't register handlers\n");
-       }
-       return status;
-}
diff --git a/drivers/staging/ramster/tmem.c b/drivers/staging/ramster/tmem.c
deleted file mode 100644 (file)
index 8f2f689..0000000
+++ /dev/null
@@ -1,851 +0,0 @@
-/*
- * In-kernel transcendent memory (generic implementation)
- *
- * Copyright (c) 2009-2011, Dan Magenheimer, Oracle Corp.
- *
- * The primary purpose of Transcedent Memory ("tmem") is to map object-oriented
- * "handles" (triples containing a pool id, and object id, and an index), to
- * pages in a page-accessible memory (PAM).  Tmem references the PAM pages via
- * an abstract "pampd" (PAM page-descriptor), which can be operated on by a
- * set of functions (pamops).  Each pampd contains some representation of
- * PAGE_SIZE bytes worth of data. Tmem must support potentially millions of
- * pages and must be able to insert, find, and delete these pages at a
- * potential frequency of thousands per second concurrently across many CPUs,
- * (and, if used with KVM, across many vcpus across many guests).
- * Tmem is tracked with a hierarchy of data structures, organized by
- * the elements in a handle-tuple: pool_id, object_id, and page index.
- * One or more "clients" (e.g. guests) each provide one or more tmem_pools.
- * Each pool, contains a hash table of rb_trees of tmem_objs.  Each
- * tmem_obj contains a radix-tree-like tree of pointers, with intermediate
- * nodes called tmem_objnodes.  Each leaf pointer in this tree points to
- * a pampd, which is accessible only through a small set of callbacks
- * registered by the PAM implementation (see tmem_register_pamops). Tmem
- * does all memory allocation via a set of callbacks registered by the tmem
- * host implementation (e.g. see tmem_register_hostops).
- */
-
-#include <linux/list.h>
-#include <linux/spinlock.h>
-#include <linux/atomic.h>
-#include <linux/delay.h>
-
-#include "tmem.h"
-
-/* data structure sentinels used for debugging... see tmem.h */
-#define POOL_SENTINEL 0x87658765
-#define OBJ_SENTINEL 0x12345678
-#define OBJNODE_SENTINEL 0xfedcba09
-
-/*
- * A tmem host implementation must use this function to register callbacks
- * for memory allocation.
- */
-static struct tmem_hostops tmem_hostops;
-
-static void tmem_objnode_tree_init(void);
-
-void tmem_register_hostops(struct tmem_hostops *m)
-{
-       tmem_objnode_tree_init();
-       tmem_hostops = *m;
-}
-
-/*
- * A tmem host implementation must use this function to register
- * callbacks for a page-accessible memory (PAM) implementation
- */
-static struct tmem_pamops tmem_pamops;
-
-void tmem_register_pamops(struct tmem_pamops *m)
-{
-       tmem_pamops = *m;
-}
-
-/*
- * Oid's are potentially very sparse and tmem_objs may have an indeterminately
- * short life, being added and deleted at a relatively high frequency.
- * So an rb_tree is an ideal data structure to manage tmem_objs.  But because
- * of the potentially huge number of tmem_objs, each pool manages a hashtable
- * of rb_trees to reduce search, insert, delete, and rebalancing time.
- * Each hashbucket also has a lock to manage concurrent access.
- *
- * The following routines manage tmem_objs.  When any tmem_obj is accessed,
- * the hashbucket lock must be held.
- */
-
-/* searches for object==oid in pool, returns locked object if found */
-static struct tmem_obj *tmem_obj_find(struct tmem_hashbucket *hb,
-                                       struct tmem_oid *oidp)
-{
-       struct rb_node *rbnode;
-       struct tmem_obj *obj;
-
-       rbnode = hb->obj_rb_root.rb_node;
-       while (rbnode) {
-               BUG_ON(RB_EMPTY_NODE(rbnode));
-               obj = rb_entry(rbnode, struct tmem_obj, rb_tree_node);
-               switch (tmem_oid_compare(oidp, &obj->oid)) {
-               case 0: /* equal */
-                       goto out;
-               case -1:
-                       rbnode = rbnode->rb_left;
-                       break;
-               case 1:
-                       rbnode = rbnode->rb_right;
-                       break;
-               }
-       }
-       obj = NULL;
-out:
-       return obj;
-}
-
-static void tmem_pampd_destroy_all_in_obj(struct tmem_obj *);
-
-/* free an object that has no more pampds in it */
-static void tmem_obj_free(struct tmem_obj *obj, struct tmem_hashbucket *hb)
-{
-       struct tmem_pool *pool;
-
-       BUG_ON(obj == NULL);
-       ASSERT_SENTINEL(obj, OBJ);
-       BUG_ON(obj->pampd_count > 0);
-       pool = obj->pool;
-       BUG_ON(pool == NULL);
-       if (obj->objnode_tree_root != NULL) /* may be "stump" with no leaves */
-               tmem_pampd_destroy_all_in_obj(obj);
-       BUG_ON(obj->objnode_tree_root != NULL);
-       BUG_ON((long)obj->objnode_count != 0);
-       atomic_dec(&pool->obj_count);
-       BUG_ON(atomic_read(&pool->obj_count) < 0);
-       INVERT_SENTINEL(obj, OBJ);
-       obj->pool = NULL;
-       tmem_oid_set_invalid(&obj->oid);
-       rb_erase(&obj->rb_tree_node, &hb->obj_rb_root);
-}
-
-/*
- * initialize, and insert an tmem_object_root (called only if find failed)
- */
-static void tmem_obj_init(struct tmem_obj *obj, struct tmem_hashbucket *hb,
-                                       struct tmem_pool *pool,
-                                       struct tmem_oid *oidp)
-{
-       struct rb_root *root = &hb->obj_rb_root;
-       struct rb_node **new = &(root->rb_node), *parent = NULL;
-       struct tmem_obj *this;
-
-       BUG_ON(pool == NULL);
-       atomic_inc(&pool->obj_count);
-       obj->objnode_tree_height = 0;
-       obj->objnode_tree_root = NULL;
-       obj->pool = pool;
-       obj->oid = *oidp;
-       obj->objnode_count = 0;
-       obj->pampd_count = 0;
-       (*tmem_pamops.new_obj)(obj);
-       SET_SENTINEL(obj, OBJ);
-       while (*new) {
-               BUG_ON(RB_EMPTY_NODE(*new));
-               this = rb_entry(*new, struct tmem_obj, rb_tree_node);
-               parent = *new;
-               switch (tmem_oid_compare(oidp, &this->oid)) {
-               case 0:
-                       BUG(); /* already present; should never happen! */
-                       break;
-               case -1:
-                       new = &(*new)->rb_left;
-                       break;
-               case 1:
-                       new = &(*new)->rb_right;
-                       break;
-               }
-       }
-       rb_link_node(&obj->rb_tree_node, parent, new);
-       rb_insert_color(&obj->rb_tree_node, root);
-}
-
-/*
- * Tmem is managed as a set of tmem_pools with certain attributes, such as
- * "ephemeral" vs "persistent".  These attributes apply to all tmem_objs
- * and all pampds that belong to a tmem_pool.  A tmem_pool is created
- * or deleted relatively rarely (for example, when a filesystem is
- * mounted or unmounted.
- */
-
-/* flush all data from a pool and, optionally, free it */
-static void tmem_pool_flush(struct tmem_pool *pool, bool destroy)
-{
-       struct rb_node *rbnode;
-       struct tmem_obj *obj;
-       struct tmem_hashbucket *hb = &pool->hashbucket[0];
-       int i;
-
-       BUG_ON(pool == NULL);
-       for (i = 0; i < TMEM_HASH_BUCKETS; i++, hb++) {
-               spin_lock(&hb->lock);
-               rbnode = rb_first(&hb->obj_rb_root);
-               while (rbnode != NULL) {
-                       obj = rb_entry(rbnode, struct tmem_obj, rb_tree_node);
-                       rbnode = rb_next(rbnode);
-                       tmem_pampd_destroy_all_in_obj(obj);
-                       tmem_obj_free(obj, hb);
-                       (*tmem_hostops.obj_free)(obj, pool);
-               }
-               spin_unlock(&hb->lock);
-       }
-       if (destroy)
-               list_del(&pool->pool_list);
-}
-
-/*
- * A tmem_obj contains a radix-tree-like tree in which the intermediate
- * nodes are called tmem_objnodes.  (The kernel lib/radix-tree.c implementation
- * is very specialized and tuned for specific uses and is not particularly
- * suited for use from this code, though some code from the core algorithms has
- * been reused, thus the copyright notices below).  Each tmem_objnode contains
- * a set of pointers which point to either a set of intermediate tmem_objnodes
- * or a set of of pampds.
- *
- * Portions Copyright (C) 2001 Momchil Velikov
- * Portions Copyright (C) 2001 Christoph Hellwig
- * Portions Copyright (C) 2005 SGI, Christoph Lameter <clameter@sgi.com>
- */
-
-struct tmem_objnode_tree_path {
-       struct tmem_objnode *objnode;
-       int offset;
-};
-
-/* objnode height_to_maxindex translation */
-static unsigned long tmem_objnode_tree_h2max[OBJNODE_TREE_MAX_PATH + 1];
-
-static void tmem_objnode_tree_init(void)
-{
-       unsigned int ht, tmp;
-
-       for (ht = 0; ht < ARRAY_SIZE(tmem_objnode_tree_h2max); ht++) {
-               tmp = ht * OBJNODE_TREE_MAP_SHIFT;
-               if (tmp >= OBJNODE_TREE_INDEX_BITS)
-                       tmem_objnode_tree_h2max[ht] = ~0UL;
-               else
-                       tmem_objnode_tree_h2max[ht] =
-                           (~0UL >> (OBJNODE_TREE_INDEX_BITS - tmp - 1)) >> 1;
-       }
-}
-
-static struct tmem_objnode *tmem_objnode_alloc(struct tmem_obj *obj)
-{
-       struct tmem_objnode *objnode;
-
-       ASSERT_SENTINEL(obj, OBJ);
-       BUG_ON(obj->pool == NULL);
-       ASSERT_SENTINEL(obj->pool, POOL);
-       objnode = (*tmem_hostops.objnode_alloc)(obj->pool);
-       if (unlikely(objnode == NULL))
-               goto out;
-       objnode->obj = obj;
-       SET_SENTINEL(objnode, OBJNODE);
-       memset(&objnode->slots, 0, sizeof(objnode->slots));
-       objnode->slots_in_use = 0;
-       obj->objnode_count++;
-out:
-       return objnode;
-}
-
-static void tmem_objnode_free(struct tmem_objnode *objnode)
-{
-       struct tmem_pool *pool;
-       int i;
-
-       BUG_ON(objnode == NULL);
-       for (i = 0; i < OBJNODE_TREE_MAP_SIZE; i++)
-               BUG_ON(objnode->slots[i] != NULL);
-       ASSERT_SENTINEL(objnode, OBJNODE);
-       INVERT_SENTINEL(objnode, OBJNODE);
-       BUG_ON(objnode->obj == NULL);
-       ASSERT_SENTINEL(objnode->obj, OBJ);
-       pool = objnode->obj->pool;
-       BUG_ON(pool == NULL);
-       ASSERT_SENTINEL(pool, POOL);
-       objnode->obj->objnode_count--;
-       objnode->obj = NULL;
-       (*tmem_hostops.objnode_free)(objnode, pool);
-}
-
-/*
- * lookup index in object and return associated pampd (or NULL if not found)
- */
-static void **__tmem_pampd_lookup_in_obj(struct tmem_obj *obj, uint32_t index)
-{
-       unsigned int height, shift;
-       struct tmem_objnode **slot = NULL;
-
-       BUG_ON(obj == NULL);
-       ASSERT_SENTINEL(obj, OBJ);
-       BUG_ON(obj->pool == NULL);
-       ASSERT_SENTINEL(obj->pool, POOL);
-
-       height = obj->objnode_tree_height;
-       if (index > tmem_objnode_tree_h2max[obj->objnode_tree_height])
-               goto out;
-       if (height == 0 && obj->objnode_tree_root) {
-               slot = &obj->objnode_tree_root;
-               goto out;
-       }
-       shift = (height-1) * OBJNODE_TREE_MAP_SHIFT;
-       slot = &obj->objnode_tree_root;
-       while (height > 0) {
-               if (*slot == NULL)
-                       goto out;
-               slot = (struct tmem_objnode **)
-                       ((*slot)->slots +
-                        ((index >> shift) & OBJNODE_TREE_MAP_MASK));
-               shift -= OBJNODE_TREE_MAP_SHIFT;
-               height--;
-       }
-out:
-       return slot != NULL ? (void **)slot : NULL;
-}
-
-static void *tmem_pampd_lookup_in_obj(struct tmem_obj *obj, uint32_t index)
-{
-       struct tmem_objnode **slot;
-
-       slot = (struct tmem_objnode **)__tmem_pampd_lookup_in_obj(obj, index);
-       return slot != NULL ? *slot : NULL;
-}
-
-static void *tmem_pampd_replace_in_obj(struct tmem_obj *obj, uint32_t index,
-                                       void *new_pampd, bool no_free)
-{
-       struct tmem_objnode **slot;
-       void *ret = NULL;
-
-       slot = (struct tmem_objnode **)__tmem_pampd_lookup_in_obj(obj, index);
-       if ((slot != NULL) && (*slot != NULL)) {
-               void *old_pampd = *(void **)slot;
-               *(void **)slot = new_pampd;
-               if (!no_free)
-                       (*tmem_pamops.free)(old_pampd, obj->pool,
-                                               NULL, 0, false);
-               ret = new_pampd;
-       }
-       return ret;
-}
-
-static int tmem_pampd_add_to_obj(struct tmem_obj *obj, uint32_t index,
-                                       void *pampd)
-{
-       int ret = 0;
-       struct tmem_objnode *objnode = NULL, *newnode, *slot;
-       unsigned int height, shift;
-       int offset = 0;
-
-       /* if necessary, extend the tree to be higher  */
-       if (index > tmem_objnode_tree_h2max[obj->objnode_tree_height]) {
-               height = obj->objnode_tree_height + 1;
-               if (index > tmem_objnode_tree_h2max[height])
-                       while (index > tmem_objnode_tree_h2max[height])
-                               height++;
-               if (obj->objnode_tree_root == NULL) {
-                       obj->objnode_tree_height = height;
-                       goto insert;
-               }
-               do {
-                       newnode = tmem_objnode_alloc(obj);
-                       if (!newnode) {
-                               ret = -ENOMEM;
-                               goto out;
-                       }
-                       newnode->slots[0] = obj->objnode_tree_root;
-                       newnode->slots_in_use = 1;
-                       obj->objnode_tree_root = newnode;
-                       obj->objnode_tree_height++;
-               } while (height > obj->objnode_tree_height);
-       }
-insert:
-       slot = obj->objnode_tree_root;
-       height = obj->objnode_tree_height;
-       shift = (height-1) * OBJNODE_TREE_MAP_SHIFT;
-       while (height > 0) {
-               if (slot == NULL) {
-                       /* add a child objnode.  */
-                       slot = tmem_objnode_alloc(obj);
-                       if (!slot) {
-                               ret = -ENOMEM;
-                               goto out;
-                       }
-                       if (objnode) {
-
-                               objnode->slots[offset] = slot;
-                               objnode->slots_in_use++;
-                       } else
-                               obj->objnode_tree_root = slot;
-               }
-               /* go down a level */
-               offset = (index >> shift) & OBJNODE_TREE_MAP_MASK;
-               objnode = slot;
-               slot = objnode->slots[offset];
-               shift -= OBJNODE_TREE_MAP_SHIFT;
-               height--;
-       }
-       BUG_ON(slot != NULL);
-       if (objnode) {
-               objnode->slots_in_use++;
-               objnode->slots[offset] = pampd;
-       } else
-               obj->objnode_tree_root = pampd;
-       obj->pampd_count++;
-out:
-       return ret;
-}
-
-static void *tmem_pampd_delete_from_obj(struct tmem_obj *obj, uint32_t index)
-{
-       struct tmem_objnode_tree_path path[OBJNODE_TREE_MAX_PATH + 1];
-       struct tmem_objnode_tree_path *pathp = path;
-       struct tmem_objnode *slot = NULL;
-       unsigned int height, shift;
-       int offset;
-
-       BUG_ON(obj == NULL);
-       ASSERT_SENTINEL(obj, OBJ);
-       BUG_ON(obj->pool == NULL);
-       ASSERT_SENTINEL(obj->pool, POOL);
-       height = obj->objnode_tree_height;
-       if (index > tmem_objnode_tree_h2max[height])
-               goto out;
-       slot = obj->objnode_tree_root;
-       if (height == 0 && obj->objnode_tree_root) {
-               obj->objnode_tree_root = NULL;
-               goto out;
-       }
-       shift = (height - 1) * OBJNODE_TREE_MAP_SHIFT;
-       pathp->objnode = NULL;
-       do {
-               if (slot == NULL)
-                       goto out;
-               pathp++;
-               offset = (index >> shift) & OBJNODE_TREE_MAP_MASK;
-               pathp->offset = offset;
-               pathp->objnode = slot;
-               slot = slot->slots[offset];
-               shift -= OBJNODE_TREE_MAP_SHIFT;
-               height--;
-       } while (height > 0);
-       if (slot == NULL)
-               goto out;
-       while (pathp->objnode) {
-               pathp->objnode->slots[pathp->offset] = NULL;
-               pathp->objnode->slots_in_use--;
-               if (pathp->objnode->slots_in_use) {
-                       if (pathp->objnode == obj->objnode_tree_root) {
-                               while (obj->objnode_tree_height > 0 &&
-                                 obj->objnode_tree_root->slots_in_use == 1 &&
-                                 obj->objnode_tree_root->slots[0]) {
-                                       struct tmem_objnode *to_free =
-                                               obj->objnode_tree_root;
-
-                                       obj->objnode_tree_root =
-                                                       to_free->slots[0];
-                                       obj->objnode_tree_height--;
-                                       to_free->slots[0] = NULL;
-                                       to_free->slots_in_use = 0;
-                                       tmem_objnode_free(to_free);
-                               }
-                       }
-                       goto out;
-               }
-               tmem_objnode_free(pathp->objnode); /* 0 slots used, free it */
-               pathp--;
-       }
-       obj->objnode_tree_height = 0;
-       obj->objnode_tree_root = NULL;
-
-out:
-       if (slot != NULL)
-               obj->pampd_count--;
-       BUG_ON(obj->pampd_count < 0);
-       return slot;
-}
-
-/* recursively walk the objnode_tree destroying pampds and objnodes */
-static void tmem_objnode_node_destroy(struct tmem_obj *obj,
-                                       struct tmem_objnode *objnode,
-                                       unsigned int ht)
-{
-       int i;
-
-       if (ht == 0)
-               return;
-       for (i = 0; i < OBJNODE_TREE_MAP_SIZE; i++) {
-               if (objnode->slots[i]) {
-                       if (ht == 1) {
-                               obj->pampd_count--;
-                               (*tmem_pamops.free)(objnode->slots[i],
-                                               obj->pool, NULL, 0, true);
-                               objnode->slots[i] = NULL;
-                               continue;
-                       }
-                       tmem_objnode_node_destroy(obj, objnode->slots[i], ht-1);
-                       tmem_objnode_free(objnode->slots[i]);
-                       objnode->slots[i] = NULL;
-               }
-       }
-}
-
-static void tmem_pampd_destroy_all_in_obj(struct tmem_obj *obj)
-{
-       if (obj->objnode_tree_root == NULL)
-               return;
-       if (obj->objnode_tree_height == 0) {
-               obj->pampd_count--;
-               (*tmem_pamops.free)(obj->objnode_tree_root,
-                                       obj->pool, NULL, 0, true);
-       } else {
-               tmem_objnode_node_destroy(obj, obj->objnode_tree_root,
-                                       obj->objnode_tree_height);
-               tmem_objnode_free(obj->objnode_tree_root);
-               obj->objnode_tree_height = 0;
-       }
-       obj->objnode_tree_root = NULL;
-       (*tmem_pamops.free_obj)(obj->pool, obj);
-}
-
-/*
- * Tmem is operated on by a set of well-defined actions:
- * "put", "get", "flush", "flush_object", "new pool" and "destroy pool".
- * (The tmem ABI allows for subpages and exchanges but these operations
- * are not included in this implementation.)
- *
- * These "tmem core" operations are implemented in the following functions.
- */
-
-/*
- * "Put" a page, e.g. copy a page from the kernel into newly allocated
- * PAM space (if such space is available).  Tmem_put is complicated by
- * a corner case: What if a page with matching handle already exists in
- * tmem?  To guarantee coherency, one of two actions is necessary: Either
- * the data for the page must be overwritten, or the page must be
- * "flushed" so that the data is not accessible to a subsequent "get".
- * Since these "duplicate puts" are relatively rare, this implementation
- * always flushes for simplicity.
- */
-int tmem_put(struct tmem_pool *pool, struct tmem_oid *oidp, uint32_t index,
-               char *data, size_t size, bool raw, int ephemeral)
-{
-       struct tmem_obj *obj = NULL, *objfound = NULL, *objnew = NULL;
-       void *pampd = NULL, *pampd_del = NULL;
-       int ret = -ENOMEM;
-       struct tmem_hashbucket *hb;
-
-       hb = &pool->hashbucket[tmem_oid_hash(oidp)];
-       spin_lock(&hb->lock);
-       obj = objfound = tmem_obj_find(hb, oidp);
-       if (obj != NULL) {
-               pampd = tmem_pampd_lookup_in_obj(objfound, index);
-               if (pampd != NULL) {
-                       /* if found, is a dup put, flush the old one */
-                       pampd_del = tmem_pampd_delete_from_obj(obj, index);
-                       BUG_ON(pampd_del != pampd);
-                       (*tmem_pamops.free)(pampd, pool, oidp, index, true);
-                       if (obj->pampd_count == 0) {
-                               objnew = obj;
-                               objfound = NULL;
-                       }
-                       pampd = NULL;
-               }
-       } else {
-               obj = objnew = (*tmem_hostops.obj_alloc)(pool);
-               if (unlikely(obj == NULL)) {
-                       ret = -ENOMEM;
-                       goto out;
-               }
-               tmem_obj_init(obj, hb, pool, oidp);
-       }
-       BUG_ON(obj == NULL);
-       BUG_ON(((objnew != obj) && (objfound != obj)) || (objnew == objfound));
-       pampd = (*tmem_pamops.create)(data, size, raw, ephemeral,
-                                       obj->pool, &obj->oid, index);
-       if (unlikely(pampd == NULL))
-               goto free;
-       ret = tmem_pampd_add_to_obj(obj, index, pampd);
-       if (unlikely(ret == -ENOMEM))
-               /* may have partially built objnode tree ("stump") */
-               goto delete_and_free;
-       goto out;
-
-delete_and_free:
-       (void)tmem_pampd_delete_from_obj(obj, index);
-free:
-       if (pampd)
-               (*tmem_pamops.free)(pampd, pool, NULL, 0, true);
-       if (objnew) {
-               tmem_obj_free(objnew, hb);
-               (*tmem_hostops.obj_free)(objnew, pool);
-       }
-out:
-       spin_unlock(&hb->lock);
-       return ret;
-}
-
-void *tmem_localify_get_pampd(struct tmem_pool *pool, struct tmem_oid *oidp,
-                               uint32_t index, struct tmem_obj **ret_obj,
-                               void **saved_hb)
-{
-       struct tmem_hashbucket *hb;
-       struct tmem_obj *obj = NULL;
-       void *pampd = NULL;
-
-       hb = &pool->hashbucket[tmem_oid_hash(oidp)];
-       spin_lock(&hb->lock);
-       obj = tmem_obj_find(hb, oidp);
-       if (likely(obj != NULL))
-               pampd = tmem_pampd_lookup_in_obj(obj, index);
-       *ret_obj = obj;
-       *saved_hb = (void *)hb;
-       /* note, hashbucket remains locked */
-       return pampd;
-}
-
-void tmem_localify_finish(struct tmem_obj *obj, uint32_t index,
-                         void *pampd, void *saved_hb, bool delete)
-{
-       struct tmem_hashbucket *hb = (struct tmem_hashbucket *)saved_hb;
-
-       BUG_ON(!spin_is_locked(&hb->lock));
-       if (pampd != NULL) {
-               BUG_ON(obj == NULL);
-               (void)tmem_pampd_replace_in_obj(obj, index, pampd, 1);
-       } else if (delete) {
-               BUG_ON(obj == NULL);
-               (void)tmem_pampd_delete_from_obj(obj, index);
-       }
-       spin_unlock(&hb->lock);
-}
-
-static int tmem_repatriate(void **ppampd, struct tmem_hashbucket *hb,
-                               struct tmem_pool *pool, struct tmem_oid *oidp,
-                               uint32_t index, bool free, char *data)
-{
-       void *old_pampd = *ppampd, *new_pampd = NULL;
-       bool intransit = false;
-       int ret = 0;
-
-
-       if (!is_ephemeral(pool))
-               new_pampd = (*tmem_pamops.repatriate_preload)(
-                               old_pampd, pool, oidp, index, &intransit);
-       if (intransit)
-               ret = -EAGAIN;
-       else if (new_pampd != NULL)
-               *ppampd = new_pampd;
-       /* must release the hb->lock else repatriate can't sleep */
-       spin_unlock(&hb->lock);
-       if (!intransit)
-               ret = (*tmem_pamops.repatriate)(old_pampd, new_pampd, pool,
-                                               oidp, index, free, data);
-       return ret;
-}
-
-/*
- * "Get" a page, e.g. if one can be found, copy the tmem page with the
- * matching handle from PAM space to the kernel.  By tmem definition,
- * when a "get" is successful on an ephemeral page, the page is "flushed",
- * and when a "get" is successful on a persistent page, the page is retained
- * in tmem.  Note that to preserve
- * coherency, "get" can never be skipped if tmem contains the data.
- * That is, if a get is done with a certain handle and fails, any
- * subsequent "get" must also fail (unless of course there is a
- * "put" done with the same handle).
-
- */
-int tmem_get(struct tmem_pool *pool, struct tmem_oid *oidp, uint32_t index,
-               char *data, size_t *size, bool raw, int get_and_free)
-{
-       struct tmem_obj *obj;
-       void *pampd;
-       bool ephemeral = is_ephemeral(pool);
-       int ret = -1;
-       struct tmem_hashbucket *hb;
-       bool free = (get_and_free == 1) || ((get_and_free == 0) && ephemeral);
-       bool lock_held = 0;
-       void **ppampd;
-
-again:
-       hb = &pool->hashbucket[tmem_oid_hash(oidp)];
-       spin_lock(&hb->lock);
-       lock_held = 1;
-       obj = tmem_obj_find(hb, oidp);
-       if (obj == NULL)
-               goto out;
-       ppampd = __tmem_pampd_lookup_in_obj(obj, index);
-       if (ppampd == NULL)
-               goto out;
-       if (tmem_pamops.is_remote(*ppampd)) {
-               ret = tmem_repatriate(ppampd, hb, pool, oidp,
-                                       index, free, data);
-               lock_held = 0; /* note hb->lock has been unlocked */
-               if (ret == -EAGAIN) {
-                       /* rare I think, but should cond_resched()??? */
-                       usleep_range(10, 1000);
-                       goto again;
-               } else if (ret != 0) {
-                       if (ret != -ENOENT)
-                               pr_err("UNTESTED case in tmem_get, ret=%d\n",
-                                               ret);
-                       ret = -1;
-                       goto out;
-               }
-               goto out;
-       }
-       if (free)
-               pampd = tmem_pampd_delete_from_obj(obj, index);
-       else
-               pampd = tmem_pampd_lookup_in_obj(obj, index);
-       if (pampd == NULL)
-               goto out;
-       if (free) {
-               if (obj->pampd_count == 0) {
-                       tmem_obj_free(obj, hb);
-                       (*tmem_hostops.obj_free)(obj, pool);
-                       obj = NULL;
-               }
-       }
-       if (free)
-               ret = (*tmem_pamops.get_data_and_free)(
-                               data, size, raw, pampd, pool, oidp, index);
-       else
-               ret = (*tmem_pamops.get_data)(
-                               data, size, raw, pampd, pool, oidp, index);
-       if (ret < 0)
-               goto out;
-       ret = 0;
-out:
-       if (lock_held)
-               spin_unlock(&hb->lock);
-       return ret;
-}
-
-/*
- * If a page in tmem matches the handle, "flush" this page from tmem such
- * that any subsequent "get" does not succeed (unless, of course, there
- * was another "put" with the same handle).
- */
-int tmem_flush_page(struct tmem_pool *pool,
-                               struct tmem_oid *oidp, uint32_t index)
-{
-       struct tmem_obj *obj;
-       void *pampd;
-       int ret = -1;
-       struct tmem_hashbucket *hb;
-
-       hb = &pool->hashbucket[tmem_oid_hash(oidp)];
-       spin_lock(&hb->lock);
-       obj = tmem_obj_find(hb, oidp);
-       if (obj == NULL)
-               goto out;
-       pampd = tmem_pampd_delete_from_obj(obj, index);
-       if (pampd == NULL)
-               goto out;
-       (*tmem_pamops.free)(pampd, pool, oidp, index, true);
-       if (obj->pampd_count == 0) {
-               tmem_obj_free(obj, hb);
-               (*tmem_hostops.obj_free)(obj, pool);
-       }
-       ret = 0;
-
-out:
-       spin_unlock(&hb->lock);
-       return ret;
-}
-
-/*
- * If a page in tmem matches the handle, replace the page so that any
- * subsequent "get" gets the new page.  Returns the new page if
- * there was a page to replace, else returns NULL.
- */
-int tmem_replace(struct tmem_pool *pool, struct tmem_oid *oidp,
-                       uint32_t index, void *new_pampd)
-{
-       struct tmem_obj *obj;
-       int ret = -1;
-       struct tmem_hashbucket *hb;
-
-       hb = &pool->hashbucket[tmem_oid_hash(oidp)];
-       spin_lock(&hb->lock);
-       obj = tmem_obj_find(hb, oidp);
-       if (obj == NULL)
-               goto out;
-       new_pampd = tmem_pampd_replace_in_obj(obj, index, new_pampd, 0);
-       ret = (*tmem_pamops.replace_in_obj)(new_pampd, obj);
-out:
-       spin_unlock(&hb->lock);
-       return ret;
-}
-
-/*
- * "Flush" all pages in tmem matching this oid.
- */
-int tmem_flush_object(struct tmem_pool *pool, struct tmem_oid *oidp)
-{
-       struct tmem_obj *obj;
-       struct tmem_hashbucket *hb;
-       int ret = -1;
-
-       hb = &pool->hashbucket[tmem_oid_hash(oidp)];
-       spin_lock(&hb->lock);
-       obj = tmem_obj_find(hb, oidp);
-       if (obj == NULL)
-               goto out;
-       tmem_pampd_destroy_all_in_obj(obj);
-       tmem_obj_free(obj, hb);
-       (*tmem_hostops.obj_free)(obj, pool);
-       ret = 0;
-
-out:
-       spin_unlock(&hb->lock);
-       return ret;
-}
-
-/*
- * "Flush" all pages (and tmem_objs) from this tmem_pool and disable
- * all subsequent access to this tmem_pool.
- */
-int tmem_destroy_pool(struct tmem_pool *pool)
-{
-       int ret = -1;
-
-       if (pool == NULL)
-               goto out;
-       tmem_pool_flush(pool, 1);
-       ret = 0;
-out:
-       return ret;
-}
-
-static LIST_HEAD(tmem_global_pool_list);
-
-/*
- * Create a new tmem_pool with the provided flag and return
- * a pool id provided by the tmem host implementation.
- */
-void tmem_new_pool(struct tmem_pool *pool, uint32_t flags)
-{
-       int persistent = flags & TMEM_POOL_PERSIST;
-       int shared = flags & TMEM_POOL_SHARED;
-       struct tmem_hashbucket *hb = &pool->hashbucket[0];
-       int i;
-
-       for (i = 0; i < TMEM_HASH_BUCKETS; i++, hb++) {
-               hb->obj_rb_root = RB_ROOT;
-               spin_lock_init(&hb->lock);
-       }
-       INIT_LIST_HEAD(&pool->pool_list);
-       atomic_set(&pool->obj_count, 0);
-       SET_SENTINEL(pool, POOL);
-       list_add_tail(&pool->pool_list, &tmem_global_pool_list);
-       pool->persistent = persistent;
-       pool->shared = shared;
-}
diff --git a/drivers/staging/ramster/tmem.h b/drivers/staging/ramster/tmem.h
deleted file mode 100644 (file)
index 47f1918..0000000
+++ /dev/null
@@ -1,244 +0,0 @@
-/*
- * tmem.h
- *
- * Transcendent memory
- *
- * Copyright (c) 2009-2011, Dan Magenheimer, Oracle Corp.
- */
-
-#ifndef _TMEM_H_
-#define _TMEM_H_
-
-#include <linux/highmem.h>
-#include <linux/hash.h>
-#include <linux/atomic.h>
-
-/*
- * These are pre-defined by the Xen<->Linux ABI
- */
-#define TMEM_PUT_PAGE                  4
-#define TMEM_GET_PAGE                  5
-#define TMEM_FLUSH_PAGE                        6
-#define TMEM_FLUSH_OBJECT              7
-#define TMEM_POOL_PERSIST              1
-#define TMEM_POOL_SHARED               2
-#define TMEM_POOL_PRECOMPRESSED                4
-#define TMEM_POOL_PAGESIZE_SHIFT       4
-#define TMEM_POOL_PAGESIZE_MASK                0xf
-#define TMEM_POOL_RESERVED_BITS                0x00ffff00
-
-/*
- * sentinels have proven very useful for debugging but can be removed
- * or disabled before final merge.
- */
-#define SENTINELS
-#ifdef SENTINELS
-#define DECL_SENTINEL uint32_t sentinel;
-#define SET_SENTINEL(_x, _y) (_x->sentinel = _y##_SENTINEL)
-#define INVERT_SENTINEL(_x, _y) (_x->sentinel = ~_y##_SENTINEL)
-#define ASSERT_SENTINEL(_x, _y) WARN_ON(_x->sentinel != _y##_SENTINEL)
-#define ASSERT_INVERTED_SENTINEL(_x, _y) WARN_ON(_x->sentinel != ~_y##_SENTINEL)
-#else
-#define DECL_SENTINEL
-#define SET_SENTINEL(_x, _y) do { } while (0)
-#define INVERT_SENTINEL(_x, _y) do { } while (0)
-#define ASSERT_SENTINEL(_x, _y) do { } while (0)
-#define ASSERT_INVERTED_SENTINEL(_x, _y) do { } while (0)
-#endif
-
-#define ASSERT_SPINLOCK(_l)    WARN_ON(!spin_is_locked(_l))
-
-/*
- * A pool is the highest-level data structure managed by tmem and
- * usually corresponds to a large independent set of pages such as
- * a filesystem.  Each pool has an id, and certain attributes and counters.
- * It also contains a set of hash buckets, each of which contains an rbtree
- * of objects and a lock to manage concurrency within the pool.
- */
-
-#define TMEM_HASH_BUCKET_BITS  8
-#define TMEM_HASH_BUCKETS      (1<<TMEM_HASH_BUCKET_BITS)
-
-struct tmem_hashbucket {
-       struct rb_root obj_rb_root;
-       spinlock_t lock;
-};
-
-struct tmem_pool {
-       void *client; /* "up" for some clients, avoids table lookup */
-       struct list_head pool_list;
-       uint32_t pool_id;
-       bool persistent;
-       bool shared;
-       atomic_t obj_count;
-       atomic_t refcount;
-       struct tmem_hashbucket hashbucket[TMEM_HASH_BUCKETS];
-       DECL_SENTINEL
-};
-
-#define is_persistent(_p)  (_p->persistent)
-#define is_ephemeral(_p)   (!(_p->persistent))
-
-/*
- * An object id ("oid") is large: 192-bits (to ensure, for example, files
- * in a modern filesystem can be uniquely identified).
- */
-
-struct tmem_oid {
-       uint64_t oid[3];
-};
-
-struct tmem_xhandle {
-       uint8_t client_id;
-       uint8_t xh_data_cksum;
-       uint16_t xh_data_size;
-       uint16_t pool_id;
-       struct tmem_oid oid;
-       uint32_t index;
-       void *extra;
-};
-
-static inline struct tmem_xhandle tmem_xhandle_fill(uint16_t client_id,
-                                       struct tmem_pool *pool,
-                                       struct tmem_oid *oidp,
-                                       uint32_t index)
-{
-       struct tmem_xhandle xh;
-       xh.client_id = client_id;
-       xh.xh_data_cksum = (uint8_t)-1;
-       xh.xh_data_size = (uint16_t)-1;
-       xh.pool_id = pool->pool_id;
-       xh.oid = *oidp;
-       xh.index = index;
-       return xh;
-}
-
-static inline void tmem_oid_set_invalid(struct tmem_oid *oidp)
-{
-       oidp->oid[0] = oidp->oid[1] = oidp->oid[2] = -1UL;
-}
-
-static inline bool tmem_oid_valid(struct tmem_oid *oidp)
-{
-       return oidp->oid[0] != -1UL || oidp->oid[1] != -1UL ||
-               oidp->oid[2] != -1UL;
-}
-
-static inline int tmem_oid_compare(struct tmem_oid *left,
-                                       struct tmem_oid *right)
-{
-       int ret;
-
-       if (left->oid[2] == right->oid[2]) {
-               if (left->oid[1] == right->oid[1]) {
-                       if (left->oid[0] == right->oid[0])
-                               ret = 0;
-                       else if (left->oid[0] < right->oid[0])
-                               ret = -1;
-                       else
-                               return 1;
-               } else if (left->oid[1] < right->oid[1])
-                       ret = -1;
-               else
-                       ret = 1;
-       } else if (left->oid[2] < right->oid[2])
-               ret = -1;
-       else
-               ret = 1;
-       return ret;
-}
-
-static inline unsigned tmem_oid_hash(struct tmem_oid *oidp)
-{
-       return hash_long(oidp->oid[0] ^ oidp->oid[1] ^ oidp->oid[2],
-                               TMEM_HASH_BUCKET_BITS);
-}
-
-/*
- * A tmem_obj contains an identifier (oid), pointers to the parent
- * pool and the rb_tree to which it belongs, counters, and an ordered
- * set of pampds, structured in a radix-tree-like tree.  The intermediate
- * nodes of the tree are called tmem_objnodes.
- */
-
-struct tmem_objnode;
-
-struct tmem_obj {
-       struct tmem_oid oid;
-       struct tmem_pool *pool;
-       struct rb_node rb_tree_node;
-       struct tmem_objnode *objnode_tree_root;
-       unsigned int objnode_tree_height;
-       unsigned long objnode_count;
-       long pampd_count;
-       /* for current design of ramster, all pages belonging to
-        * an object reside on the same remotenode and extra is
-        * used to record the number of the remotenode so a
-        * flush-object operation can specify it */
-       void *extra; /* for use by pampd implementation */
-       DECL_SENTINEL
-};
-
-#define OBJNODE_TREE_MAP_SHIFT 6
-#define OBJNODE_TREE_MAP_SIZE (1UL << OBJNODE_TREE_MAP_SHIFT)
-#define OBJNODE_TREE_MAP_MASK (OBJNODE_TREE_MAP_SIZE-1)
-#define OBJNODE_TREE_INDEX_BITS (8 /* CHAR_BIT */ * sizeof(unsigned long))
-#define OBJNODE_TREE_MAX_PATH \
-               (OBJNODE_TREE_INDEX_BITS/OBJNODE_TREE_MAP_SHIFT + 2)
-
-struct tmem_objnode {
-       struct tmem_obj *obj;
-       DECL_SENTINEL
-       void *slots[OBJNODE_TREE_MAP_SIZE];
-       unsigned int slots_in_use;
-};
-
-/* pampd abstract datatype methods provided by the PAM implementation */
-struct tmem_pamops {
-       void *(*create)(char *, size_t, bool, int,
-                       struct tmem_pool *, struct tmem_oid *, uint32_t);
-       int (*get_data)(char *, size_t *, bool, void *, struct tmem_pool *,
-                               struct tmem_oid *, uint32_t);
-       int (*get_data_and_free)(char *, size_t *, bool, void *,
-                               struct tmem_pool *, struct tmem_oid *,
-                               uint32_t);
-       void (*free)(void *, struct tmem_pool *,
-                               struct tmem_oid *, uint32_t, bool);
-       void (*free_obj)(struct tmem_pool *, struct tmem_obj *);
-       bool (*is_remote)(void *);
-       void *(*repatriate_preload)(void *, struct tmem_pool *,
-                                       struct tmem_oid *, uint32_t, bool *);
-       int (*repatriate)(void *, void *, struct tmem_pool *,
-                               struct tmem_oid *, uint32_t, bool, void *);
-       void (*new_obj)(struct tmem_obj *);
-       int (*replace_in_obj)(void *, struct tmem_obj *);
-};
-extern void tmem_register_pamops(struct tmem_pamops *m);
-
-/* memory allocation methods provided by the host implementation */
-struct tmem_hostops {
-       struct tmem_obj *(*obj_alloc)(struct tmem_pool *);
-       void (*obj_free)(struct tmem_obj *, struct tmem_pool *);
-       struct tmem_objnode *(*objnode_alloc)(struct tmem_pool *);
-       void (*objnode_free)(struct tmem_objnode *, struct tmem_pool *);
-};
-extern void tmem_register_hostops(struct tmem_hostops *m);
-
-/* core tmem accessor functions */
-extern int tmem_put(struct tmem_pool *, struct tmem_oid *, uint32_t index,
-                       char *, size_t, bool, int);
-extern int tmem_get(struct tmem_pool *, struct tmem_oid *, uint32_t index,
-                       char *, size_t *, bool, int);
-extern int tmem_replace(struct tmem_pool *, struct tmem_oid *, uint32_t index,
-                       void *);
-extern void *tmem_localify_get_pampd(struct tmem_pool *, struct tmem_oid *,
-                                  uint32_t index, struct tmem_obj **,
-                                  void **);
-extern void tmem_localify_finish(struct tmem_obj *, uint32_t index,
-                                void *, void *, bool);
-extern int tmem_flush_page(struct tmem_pool *, struct tmem_oid *,
-                       uint32_t index);
-extern int tmem_flush_object(struct tmem_pool *, struct tmem_oid *);
-extern int tmem_destroy_pool(struct tmem_pool *);
-extern void tmem_new_pool(struct tmem_pool *, uint32_t);
-#endif /* _TMEM_H */
diff --git a/drivers/staging/ramster/zcache-main.c b/drivers/staging/ramster/zcache-main.c
deleted file mode 100644 (file)
index 851e94a..0000000
+++ /dev/null
@@ -1,3264 +0,0 @@
-/*
- * zcache.c
- *
- * Copyright (c) 2010-2012, Dan Magenheimer, Oracle Corp.
- * Copyright (c) 2010,2011, Nitin Gupta
- *
- * Zcache provides an in-kernel "host implementation" for transcendent memory
- * and, thus indirectly, for cleancache and frontswap.  Zcache includes two
- * page-accessible memory [1] interfaces, both utilizing lzo1x compression:
- * 1) "compression buddies" ("zbud") is used for ephemeral pages
- * 2) xvmalloc is used for persistent pages.
- * Xvmalloc (based on the TLSF allocator) has very low fragmentation
- * so maximizes space efficiency, while zbud allows pairs (and potentially,
- * in the future, more than a pair of) compressed pages to be closely linked
- * so that reclaiming can be done via the kernel's physical-page-oriented
- * "shrinker" interface.
- *
- * [1] For a definition of page-accessible memory (aka PAM), see:
- *   http://marc.info/?l=linux-mm&m=127811271605009
- *  RAMSTER TODO:
- *   - handle remotifying of buddied pages (see zbud_remotify_zbpg)
- *   - kernel boot params: nocleancache/nofrontswap don't always work?!?
- */
-
-#include <linux/module.h>
-#include <linux/cpu.h>
-#include <linux/highmem.h>
-#include <linux/list.h>
-#include <linux/lzo.h>
-#include <linux/slab.h>
-#include <linux/spinlock.h>
-#include <linux/types.h>
-#include <linux/atomic.h>
-#include <linux/math64.h>
-#include "tmem.h"
-#include "zcache.h"
-#include "ramster.h"
-#include "cluster/tcp.h"
-
-#include "../zram/xvmalloc.h" /* if built in drivers/staging */
-
-#define        RAMSTER_TESTING
-
-#if (!defined(CONFIG_CLEANCACHE) && !defined(CONFIG_FRONTSWAP))
-#error "ramster is useless without CONFIG_CLEANCACHE or CONFIG_FRONTSWAP"
-#endif
-#ifdef CONFIG_CLEANCACHE
-#include <linux/cleancache.h>
-#endif
-#ifdef CONFIG_FRONTSWAP
-#include <linux/frontswap.h>
-#endif
-
-enum ramster_remotify_op {
-       RAMSTER_REMOTIFY_EPH_PUT,
-       RAMSTER_REMOTIFY_PERS_PUT,
-       RAMSTER_REMOTIFY_FLUSH_PAGE,
-       RAMSTER_REMOTIFY_FLUSH_OBJ,
-       RAMSTER_INTRANSIT_PERS
-};
-
-struct ramster_remotify_hdr {
-       enum ramster_remotify_op op;
-       struct list_head list;
-};
-
-#define ZBH_SENTINEL  0x43214321
-#define ZBPG_SENTINEL  0xdeadbeef
-
-#define ZBUD_MAX_BUDS 2
-
-struct zbud_hdr {
-       struct ramster_remotify_hdr rem_op;
-       uint16_t client_id;
-       uint16_t pool_id;
-       struct tmem_oid oid;
-       uint32_t index;
-       uint16_t size; /* compressed size in bytes, zero means unused */
-       DECL_SENTINEL
-};
-
-#define ZVH_SENTINEL  0x43214321
-static const int zv_max_page_size = (PAGE_SIZE / 8) * 7;
-
-struct zv_hdr {
-       struct ramster_remotify_hdr rem_op;
-       uint16_t client_id;
-       uint16_t pool_id;
-       struct tmem_oid oid;
-       uint32_t index;
-       DECL_SENTINEL
-};
-
-struct flushlist_node {
-       struct ramster_remotify_hdr rem_op;
-       struct tmem_xhandle xh;
-};
-
-union {
-       struct ramster_remotify_hdr rem_op;
-       struct zv_hdr zv;
-       struct zbud_hdr zbud;
-       struct flushlist_node flist;
-} remotify_list_node;
-
-static LIST_HEAD(zcache_rem_op_list);
-static DEFINE_SPINLOCK(zcache_rem_op_list_lock);
-
-#if 0
-/* this is more aggressive but may cause other problems? */
-#define ZCACHE_GFP_MASK        (GFP_ATOMIC | __GFP_NORETRY | __GFP_NOWARN)
-#else
-#define ZCACHE_GFP_MASK \
-       (__GFP_FS | __GFP_NORETRY | __GFP_NOWARN | __GFP_NOMEMALLOC)
-#endif
-
-#define MAX_POOLS_PER_CLIENT 16
-
-#define MAX_CLIENTS 16
-#define LOCAL_CLIENT ((uint16_t)-1)
-
-MODULE_LICENSE("GPL");
-
-struct zcache_client {
-       struct tmem_pool *tmem_pools[MAX_POOLS_PER_CLIENT];
-       struct xv_pool *xvpool;
-       bool allocated;
-       atomic_t refcount;
-};
-
-static struct zcache_client zcache_host;
-static struct zcache_client zcache_clients[MAX_CLIENTS];
-
-static inline uint16_t get_client_id_from_client(struct zcache_client *cli)
-{
-       BUG_ON(cli == NULL);
-       if (cli == &zcache_host)
-               return LOCAL_CLIENT;
-       return cli - &zcache_clients[0];
-}
-
-static inline bool is_local_client(struct zcache_client *cli)
-{
-       return cli == &zcache_host;
-}
-
-/**********
- * Compression buddies ("zbud") provides for packing two (or, possibly
- * in the future, more) compressed ephemeral pages into a single "raw"
- * (physical) page and tracking them with data structures so that
- * the raw pages can be easily reclaimed.
- *
- * A zbud page ("zbpg") is an aligned page containing a list_head,
- * a lock, and two "zbud headers".  The remainder of the physical
- * page is divided up into aligned 64-byte "chunks" which contain
- * the compressed data for zero, one, or two zbuds.  Each zbpg
- * resides on: (1) an "unused list" if it has no zbuds; (2) a
- * "buddied" list if it is fully populated  with two zbuds; or
- * (3) one of PAGE_SIZE/64 "unbuddied" lists indexed by how many chunks
- * the one unbuddied zbud uses.  The data inside a zbpg cannot be
- * read or written unless the zbpg's lock is held.
- */
-
-struct zbud_page {
-       struct list_head bud_list;
-       spinlock_t lock;
-       struct zbud_hdr buddy[ZBUD_MAX_BUDS];
-       DECL_SENTINEL
-       /* followed by NUM_CHUNK aligned CHUNK_SIZE-byte chunks */
-};
-
-#define CHUNK_SHIFT    6
-#define CHUNK_SIZE     (1 << CHUNK_SHIFT)
-#define CHUNK_MASK     (~(CHUNK_SIZE-1))
-#define NCHUNKS                (((PAGE_SIZE - sizeof(struct zbud_page)) & \
-                               CHUNK_MASK) >> CHUNK_SHIFT)
-#define MAX_CHUNK      (NCHUNKS-1)
-
-static struct {
-       struct list_head list;
-       unsigned count;
-} zbud_unbuddied[NCHUNKS];
-/* list N contains pages with N chunks USED and NCHUNKS-N unused */
-/* element 0 is never used but optimizing that isn't worth it */
-static unsigned long zbud_cumul_chunk_counts[NCHUNKS];
-
-struct list_head zbud_buddied_list;
-static unsigned long zcache_zbud_buddied_count;
-
-/* protects the buddied list and all unbuddied lists */
-static DEFINE_SPINLOCK(zbud_budlists_spinlock);
-
-static atomic_t zcache_zbud_curr_raw_pages;
-static atomic_t zcache_zbud_curr_zpages;
-static unsigned long zcache_zbud_curr_zbytes;
-static unsigned long zcache_zbud_cumul_zpages;
-static unsigned long zcache_zbud_cumul_zbytes;
-static unsigned long zcache_compress_poor;
-static unsigned long zcache_policy_percent_exceeded;
-static unsigned long zcache_mean_compress_poor;
-
-/*
- * RAMster counters
- * - Remote pages are pages with a local pampd but the data is remote
- * - Foreign pages are pages stored locally but belonging to another node
- */
-static atomic_t ramster_remote_pers_pages = ATOMIC_INIT(0);
-static unsigned long ramster_pers_remotify_enable;
-static unsigned long ramster_eph_remotify_enable;
-static unsigned long ramster_eph_pages_remoted;
-static unsigned long ramster_eph_pages_remote_failed;
-static unsigned long ramster_pers_pages_remoted;
-static unsigned long ramster_pers_pages_remote_failed;
-static unsigned long ramster_pers_pages_remote_nomem;
-static unsigned long ramster_remote_objects_flushed;
-static unsigned long ramster_remote_object_flushes_failed;
-static unsigned long ramster_remote_pages_flushed;
-static unsigned long ramster_remote_page_flushes_failed;
-static unsigned long ramster_remote_eph_pages_succ_get;
-static unsigned long ramster_remote_pers_pages_succ_get;
-static unsigned long ramster_remote_eph_pages_unsucc_get;
-static unsigned long ramster_remote_pers_pages_unsucc_get;
-static atomic_t ramster_curr_flnode_count = ATOMIC_INIT(0);
-static unsigned long ramster_curr_flnode_count_max;
-static atomic_t ramster_foreign_eph_pampd_count = ATOMIC_INIT(0);
-static unsigned long ramster_foreign_eph_pampd_count_max;
-static atomic_t ramster_foreign_pers_pampd_count = ATOMIC_INIT(0);
-static unsigned long ramster_foreign_pers_pampd_count_max;
-
-/* forward references */
-static void *zcache_get_free_page(void);
-static void zcache_free_page(void *p);
-
-/*
- * zbud helper functions
- */
-
-static inline unsigned zbud_max_buddy_size(void)
-{
-       return MAX_CHUNK << CHUNK_SHIFT;
-}
-
-static inline unsigned zbud_size_to_chunks(unsigned size)
-{
-       BUG_ON(size == 0 || size > zbud_max_buddy_size());
-       return (size + CHUNK_SIZE - 1) >> CHUNK_SHIFT;
-}
-
-static inline int zbud_budnum(struct zbud_hdr *zh)
-{
-       unsigned offset = (unsigned long)zh & (PAGE_SIZE - 1);
-       struct zbud_page *zbpg = NULL;
-       unsigned budnum = -1U;
-       int i;
-
-       for (i = 0; i < ZBUD_MAX_BUDS; i++)
-               if (offset == offsetof(typeof(*zbpg), buddy[i])) {
-                       budnum = i;
-                       break;
-               }
-       BUG_ON(budnum == -1U);
-       return budnum;
-}
-
-static char *zbud_data(struct zbud_hdr *zh, unsigned size)
-{
-       struct zbud_page *zbpg;
-       char *p;
-       unsigned budnum;
-
-       ASSERT_SENTINEL(zh, ZBH);
-       budnum = zbud_budnum(zh);
-       BUG_ON(size == 0 || size > zbud_max_buddy_size());
-       zbpg = container_of(zh, struct zbud_page, buddy[budnum]);
-       ASSERT_SPINLOCK(&zbpg->lock);
-       p = (char *)zbpg;
-       if (budnum == 0)
-               p += ((sizeof(struct zbud_page) + CHUNK_SIZE - 1) &
-                                                       CHUNK_MASK);
-       else if (budnum == 1)
-               p += PAGE_SIZE - ((size + CHUNK_SIZE - 1) & CHUNK_MASK);
-       return p;
-}
-
-static void zbud_copy_from_pampd(char *data, size_t *size, struct zbud_hdr *zh)
-{
-       struct zbud_page *zbpg;
-       char *p;
-       unsigned budnum;
-
-       ASSERT_SENTINEL(zh, ZBH);
-       budnum = zbud_budnum(zh);
-       zbpg = container_of(zh, struct zbud_page, buddy[budnum]);
-       spin_lock(&zbpg->lock);
-       BUG_ON(zh->size > *size);
-       p = (char *)zbpg;
-       if (budnum == 0)
-               p += ((sizeof(struct zbud_page) + CHUNK_SIZE - 1) &
-                                                       CHUNK_MASK);
-       else if (budnum == 1)
-               p += PAGE_SIZE - ((zh->size + CHUNK_SIZE - 1) & CHUNK_MASK);
-       /* client should be filled in by caller */
-       memcpy(data, p, zh->size);
-       *size = zh->size;
-       spin_unlock(&zbpg->lock);
-}
-
-/*
- * zbud raw page management
- */
-
-static struct zbud_page *zbud_alloc_raw_page(void)
-{
-       struct zbud_page *zbpg = NULL;
-       struct zbud_hdr *zh0, *zh1;
-               zbpg = zcache_get_free_page();
-       if (likely(zbpg != NULL)) {
-               INIT_LIST_HEAD(&zbpg->bud_list);
-               zh0 = &zbpg->buddy[0]; zh1 = &zbpg->buddy[1];
-               spin_lock_init(&zbpg->lock);
-               atomic_inc(&zcache_zbud_curr_raw_pages);
-               INIT_LIST_HEAD(&zbpg->bud_list);
-               SET_SENTINEL(zbpg, ZBPG);
-               zh0->size = 0; zh1->size = 0;
-               tmem_oid_set_invalid(&zh0->oid);
-               tmem_oid_set_invalid(&zh1->oid);
-       }
-       return zbpg;
-}
-
-static void zbud_free_raw_page(struct zbud_page *zbpg)
-{
-       struct zbud_hdr *zh0 = &zbpg->buddy[0], *zh1 = &zbpg->buddy[1];
-
-       ASSERT_SENTINEL(zbpg, ZBPG);
-       BUG_ON(!list_empty(&zbpg->bud_list));
-       ASSERT_SPINLOCK(&zbpg->lock);
-       BUG_ON(zh0->size != 0 || tmem_oid_valid(&zh0->oid));
-       BUG_ON(zh1->size != 0 || tmem_oid_valid(&zh1->oid));
-       INVERT_SENTINEL(zbpg, ZBPG);
-       spin_unlock(&zbpg->lock);
-       atomic_dec(&zcache_zbud_curr_raw_pages);
-       zcache_free_page(zbpg);
-}
-
-/*
- * core zbud handling routines
- */
-
-static unsigned zbud_free(struct zbud_hdr *zh)
-{
-       unsigned size;
-
-       ASSERT_SENTINEL(zh, ZBH);
-       BUG_ON(!tmem_oid_valid(&zh->oid));
-       size = zh->size;
-       BUG_ON(zh->size == 0 || zh->size > zbud_max_buddy_size());
-       zh->size = 0;
-       tmem_oid_set_invalid(&zh->oid);
-       INVERT_SENTINEL(zh, ZBH);
-       zcache_zbud_curr_zbytes -= size;
-       atomic_dec(&zcache_zbud_curr_zpages);
-       return size;
-}
-
-static void zbud_free_and_delist(struct zbud_hdr *zh)
-{
-       unsigned chunks;
-       struct zbud_hdr *zh_other;
-       unsigned budnum = zbud_budnum(zh), size;
-       struct zbud_page *zbpg =
-               container_of(zh, struct zbud_page, buddy[budnum]);
-
-       /* FIXME, should be BUG_ON, pool destruction path doesn't disable
-        * interrupts tmem_destroy_pool()->tmem_pampd_destroy_all_in_obj()->
-        * tmem_objnode_node_destroy()-> zcache_pampd_free() */
-       WARN_ON(!irqs_disabled());
-       spin_lock(&zbpg->lock);
-       if (list_empty(&zbpg->bud_list)) {
-               /* ignore zombie page... see zbud_evict_pages() */
-               spin_unlock(&zbpg->lock);
-               return;
-       }
-       size = zbud_free(zh);
-       ASSERT_SPINLOCK(&zbpg->lock);
-       zh_other = &zbpg->buddy[(budnum == 0) ? 1 : 0];
-       if (zh_other->size == 0) { /* was unbuddied: unlist and free */
-               chunks = zbud_size_to_chunks(size) ;
-               spin_lock(&zbud_budlists_spinlock);
-               BUG_ON(list_empty(&zbud_unbuddied[chunks].list));
-               list_del_init(&zbpg->bud_list);
-               zbud_unbuddied[chunks].count--;
-               spin_unlock(&zbud_budlists_spinlock);
-               zbud_free_raw_page(zbpg);
-       } else { /* was buddied: move remaining buddy to unbuddied list */
-               chunks = zbud_size_to_chunks(zh_other->size) ;
-               spin_lock(&zbud_budlists_spinlock);
-               list_del_init(&zbpg->bud_list);
-               zcache_zbud_buddied_count--;
-               list_add_tail(&zbpg->bud_list, &zbud_unbuddied[chunks].list);
-               zbud_unbuddied[chunks].count++;
-               spin_unlock(&zbud_budlists_spinlock);
-               spin_unlock(&zbpg->lock);
-       }
-}
-
-static struct zbud_hdr *zbud_create(uint16_t client_id, uint16_t pool_id,
-                                       struct tmem_oid *oid,
-                                       uint32_t index, struct page *page,
-                                       void *cdata, unsigned size)
-{
-       struct zbud_hdr *zh0, *zh1, *zh = NULL;
-       struct zbud_page *zbpg = NULL, *ztmp;
-       unsigned nchunks;
-       char *to;
-       int i, found_good_buddy = 0;
-
-       nchunks = zbud_size_to_chunks(size) ;
-       for (i = MAX_CHUNK - nchunks + 1; i > 0; i--) {
-               spin_lock(&zbud_budlists_spinlock);
-               if (!list_empty(&zbud_unbuddied[i].list)) {
-                       list_for_each_entry_safe(zbpg, ztmp,
-                                   &zbud_unbuddied[i].list, bud_list) {
-                               if (spin_trylock(&zbpg->lock)) {
-                                       found_good_buddy = i;
-                                       goto found_unbuddied;
-                               }
-                       }
-               }
-               spin_unlock(&zbud_budlists_spinlock);
-       }
-       /* didn't find a good buddy, try allocating a new page */
-       zbpg = zbud_alloc_raw_page();
-       if (unlikely(zbpg == NULL))
-               goto out;
-       /* ok, have a page, now compress the data before taking locks */
-       spin_lock(&zbud_budlists_spinlock);
-       spin_lock(&zbpg->lock);
-       list_add_tail(&zbpg->bud_list, &zbud_unbuddied[nchunks].list);
-       zbud_unbuddied[nchunks].count++;
-       zh = &zbpg->buddy[0];
-       goto init_zh;
-
-found_unbuddied:
-       ASSERT_SPINLOCK(&zbpg->lock);
-       zh0 = &zbpg->buddy[0]; zh1 = &zbpg->buddy[1];
-       BUG_ON(!((zh0->size == 0) ^ (zh1->size == 0)));
-       if (zh0->size != 0) { /* buddy0 in use, buddy1 is vacant */
-               ASSERT_SENTINEL(zh0, ZBH);
-               zh = zh1;
-       } else if (zh1->size != 0) { /* buddy1 in use, buddy0 is vacant */
-               ASSERT_SENTINEL(zh1, ZBH);
-               zh = zh0;
-       } else
-               BUG();
-       list_del_init(&zbpg->bud_list);
-       zbud_unbuddied[found_good_buddy].count--;
-       list_add_tail(&zbpg->bud_list, &zbud_buddied_list);
-       zcache_zbud_buddied_count++;
-
-init_zh:
-       SET_SENTINEL(zh, ZBH);
-       zh->size = size;
-       zh->index = index;
-       zh->oid = *oid;
-       zh->pool_id = pool_id;
-       zh->client_id = client_id;
-       to = zbud_data(zh, size);
-       memcpy(to, cdata, size);
-       spin_unlock(&zbpg->lock);
-       spin_unlock(&zbud_budlists_spinlock);
-       zbud_cumul_chunk_counts[nchunks]++;
-       atomic_inc(&zcache_zbud_curr_zpages);
-       zcache_zbud_cumul_zpages++;
-       zcache_zbud_curr_zbytes += size;
-       zcache_zbud_cumul_zbytes += size;
-out:
-       return zh;
-}
-
-static int zbud_decompress(struct page *page, struct zbud_hdr *zh)
-{
-       struct zbud_page *zbpg;
-       unsigned budnum = zbud_budnum(zh);
-       size_t out_len = PAGE_SIZE;
-       char *to_va, *from_va;
-       unsigned size;
-       int ret = 0;
-
-       zbpg = container_of(zh, struct zbud_page, buddy[budnum]);
-       spin_lock(&zbpg->lock);
-       if (list_empty(&zbpg->bud_list)) {
-               /* ignore zombie page... see zbud_evict_pages() */
-               ret = -EINVAL;
-               goto out;
-       }
-       ASSERT_SENTINEL(zh, ZBH);
-       BUG_ON(zh->size == 0 || zh->size > zbud_max_buddy_size());
-       to_va = kmap_atomic(page, KM_USER0);
-       size = zh->size;
-       from_va = zbud_data(zh, size);
-       ret = lzo1x_decompress_safe(from_va, size, to_va, &out_len);
-       BUG_ON(ret != LZO_E_OK);
-       BUG_ON(out_len != PAGE_SIZE);
-       kunmap_atomic(to_va, KM_USER0);
-out:
-       spin_unlock(&zbpg->lock);
-       return ret;
-}
-
-/*
- * The following routines handle shrinking of ephemeral pages by evicting
- * pages "least valuable" first.
- */
-
-static unsigned long zcache_evicted_raw_pages;
-static unsigned long zcache_evicted_buddied_pages;
-static unsigned long zcache_evicted_unbuddied_pages;
-
-static struct tmem_pool *zcache_get_pool_by_id(uint16_t cli_id,
-                                               uint16_t poolid);
-static void zcache_put_pool(struct tmem_pool *pool);
-
-/*
- * Flush and free all zbuds in a zbpg, then free the pageframe
- */
-static void zbud_evict_zbpg(struct zbud_page *zbpg)
-{
-       struct zbud_hdr *zh;
-       int i, j;
-       uint32_t pool_id[ZBUD_MAX_BUDS], client_id[ZBUD_MAX_BUDS];
-       uint32_t index[ZBUD_MAX_BUDS];
-       struct tmem_oid oid[ZBUD_MAX_BUDS];
-       struct tmem_pool *pool;
-       unsigned long flags;
-
-       ASSERT_SPINLOCK(&zbpg->lock);
-       for (i = 0, j = 0; i < ZBUD_MAX_BUDS; i++) {
-               zh = &zbpg->buddy[i];
-               if (zh->size) {
-                       client_id[j] = zh->client_id;
-                       pool_id[j] = zh->pool_id;
-                       oid[j] = zh->oid;
-                       index[j] = zh->index;
-                       j++;
-               }
-       }
-       spin_unlock(&zbpg->lock);
-       for (i = 0; i < j; i++) {
-               pool = zcache_get_pool_by_id(client_id[i], pool_id[i]);
-               BUG_ON(pool == NULL);
-               local_irq_save(flags);
-               /* these flushes should dispose of any local storage */
-               tmem_flush_page(pool, &oid[i], index[i]);
-               local_irq_restore(flags);
-               zcache_put_pool(pool);
-       }
-}
-
-/*
- * Free nr pages.  This code is funky because we want to hold the locks
- * protecting various lists for as short a time as possible, and in some
- * circumstances the list may change asynchronously when the list lock is
- * not held.  In some cases we also trylock not only to avoid waiting on a
- * page in use by another cpu, but also to avoid potential deadlock due to
- * lock inversion.
- */
-static void zbud_evict_pages(int nr)
-{
-       struct zbud_page *zbpg;
-       int i, newly_unused_pages = 0;
-
-
-       /* now try freeing unbuddied pages, starting with least space avail */
-       for (i = 0; i < MAX_CHUNK; i++) {
-retry_unbud_list_i:
-               spin_lock_bh(&zbud_budlists_spinlock);
-               if (list_empty(&zbud_unbuddied[i].list)) {
-                       spin_unlock_bh(&zbud_budlists_spinlock);
-                       continue;
-               }
-               list_for_each_entry(zbpg, &zbud_unbuddied[i].list, bud_list) {
-                       if (unlikely(!spin_trylock(&zbpg->lock)))
-                               continue;
-                       zbud_unbuddied[i].count--;
-                       spin_unlock(&zbud_budlists_spinlock);
-                       zcache_evicted_unbuddied_pages++;
-                       /* want budlists unlocked when doing zbpg eviction */
-                       zbud_evict_zbpg(zbpg);
-                       newly_unused_pages++;
-                       local_bh_enable();
-                       if (--nr <= 0)
-                               goto evict_unused;
-                       goto retry_unbud_list_i;
-               }
-               spin_unlock_bh(&zbud_budlists_spinlock);
-       }
-
-       /* as a last resort, free buddied pages */
-retry_bud_list:
-       spin_lock_bh(&zbud_budlists_spinlock);
-       if (list_empty(&zbud_buddied_list)) {
-               spin_unlock_bh(&zbud_budlists_spinlock);
-               goto evict_unused;
-       }
-       list_for_each_entry(zbpg, &zbud_buddied_list, bud_list) {
-               if (unlikely(!spin_trylock(&zbpg->lock)))
-                       continue;
-               zcache_zbud_buddied_count--;
-               spin_unlock(&zbud_budlists_spinlock);
-               zcache_evicted_buddied_pages++;
-               /* want budlists unlocked when doing zbpg eviction */
-               zbud_evict_zbpg(zbpg);
-               newly_unused_pages++;
-               local_bh_enable();
-               if (--nr <= 0)
-                       goto evict_unused;
-               goto retry_bud_list;
-       }
-       spin_unlock_bh(&zbud_budlists_spinlock);
-
-evict_unused:
-       return;
-}
-
-static DEFINE_PER_CPU(unsigned char *, zcache_remoteputmem);
-
-static int zbud_remotify_zbud(struct tmem_xhandle *xh, char *data,
-                               size_t size)
-{
-       struct tmem_pool *pool;
-       int i, remotenode, ret = -1;
-       unsigned char cksum, *p;
-       unsigned long flags;
-
-       for (p = data, cksum = 0, i = 0; i < size; i++)
-               cksum += *p;
-       ret = ramster_remote_put(xh, data, size, true, &remotenode);
-       if (ret == 0) {
-               /* data was successfully remoted so change the local version
-                * to point to the remote node where it landed */
-               pool = zcache_get_pool_by_id(LOCAL_CLIENT, xh->pool_id);
-               BUG_ON(pool == NULL);
-               local_irq_save(flags);
-               /* tmem_replace will also free up any local space */
-               (void)tmem_replace(pool, &xh->oid, xh->index,
-                       pampd_make_remote(remotenode, size, cksum));
-               local_irq_restore(flags);
-               zcache_put_pool(pool);
-               ramster_eph_pages_remoted++;
-               ret = 0;
-       } else
-               ramster_eph_pages_remote_failed++;
-       return ret;
-}
-
-static int zbud_remotify_zbpg(struct zbud_page *zbpg)
-{
-       struct zbud_hdr *zh1, *zh2 = NULL;
-       struct tmem_xhandle xh1, xh2 = { 0 };
-       char *data1 = NULL, *data2 = NULL;
-       size_t size1 = 0, size2 = 0;
-       int ret = 0;
-       unsigned char *tmpmem = __get_cpu_var(zcache_remoteputmem);
-
-       ASSERT_SPINLOCK(&zbpg->lock);
-       if (zbpg->buddy[0].size == 0)
-               zh1 = &zbpg->buddy[1];
-       else if (zbpg->buddy[1].size == 0)
-               zh1 = &zbpg->buddy[0];
-       else {
-               zh1 = &zbpg->buddy[0];
-               zh2 = &zbpg->buddy[1];
-       }
-       /* don't remotify pages that are already remotified */
-       if (zh1->client_id != LOCAL_CLIENT)
-               zh1 = NULL;
-       if ((zh2 != NULL) && (zh2->client_id != LOCAL_CLIENT))
-               zh2 = NULL;
-
-       /* copy the data and metadata so can release lock */
-       if (zh1 != NULL) {
-               xh1.client_id = zh1->client_id;
-               xh1.pool_id = zh1->pool_id;
-               xh1.oid = zh1->oid;
-               xh1.index = zh1->index;
-               size1 = zh1->size;
-               data1 = zbud_data(zh1, size1);
-               memcpy(tmpmem, zbud_data(zh1, size1), size1);
-               data1 = tmpmem;
-               tmpmem += size1;
-       }
-       if (zh2 != NULL) {
-               xh2.client_id = zh2->client_id;
-               xh2.pool_id = zh2->pool_id;
-               xh2.oid = zh2->oid;
-               xh2.index = zh2->index;
-               size2 = zh2->size;
-               memcpy(tmpmem, zbud_data(zh2, size2), size2);
-               data2 = tmpmem;
-       }
-       spin_unlock(&zbpg->lock);
-       preempt_enable();
-
-       /* OK, no locks held anymore, remotify one or both zbuds */
-       if (zh1 != NULL)
-               ret = zbud_remotify_zbud(&xh1, data1, size1);
-       if (zh2 != NULL)
-               ret |= zbud_remotify_zbud(&xh2, data2, size2);
-       return ret;
-}
-
-void zbud_remotify_pages(int nr)
-{
-       struct zbud_page *zbpg;
-       int i, ret;
-
-       /*
-        * for now just try remotifying unbuddied pages, starting with
-        * least space avail
-        */
-       for (i = 0; i < MAX_CHUNK; i++) {
-retry_unbud_list_i:
-               preempt_disable();  /* enable in zbud_remotify_zbpg */
-               spin_lock_bh(&zbud_budlists_spinlock);
-               if (list_empty(&zbud_unbuddied[i].list)) {
-                       spin_unlock_bh(&zbud_budlists_spinlock);
-                       preempt_enable();
-                       continue; /* next i in for loop */
-               }
-               list_for_each_entry(zbpg, &zbud_unbuddied[i].list, bud_list) {
-                       if (unlikely(!spin_trylock(&zbpg->lock)))
-                               continue; /* next list_for_each_entry */
-                       zbud_unbuddied[i].count--;
-                       /* want budlists unlocked when doing zbpg remotify */
-                       spin_unlock_bh(&zbud_budlists_spinlock);
-                       ret = zbud_remotify_zbpg(zbpg);
-                       /* preemption is re-enabled in zbud_remotify_zbpg */
-                       if (ret == 0) {
-                               if (--nr <= 0)
-                                       goto out;
-                               goto retry_unbud_list_i;
-                       }
-                       /* if fail to remotify any page, quit */
-                       pr_err("TESTING zbud_remotify_pages failed on page,"
-                               " trying to re-add\n");
-                       spin_lock_bh(&zbud_budlists_spinlock);
-                       spin_lock(&zbpg->lock);
-                       list_add_tail(&zbpg->bud_list, &zbud_unbuddied[i].list);
-                       zbud_unbuddied[i].count++;
-                       spin_unlock(&zbpg->lock);
-                       spin_unlock_bh(&zbud_budlists_spinlock);
-                       pr_err("TESTING zbud_remotify_pages failed on page,"
-                               " finished re-add\n");
-                       goto out;
-               }
-               spin_unlock_bh(&zbud_budlists_spinlock);
-               preempt_enable();
-       }
-
-next_buddied_zbpg:
-       preempt_disable();  /* enable in zbud_remotify_zbpg */
-       spin_lock_bh(&zbud_budlists_spinlock);
-       if (list_empty(&zbud_buddied_list))
-               goto unlock_out;
-       list_for_each_entry(zbpg, &zbud_buddied_list, bud_list) {
-               if (unlikely(!spin_trylock(&zbpg->lock)))
-                       continue; /* next list_for_each_entry */
-               zcache_zbud_buddied_count--;
-               /* want budlists unlocked when doing zbpg remotify */
-               spin_unlock_bh(&zbud_budlists_spinlock);
-               ret = zbud_remotify_zbpg(zbpg);
-               /* preemption is re-enabled in zbud_remotify_zbpg */
-               if (ret == 0) {
-                       if (--nr <= 0)
-                               goto out;
-                       goto next_buddied_zbpg;
-               }
-               /* if fail to remotify any page, quit */
-               pr_err("TESTING zbud_remotify_pages failed on BUDDIED page,"
-                       " trying to re-add\n");
-               spin_lock_bh(&zbud_budlists_spinlock);
-               spin_lock(&zbpg->lock);
-               list_add_tail(&zbpg->bud_list, &zbud_buddied_list);
-               zcache_zbud_buddied_count++;
-               spin_unlock(&zbpg->lock);
-               spin_unlock_bh(&zbud_budlists_spinlock);
-               pr_err("TESTING zbud_remotify_pages failed on BUDDIED page,"
-                       " finished re-add\n");
-               goto out;
-       }
-unlock_out:
-       spin_unlock_bh(&zbud_budlists_spinlock);
-       preempt_enable();
-out:
-       return;
-}
-
-/* the "flush list" asynchronously collects pages to remotely flush */
-#define FLUSH_ENTIRE_OBJECT ((uint32_t)-1)
-static void ramster_flnode_free(struct flushlist_node *,
-                               struct tmem_pool *);
-
-static void zcache_remote_flush_page(struct flushlist_node *flnode)
-{
-       struct tmem_xhandle *xh;
-       int remotenode, ret;
-
-       preempt_disable();
-       xh = &flnode->xh;
-       remotenode = flnode->xh.client_id;
-       ret = ramster_remote_flush(xh, remotenode);
-       if (ret >= 0)
-               ramster_remote_pages_flushed++;
-       else
-               ramster_remote_page_flushes_failed++;
-       preempt_enable_no_resched();
-       ramster_flnode_free(flnode, NULL);
-}
-
-static void zcache_remote_flush_object(struct flushlist_node *flnode)
-{
-       struct tmem_xhandle *xh;
-       int remotenode, ret;
-
-       preempt_disable();
-       xh = &flnode->xh;
-       remotenode = flnode->xh.client_id;
-       ret = ramster_remote_flush_object(xh, remotenode);
-       if (ret >= 0)
-               ramster_remote_objects_flushed++;
-       else
-               ramster_remote_object_flushes_failed++;
-       preempt_enable_no_resched();
-       ramster_flnode_free(flnode, NULL);
-}
-
-static void zcache_remote_eph_put(struct zbud_hdr *zbud)
-{
-       /* FIXME */
-}
-
-static void zcache_remote_pers_put(struct zv_hdr *zv)
-{
-       struct tmem_xhandle xh;
-       uint16_t size;
-       bool ephemeral;
-       int remotenode, ret = -1;
-       char *data;
-       struct tmem_pool *pool;
-       unsigned long flags;
-       unsigned char cksum;
-       char *p;
-       int i;
-       unsigned char *tmpmem = __get_cpu_var(zcache_remoteputmem);
-
-       ASSERT_SENTINEL(zv, ZVH);
-       BUG_ON(zv->client_id != LOCAL_CLIENT);
-       local_bh_disable();
-       xh.client_id = zv->client_id;
-       xh.pool_id = zv->pool_id;
-       xh.oid = zv->oid;
-       xh.index = zv->index;
-       size = xv_get_object_size(zv) - sizeof(*zv);
-       BUG_ON(size == 0 || size > zv_max_page_size);
-       data = (char *)zv + sizeof(*zv);
-       for (p = data, cksum = 0, i = 0; i < size; i++)
-               cksum += *p;
-       memcpy(tmpmem, data, size);
-       data = tmpmem;
-       pool = zcache_get_pool_by_id(zv->client_id, zv->pool_id);
-       ephemeral = is_ephemeral(pool);
-       zcache_put_pool(pool);
-       /* now OK to release lock set in caller */
-       spin_unlock(&zcache_rem_op_list_lock);
-       local_bh_enable();
-       preempt_disable();
-       ret = ramster_remote_put(&xh, data, size, ephemeral, &remotenode);
-       preempt_enable_no_resched();
-       if (ret != 0) {
-               /*
-                * This is some form of a memory leak... if the remote put
-                * fails, there will never be another attempt to remotify
-                * this page.  But since we've dropped the zv pointer,
-                * the page may have been freed or the data replaced
-                * so we can't just "put it back" in the remote op list.
-                * Even if we could, not sure where to put it in the list
-                * because there may be flushes that must be strictly
-                * ordered vs the put.  So leave this as a FIXME for now.
-                * But count them so we know if it becomes a problem.
-                */
-               ramster_pers_pages_remote_failed++;
-               goto out;
-       } else
-               atomic_inc(&ramster_remote_pers_pages);
-       ramster_pers_pages_remoted++;
-       /*
-        * data was successfully remoted so change the local version to
-        * point to the remote node where it landed
-        */
-       local_bh_disable();
-       pool = zcache_get_pool_by_id(LOCAL_CLIENT, xh.pool_id);
-       local_irq_save(flags);
-       (void)tmem_replace(pool, &xh.oid, xh.index,
-                       pampd_make_remote(remotenode, size, cksum));
-       local_irq_restore(flags);
-       zcache_put_pool(pool);
-       local_bh_enable();
-out:
-       return;
-}
-
-static void zcache_do_remotify_ops(int nr)
-{
-       struct ramster_remotify_hdr *rem_op;
-       union remotify_list_node *u;
-
-       while (1) {
-               if (!nr)
-                       goto out;
-               spin_lock(&zcache_rem_op_list_lock);
-               if (list_empty(&zcache_rem_op_list)) {
-                       spin_unlock(&zcache_rem_op_list_lock);
-                       goto out;
-               }
-               rem_op = list_first_entry(&zcache_rem_op_list,
-                               struct ramster_remotify_hdr, list);
-               list_del_init(&rem_op->list);
-               if (rem_op->op != RAMSTER_REMOTIFY_PERS_PUT)
-                       spin_unlock(&zcache_rem_op_list_lock);
-               u = (union remotify_list_node *)rem_op;
-               switch (rem_op->op) {
-               case RAMSTER_REMOTIFY_EPH_PUT:
-BUG();
-                       zcache_remote_eph_put((struct zbud_hdr *)rem_op);
-                       break;
-               case RAMSTER_REMOTIFY_PERS_PUT:
-                       zcache_remote_pers_put((struct zv_hdr *)rem_op);
-                       break;
-               case RAMSTER_REMOTIFY_FLUSH_PAGE:
-                       zcache_remote_flush_page((struct flushlist_node *)u);
-                       break;
-               case RAMSTER_REMOTIFY_FLUSH_OBJ:
-                       zcache_remote_flush_object((struct flushlist_node *)u);
-                       break;
-               default:
-                       BUG();
-               }
-       }
-out:
-       return;
-}
-
-/*
- * For now, just push over a few pages every few seconds to
- * ensure that it basically works
- */
-static struct workqueue_struct *ramster_remotify_workqueue;
-static void ramster_remotify_process(struct work_struct *work);
-static DECLARE_DELAYED_WORK(ramster_remotify_worker,
-               ramster_remotify_process);
-
-static void ramster_remotify_queue_delayed_work(unsigned long delay)
-{
-       if (!queue_delayed_work(ramster_remotify_workqueue,
-                               &ramster_remotify_worker, delay))
-               pr_err("ramster_remotify: bad workqueue\n");
-}
-
-
-static int use_frontswap;
-static int use_cleancache;
-static void ramster_remotify_process(struct work_struct *work)
-{
-       static bool remotify_in_progress;
-
-       BUG_ON(irqs_disabled());
-       if (remotify_in_progress)
-               ramster_remotify_queue_delayed_work(HZ);
-       else {
-               remotify_in_progress = true;
-#ifdef CONFIG_CLEANCACHE
-       if (use_cleancache && ramster_eph_remotify_enable)
-               zbud_remotify_pages(5000); /* FIXME is this a good number? */
-#endif
-#ifdef CONFIG_FRONTSWAP
-       if (use_frontswap && ramster_pers_remotify_enable)
-               zcache_do_remotify_ops(500); /* FIXME is this a good number? */
-#endif
-               remotify_in_progress = false;
-               ramster_remotify_queue_delayed_work(HZ);
-       }
-}
-
-static void ramster_remotify_init(void)
-{
-       unsigned long n = 60UL;
-       ramster_remotify_workqueue =
-               create_singlethread_workqueue("ramster_remotify");
-       ramster_remotify_queue_delayed_work(n * HZ);
-}
-
-
-static void zbud_init(void)
-{
-       int i;
-
-       INIT_LIST_HEAD(&zbud_buddied_list);
-       zcache_zbud_buddied_count = 0;
-       for (i = 0; i < NCHUNKS; i++) {
-               INIT_LIST_HEAD(&zbud_unbuddied[i].list);
-               zbud_unbuddied[i].count = 0;
-       }
-}
-
-#ifdef CONFIG_SYSFS
-/*
- * These sysfs routines show a nice distribution of how many zbpg's are
- * currently (and have ever been placed) in each unbuddied list.  It's fun
- * to watch but can probably go away before final merge.
- */
-static int zbud_show_unbuddied_list_counts(char *buf)
-{
-       int i;
-       char *p = buf;
-
-       for (i = 0; i < NCHUNKS; i++)
-               p += sprintf(p, "%u ", zbud_unbuddied[i].count);
-       return p - buf;
-}
-
-static int zbud_show_cumul_chunk_counts(char *buf)
-{
-       unsigned long i, chunks = 0, total_chunks = 0, sum_total_chunks = 0;
-       unsigned long total_chunks_lte_21 = 0, total_chunks_lte_32 = 0;
-       unsigned long total_chunks_lte_42 = 0;
-       char *p = buf;
-
-       for (i = 0; i < NCHUNKS; i++) {
-               p += sprintf(p, "%lu ", zbud_cumul_chunk_counts[i]);
-               chunks += zbud_cumul_chunk_counts[i];
-               total_chunks += zbud_cumul_chunk_counts[i];
-               sum_total_chunks += i * zbud_cumul_chunk_counts[i];
-               if (i == 21)
-                       total_chunks_lte_21 = total_chunks;
-               if (i == 32)
-                       total_chunks_lte_32 = total_chunks;
-               if (i == 42)
-                       total_chunks_lte_42 = total_chunks;
-       }
-       p += sprintf(p, "<=21:%lu <=32:%lu <=42:%lu, mean:%lu\n",
-               total_chunks_lte_21, total_chunks_lte_32, total_chunks_lte_42,
-               chunks == 0 ? 0 : sum_total_chunks / chunks);
-       return p - buf;
-}
-#endif
-
-/**********
- * This "zv" PAM implementation combines the TLSF-based xvMalloc
- * with lzo1x compression to maximize the amount of data that can
- * be packed into a physical page.
- *
- * Zv represents a PAM page with the index and object (plus a "size" value
- * necessary for decompression) immediately preceding the compressed data.
- */
-
-/* rudimentary policy limits */
-/* total number of persistent pages may not exceed this percentage */
-static unsigned int zv_page_count_policy_percent = 75;
-/*
- * byte count defining poor compression; pages with greater zsize will be
- * rejected
- */
-static unsigned int zv_max_zsize = (PAGE_SIZE / 8) * 7;
-/*
- * byte count defining poor *mean* compression; pages with greater zsize
- * will be rejected until sufficient better-compressed pages are accepted
- * driving the mean below this threshold
- */
-static unsigned int zv_max_mean_zsize = (PAGE_SIZE / 8) * 5;
-
-static atomic_t zv_curr_dist_counts[NCHUNKS];
-static atomic_t zv_cumul_dist_counts[NCHUNKS];
-
-
-static struct zv_hdr *zv_create(struct zcache_client *cli, uint32_t pool_id,
-                               struct tmem_oid *oid, uint32_t index,
-                               void *cdata, unsigned clen)
-{
-       struct page *page;
-       struct zv_hdr *zv = NULL;
-       uint32_t offset;
-       int alloc_size = clen + sizeof(struct zv_hdr);
-       int chunks = (alloc_size + (CHUNK_SIZE - 1)) >> CHUNK_SHIFT;
-       int ret;
-
-       BUG_ON(!irqs_disabled());
-       BUG_ON(chunks >= NCHUNKS);
-       ret = xv_malloc(cli->xvpool, clen + sizeof(struct zv_hdr),
-                       &page, &offset, ZCACHE_GFP_MASK);
-       if (unlikely(ret))
-               goto out;
-       atomic_inc(&zv_curr_dist_counts[chunks]);
-       atomic_inc(&zv_cumul_dist_counts[chunks]);
-       zv = kmap_atomic(page, KM_USER0) + offset;
-       zv->index = index;
-       zv->oid = *oid;
-       zv->pool_id = pool_id;
-       SET_SENTINEL(zv, ZVH);
-       INIT_LIST_HEAD(&zv->rem_op.list);
-       zv->client_id = get_client_id_from_client(cli);
-       zv->rem_op.op = RAMSTER_REMOTIFY_PERS_PUT;
-       if (zv->client_id == LOCAL_CLIENT) {
-               spin_lock(&zcache_rem_op_list_lock);
-               list_add_tail(&zv->rem_op.list, &zcache_rem_op_list);
-               spin_unlock(&zcache_rem_op_list_lock);
-       }
-       memcpy((char *)zv + sizeof(struct zv_hdr), cdata, clen);
-       kunmap_atomic(zv, KM_USER0);
-out:
-       return zv;
-}
-
-/* similar to zv_create, but just reserve space, no data yet */
-static struct zv_hdr *zv_alloc(struct tmem_pool *pool,
-                               struct tmem_oid *oid, uint32_t index,
-                               unsigned clen)
-{
-       struct zcache_client *cli = pool->client;
-       struct page *page;
-       struct zv_hdr *zv = NULL;
-       uint32_t offset;
-       int ret;
-
-       BUG_ON(!irqs_disabled());
-       BUG_ON(!is_local_client(pool->client));
-       ret = xv_malloc(cli->xvpool, clen + sizeof(struct zv_hdr),
-                       &page, &offset, ZCACHE_GFP_MASK);
-       if (unlikely(ret))
-               goto out;
-       zv = kmap_atomic(page, KM_USER0) + offset;
-       SET_SENTINEL(zv, ZVH);
-       INIT_LIST_HEAD(&zv->rem_op.list);
-       zv->client_id = LOCAL_CLIENT;
-       zv->rem_op.op = RAMSTER_INTRANSIT_PERS;
-       zv->index = index;
-       zv->oid = *oid;
-       zv->pool_id = pool->pool_id;
-       kunmap_atomic(zv, KM_USER0);
-out:
-       return zv;
-}
-
-static void zv_free(struct xv_pool *xvpool, struct zv_hdr *zv)
-{
-       unsigned long flags;
-       struct page *page;
-       uint32_t offset;
-       uint16_t size = xv_get_object_size(zv);
-       int chunks = (size + (CHUNK_SIZE - 1)) >> CHUNK_SHIFT;
-
-       ASSERT_SENTINEL(zv, ZVH);
-       BUG_ON(chunks >= NCHUNKS);
-       atomic_dec(&zv_curr_dist_counts[chunks]);
-       size -= sizeof(*zv);
-       spin_lock(&zcache_rem_op_list_lock);
-       size = xv_get_object_size(zv) - sizeof(*zv);
-       BUG_ON(size == 0);
-       INVERT_SENTINEL(zv, ZVH);
-       if (!list_empty(&zv->rem_op.list))
-               list_del_init(&zv->rem_op.list);
-       spin_unlock(&zcache_rem_op_list_lock);
-       page = virt_to_page(zv);
-       offset = (unsigned long)zv & ~PAGE_MASK;
-       local_irq_save(flags);
-       xv_free(xvpool, page, offset);
-       local_irq_restore(flags);
-}
-
-static void zv_decompress(struct page *page, struct zv_hdr *zv)
-{
-       size_t clen = PAGE_SIZE;
-       char *to_va;
-       unsigned size;
-       int ret;
-
-       ASSERT_SENTINEL(zv, ZVH);
-       size = xv_get_object_size(zv) - sizeof(*zv);
-       BUG_ON(size == 0);
-       to_va = kmap_atomic(page, KM_USER0);
-       ret = lzo1x_decompress_safe((char *)zv + sizeof(*zv),
-                                       size, to_va, &clen);
-       kunmap_atomic(to_va, KM_USER0);
-       BUG_ON(ret != LZO_E_OK);
-       BUG_ON(clen != PAGE_SIZE);
-}
-
-static void zv_copy_from_pampd(char *data, size_t *bufsize, struct zv_hdr *zv)
-{
-       unsigned size;
-
-       ASSERT_SENTINEL(zv, ZVH);
-       size = xv_get_object_size(zv) - sizeof(*zv);
-       BUG_ON(size == 0 || size > zv_max_page_size);
-       BUG_ON(size > *bufsize);
-       memcpy(data, (char *)zv + sizeof(*zv), size);
-       *bufsize = size;
-}
-
-static void zv_copy_to_pampd(struct zv_hdr *zv, char *data, size_t size)
-{
-       unsigned zv_size;
-
-       ASSERT_SENTINEL(zv, ZVH);
-       zv_size = xv_get_object_size(zv) - sizeof(*zv);
-       BUG_ON(zv_size != size);
-       BUG_ON(zv_size == 0 || zv_size > zv_max_page_size);
-       memcpy((char *)zv + sizeof(*zv), data, size);
-}
-
-#ifdef CONFIG_SYSFS
-/*
- * show a distribution of compression stats for zv pages.
- */
-
-static int zv_curr_dist_counts_show(char *buf)
-{
-       unsigned long i, n, chunks = 0, sum_total_chunks = 0;
-       char *p = buf;
-
-       for (i = 0; i < NCHUNKS; i++) {
-               n = atomic_read(&zv_curr_dist_counts[i]);
-               p += sprintf(p, "%lu ", n);
-               chunks += n;
-               sum_total_chunks += i * n;
-       }
-       p += sprintf(p, "mean:%lu\n",
-               chunks == 0 ? 0 : sum_total_chunks / chunks);
-       return p - buf;
-}
-
-static int zv_cumul_dist_counts_show(char *buf)
-{
-       unsigned long i, n, chunks = 0, sum_total_chunks = 0;
-       char *p = buf;
-
-       for (i = 0; i < NCHUNKS; i++) {
-               n = atomic_read(&zv_cumul_dist_counts[i]);
-               p += sprintf(p, "%lu ", n);
-               chunks += n;
-               sum_total_chunks += i * n;
-       }
-       p += sprintf(p, "mean:%lu\n",
-               chunks == 0 ? 0 : sum_total_chunks / chunks);
-       return p - buf;
-}
-
-/*
- * setting zv_max_zsize via sysfs causes all persistent (e.g. swap)
- * pages that don't compress to less than this value (including metadata
- * overhead) to be rejected.  We don't allow the value to get too close
- * to PAGE_SIZE.
- */
-static ssize_t zv_max_zsize_show(struct kobject *kobj,
-                                   struct kobj_attribute *attr,
-                                   char *buf)
-{
-       return sprintf(buf, "%u\n", zv_max_zsize);
-}
-
-static ssize_t zv_max_zsize_store(struct kobject *kobj,
-                                   struct kobj_attribute *attr,
-                                   const char *buf, size_t count)
-{
-       unsigned long val;
-       int err;
-
-       if (!capable(CAP_SYS_ADMIN))
-               return -EPERM;
-
-       err = strict_strtoul(buf, 10, &val);
-       if (err || (val == 0) || (val > (PAGE_SIZE / 8) * 7))
-               return -EINVAL;
-       zv_max_zsize = val;
-       return count;
-}
-
-/*
- * setting zv_max_mean_zsize via sysfs causes all persistent (e.g. swap)
- * pages that don't compress to less than this value (including metadata
- * overhead) to be rejected UNLESS the mean compression is also smaller
- * than this value.  In other words, we are load-balancing-by-zsize the
- * accepted pages.  Again, we don't allow the value to get too close
- * to PAGE_SIZE.
- */
-static ssize_t zv_max_mean_zsize_show(struct kobject *kobj,
-                                   struct kobj_attribute *attr,
-                                   char *buf)
-{
-       return sprintf(buf, "%u\n", zv_max_mean_zsize);
-}
-
-static ssize_t zv_max_mean_zsize_store(struct kobject *kobj,
-                                   struct kobj_attribute *attr,
-                                   const char *buf, size_t count)
-{
-       unsigned long val;
-       int err;
-
-       if (!capable(CAP_SYS_ADMIN))
-               return -EPERM;
-
-       err = strict_strtoul(buf, 10, &val);
-       if (err || (val == 0) || (val > (PAGE_SIZE / 8) * 7))
-               return -EINVAL;
-       zv_max_mean_zsize = val;
-       return count;
-}
-
-/*
- * setting zv_page_count_policy_percent via sysfs sets an upper bound of
- * persistent (e.g. swap) pages that will be retained according to:
- *     (zv_page_count_policy_percent * totalram_pages) / 100)
- * when that limit is reached, further puts will be rejected (until
- * some pages have been flushed).  Note that, due to compression,
- * this number may exceed 100; it defaults to 75 and we set an
- * arbitary limit of 150.  A poor choice will almost certainly result
- * in OOM's, so this value should only be changed prudently.
- */
-static ssize_t zv_page_count_policy_percent_show(struct kobject *kobj,
-                                                struct kobj_attribute *attr,
-                                                char *buf)
-{
-       return sprintf(buf, "%u\n", zv_page_count_policy_percent);
-}
-
-static ssize_t zv_page_count_policy_percent_store(struct kobject *kobj,
-                                                 struct kobj_attribute *attr,
-                                                 const char *buf, size_t count)
-{
-       unsigned long val;
-       int err;
-
-       if (!capable(CAP_SYS_ADMIN))
-               return -EPERM;
-
-       err = strict_strtoul(buf, 10, &val);
-       if (err || (val == 0) || (val > 150))
-               return -EINVAL;
-       zv_page_count_policy_percent = val;
-       return count;
-}
-
-static struct kobj_attribute zcache_zv_max_zsize_attr = {
-               .attr = { .name = "zv_max_zsize", .mode = 0644 },
-               .show = zv_max_zsize_show,
-               .store = zv_max_zsize_store,
-};
-
-static struct kobj_attribute zcache_zv_max_mean_zsize_attr = {
-               .attr = { .name = "zv_max_mean_zsize", .mode = 0644 },
-               .show = zv_max_mean_zsize_show,
-               .store = zv_max_mean_zsize_store,
-};
-
-static struct kobj_attribute zcache_zv_page_count_policy_percent_attr = {
-               .attr = { .name = "zv_page_count_policy_percent",
-                         .mode = 0644 },
-               .show = zv_page_count_policy_percent_show,
-               .store = zv_page_count_policy_percent_store,
-};
-#endif
-
-/*
- * zcache core code starts here
- */
-
-/* useful stats not collected by cleancache or frontswap */
-static unsigned long zcache_flush_total;
-static unsigned long zcache_flush_found;
-static unsigned long zcache_flobj_total;
-static unsigned long zcache_flobj_found;
-static unsigned long zcache_failed_eph_puts;
-static unsigned long zcache_nonactive_puts;
-static unsigned long zcache_failed_pers_puts;
-
-/*
- * Tmem operations assume the poolid implies the invoking client.
- * Zcache only has one client (the kernel itself): LOCAL_CLIENT.
- * RAMster has each client numbered by cluster node, and a KVM version
- * of zcache would have one client per guest and each client might
- * have a poolid==N.
- */
-static struct tmem_pool *zcache_get_pool_by_id(uint16_t cli_id, uint16_t poolid)
-{
-       struct tmem_pool *pool = NULL;
-       struct zcache_client *cli = NULL;
-
-       if (cli_id == LOCAL_CLIENT)
-               cli = &zcache_host;
-       else {
-               if (cli_id >= MAX_CLIENTS)
-                       goto out;
-               cli = &zcache_clients[cli_id];
-               if (cli == NULL)
-                       goto out;
-               atomic_inc(&cli->refcount);
-       }
-       if (poolid < MAX_POOLS_PER_CLIENT) {
-               pool = cli->tmem_pools[poolid];
-               if (pool != NULL)
-                       atomic_inc(&pool->refcount);
-       }
-out:
-       return pool;
-}
-
-static void zcache_put_pool(struct tmem_pool *pool)
-{
-       struct zcache_client *cli = NULL;
-
-       if (pool == NULL)
-               BUG();
-       cli = pool->client;
-       atomic_dec(&pool->refcount);
-       atomic_dec(&cli->refcount);
-}
-
-int zcache_new_client(uint16_t cli_id)
-{
-       struct zcache_client *cli = NULL;
-       int ret = -1;
-
-       if (cli_id == LOCAL_CLIENT)
-               cli = &zcache_host;
-       else if ((unsigned int)cli_id < MAX_CLIENTS)
-               cli = &zcache_clients[cli_id];
-       if (cli == NULL)
-               goto out;
-       if (cli->allocated)
-               goto out;
-       cli->allocated = 1;
-#ifdef CONFIG_FRONTSWAP
-       cli->xvpool = xv_create_pool();
-       if (cli->xvpool == NULL)
-               goto out;
-#endif
-       ret = 0;
-out:
-       return ret;
-}
-
-/* counters for debugging */
-static unsigned long zcache_failed_get_free_pages;
-static unsigned long zcache_failed_alloc;
-static unsigned long zcache_put_to_flush;
-
-/*
- * for now, used named slabs so can easily track usage; later can
- * either just use kmalloc, or perhaps add a slab-like allocator
- * to more carefully manage total memory utilization
- */
-static struct kmem_cache *zcache_objnode_cache;
-static struct kmem_cache *zcache_obj_cache;
-static struct kmem_cache *ramster_flnode_cache;
-static atomic_t zcache_curr_obj_count = ATOMIC_INIT(0);
-static unsigned long zcache_curr_obj_count_max;
-static atomic_t zcache_curr_objnode_count = ATOMIC_INIT(0);
-static unsigned long zcache_curr_objnode_count_max;
-
-/*
- * to avoid memory allocation recursion (e.g. due to direct reclaim), we
- * preload all necessary data structures so the hostops callbacks never
- * actually do a malloc
- */
-struct zcache_preload {
-       void *page;
-       struct tmem_obj *obj;
-       int nr;
-       struct tmem_objnode *objnodes[OBJNODE_TREE_MAX_PATH];
-       struct flushlist_node *flnode;
-};
-static DEFINE_PER_CPU(struct zcache_preload, zcache_preloads) = { 0, };
-
-static int zcache_do_preload(struct tmem_pool *pool)
-{
-       struct zcache_preload *kp;
-       struct tmem_objnode *objnode;
-       struct tmem_obj *obj;
-       struct flushlist_node *flnode;
-       void *page;
-       int ret = -ENOMEM;
-
-       if (unlikely(zcache_objnode_cache == NULL))
-               goto out;
-       if (unlikely(zcache_obj_cache == NULL))
-               goto out;
-       preempt_disable();
-       kp = &__get_cpu_var(zcache_preloads);
-       while (kp->nr < ARRAY_SIZE(kp->objnodes)) {
-               preempt_enable_no_resched();
-               objnode = kmem_cache_alloc(zcache_objnode_cache,
-                               ZCACHE_GFP_MASK);
-               if (unlikely(objnode == NULL)) {
-                       zcache_failed_alloc++;
-                       goto out;
-               }
-               preempt_disable();
-               kp = &__get_cpu_var(zcache_preloads);
-               if (kp->nr < ARRAY_SIZE(kp->objnodes))
-                       kp->objnodes[kp->nr++] = objnode;
-               else
-                       kmem_cache_free(zcache_objnode_cache, objnode);
-       }
-       preempt_enable_no_resched();
-       obj = kmem_cache_alloc(zcache_obj_cache, ZCACHE_GFP_MASK);
-       if (unlikely(obj == NULL)) {
-               zcache_failed_alloc++;
-               goto out;
-       }
-       flnode = kmem_cache_alloc(ramster_flnode_cache, ZCACHE_GFP_MASK);
-       if (unlikely(flnode == NULL)) {
-               zcache_failed_alloc++;
-               goto out;
-       }
-       if (is_ephemeral(pool)) {
-               page = (void *)__get_free_page(ZCACHE_GFP_MASK);
-               if (unlikely(page == NULL)) {
-                       zcache_failed_get_free_pages++;
-                       kmem_cache_free(zcache_obj_cache, obj);
-                       kmem_cache_free(ramster_flnode_cache, flnode);
-                       goto out;
-               }
-       }
-       preempt_disable();
-       kp = &__get_cpu_var(zcache_preloads);
-       if (kp->obj == NULL)
-               kp->obj = obj;
-       else
-               kmem_cache_free(zcache_obj_cache, obj);
-       if (kp->flnode == NULL)
-               kp->flnode = flnode;
-       else
-               kmem_cache_free(ramster_flnode_cache, flnode);
-       if (is_ephemeral(pool)) {
-               if (kp->page == NULL)
-                       kp->page = page;
-               else
-                       free_page((unsigned long)page);
-       }
-       ret = 0;
-out:
-       return ret;
-}
-
-static int ramster_do_preload_flnode_only(struct tmem_pool *pool)
-{
-       struct zcache_preload *kp;
-       struct flushlist_node *flnode;
-       int ret = -ENOMEM;
-
-       BUG_ON(!irqs_disabled());
-       if (unlikely(ramster_flnode_cache == NULL))
-               BUG();
-       kp = &__get_cpu_var(zcache_preloads);
-       flnode = kmem_cache_alloc(ramster_flnode_cache, GFP_ATOMIC);
-       if (unlikely(flnode == NULL) && kp->flnode == NULL)
-               BUG();  /* FIXME handle more gracefully, but how??? */
-       else if (kp->flnode == NULL)
-               kp->flnode = flnode;
-       else
-               kmem_cache_free(ramster_flnode_cache, flnode);
-       return ret;
-}
-
-static void *zcache_get_free_page(void)
-{
-       struct zcache_preload *kp;
-       void *page;
-
-       kp = &__get_cpu_var(zcache_preloads);
-       page = kp->page;
-       BUG_ON(page == NULL);
-       kp->page = NULL;
-       return page;
-}
-
-static void zcache_free_page(void *p)
-{
-       free_page((unsigned long)p);
-}
-
-/*
- * zcache implementation for tmem host ops
- */
-
-static struct tmem_objnode *zcache_objnode_alloc(struct tmem_pool *pool)
-{
-       struct tmem_objnode *objnode = NULL;
-       unsigned long count;
-       struct zcache_preload *kp;
-
-       kp = &__get_cpu_var(zcache_preloads);
-       if (kp->nr <= 0)
-               goto out;
-       objnode = kp->objnodes[kp->nr - 1];
-       BUG_ON(objnode == NULL);
-       kp->objnodes[kp->nr - 1] = NULL;
-       kp->nr--;
-       count = atomic_inc_return(&zcache_curr_objnode_count);
-       if (count > zcache_curr_objnode_count_max)
-               zcache_curr_objnode_count_max = count;
-out:
-       return objnode;
-}
-
-static void zcache_objnode_free(struct tmem_objnode *objnode,
-                                       struct tmem_pool *pool)
-{
-       atomic_dec(&zcache_curr_objnode_count);
-       BUG_ON(atomic_read(&zcache_curr_objnode_count) < 0);
-       kmem_cache_free(zcache_objnode_cache, objnode);
-}
-
-static struct tmem_obj *zcache_obj_alloc(struct tmem_pool *pool)
-{
-       struct tmem_obj *obj = NULL;
-       unsigned long count;
-       struct zcache_preload *kp;
-
-       kp = &__get_cpu_var(zcache_preloads);
-       obj = kp->obj;
-       BUG_ON(obj == NULL);
-       kp->obj = NULL;
-       count = atomic_inc_return(&zcache_curr_obj_count);
-       if (count > zcache_curr_obj_count_max)
-               zcache_curr_obj_count_max = count;
-       return obj;
-}
-
-static void zcache_obj_free(struct tmem_obj *obj, struct tmem_pool *pool)
-{
-       atomic_dec(&zcache_curr_obj_count);
-       BUG_ON(atomic_read(&zcache_curr_obj_count) < 0);
-       kmem_cache_free(zcache_obj_cache, obj);
-}
-
-static struct flushlist_node *ramster_flnode_alloc(struct tmem_pool *pool)
-{
-       struct flushlist_node *flnode = NULL;
-       struct zcache_preload *kp;
-       int count;
-
-       kp = &__get_cpu_var(zcache_preloads);
-       flnode = kp->flnode;
-       BUG_ON(flnode == NULL);
-       kp->flnode = NULL;
-       count = atomic_inc_return(&ramster_curr_flnode_count);
-       if (count > ramster_curr_flnode_count_max)
-               ramster_curr_flnode_count_max = count;
-       return flnode;
-}
-
-static void ramster_flnode_free(struct flushlist_node *flnode,
-                               struct tmem_pool *pool)
-{
-       atomic_dec(&ramster_curr_flnode_count);
-       BUG_ON(atomic_read(&ramster_curr_flnode_count) < 0);
-       kmem_cache_free(ramster_flnode_cache, flnode);
-}
-
-static struct tmem_hostops zcache_hostops = {
-       .obj_alloc = zcache_obj_alloc,
-       .obj_free = zcache_obj_free,
-       .objnode_alloc = zcache_objnode_alloc,
-       .objnode_free = zcache_objnode_free,
-};
-
-/*
- * zcache implementations for PAM page descriptor ops
- */
-
-static atomic_t zcache_curr_eph_pampd_count = ATOMIC_INIT(0);
-static unsigned long zcache_curr_eph_pampd_count_max;
-static atomic_t zcache_curr_pers_pampd_count = ATOMIC_INIT(0);
-static unsigned long zcache_curr_pers_pampd_count_max;
-
-/* forward reference */
-static int zcache_compress(struct page *from, void **out_va, size_t *out_len);
-
-static int zcache_pampd_eph_create(char *data, size_t size, bool raw,
-                               struct tmem_pool *pool, struct tmem_oid *oid,
-                               uint32_t index, void **pampd)
-{
-       int ret = -1;
-       void *cdata = data;
-       size_t clen = size;
-       struct zcache_client *cli = pool->client;
-       uint16_t client_id = get_client_id_from_client(cli);
-       struct page *page = NULL;
-       unsigned long count;
-
-       if (!raw) {
-               page = virt_to_page(data);
-               ret = zcache_compress(page, &cdata, &clen);
-               if (ret == 0)
-                       goto out;
-               if (clen == 0 || clen > zbud_max_buddy_size()) {
-                       zcache_compress_poor++;
-                       goto out;
-               }
-       }
-       *pampd = (void *)zbud_create(client_id, pool->pool_id, oid,
-                                       index, page, cdata, clen);
-       if (*pampd == NULL) {
-               ret = -ENOMEM;
-               goto out;
-       }
-       ret = 0;
-       count = atomic_inc_return(&zcache_curr_eph_pampd_count);
-       if (count > zcache_curr_eph_pampd_count_max)
-               zcache_curr_eph_pampd_count_max = count;
-       if (client_id != LOCAL_CLIENT) {
-               count = atomic_inc_return(&ramster_foreign_eph_pampd_count);
-               if (count > ramster_foreign_eph_pampd_count_max)
-                       ramster_foreign_eph_pampd_count_max = count;
-       }
-out:
-       return ret;
-}
-
-static int zcache_pampd_pers_create(char *data, size_t size, bool raw,
-                               struct tmem_pool *pool, struct tmem_oid *oid,
-                               uint32_t index, void **pampd)
-{
-       int ret = -1;
-       void *cdata = data;
-       size_t clen = size;
-       struct zcache_client *cli = pool->client;
-       struct page *page;
-       unsigned long count;
-       unsigned long zv_mean_zsize;
-       struct zv_hdr *zv;
-       long curr_pers_pampd_count;
-       u64 total_zsize;
-#ifdef RAMSTER_TESTING
-       static bool pampd_neg_warned;
-#endif
-
-       curr_pers_pampd_count = atomic_read(&zcache_curr_pers_pampd_count) -
-                       atomic_read(&ramster_remote_pers_pages);
-#ifdef RAMSTER_TESTING
-       /* should always be positive, but warn if accounting is off */
-       if (!pampd_neg_warned) {
-               pr_warn("ramster: bad accounting for curr_pers_pampd_count\n");
-               pampd_neg_warned = true;
-       }
-#endif
-       if (curr_pers_pampd_count >
-                   (zv_page_count_policy_percent * totalram_pages) / 100) {
-               zcache_policy_percent_exceeded++;
-               goto out;
-       }
-       if (raw)
-               goto ok_to_create;
-       page = virt_to_page(data);
-       if (zcache_compress(page, &cdata, &clen) == 0)
-               goto out;
-       /* reject if compression is too poor */
-       if (clen > zv_max_zsize) {
-               zcache_compress_poor++;
-               goto out;
-       }
-       /* reject if mean compression is too poor */
-       if ((clen > zv_max_mean_zsize) && (curr_pers_pampd_count > 0)) {
-               total_zsize = xv_get_total_size_bytes(cli->xvpool);
-               zv_mean_zsize = div_u64(total_zsize, curr_pers_pampd_count);
-               if (zv_mean_zsize > zv_max_mean_zsize) {
-                       zcache_mean_compress_poor++;
-                       goto out;
-               }
-       }
-ok_to_create:
-       *pampd = (void *)zv_create(cli, pool->pool_id, oid, index, cdata, clen);
-       if (*pampd == NULL) {
-               ret = -ENOMEM;
-               goto out;
-       }
-       ret = 0;
-       count = atomic_inc_return(&zcache_curr_pers_pampd_count);
-       if (count > zcache_curr_pers_pampd_count_max)
-               zcache_curr_pers_pampd_count_max = count;
-       if (is_local_client(cli))
-               goto out;
-       zv = *(struct zv_hdr **)pampd;
-       count = atomic_inc_return(&ramster_foreign_pers_pampd_count);
-       if (count > ramster_foreign_pers_pampd_count_max)
-               ramster_foreign_pers_pampd_count_max = count;
-out:
-       return ret;
-}
-
-static void *zcache_pampd_create(char *data, size_t size, bool raw, int eph,
-                               struct tmem_pool *pool, struct tmem_oid *oid,
-                               uint32_t index)
-{
-       void *pampd = NULL;
-       int ret;
-       bool ephemeral;
-
-       BUG_ON(preemptible());
-       ephemeral = (eph == 1) || ((eph == 0) && is_ephemeral(pool));
-       if (ephemeral)
-               ret = zcache_pampd_eph_create(data, size, raw, pool,
-                                               oid, index, &pampd);
-       else
-               ret = zcache_pampd_pers_create(data, size, raw, pool,
-                                               oid, index, &pampd);
-       /* FIXME add some counters here for failed creates? */
-       return pampd;
-}
-
-/*
- * fill the pageframe corresponding to the struct page with the data
- * from the passed pampd
- */
-static int zcache_pampd_get_data(char *data, size_t *bufsize, bool raw,
-                                       void *pampd, struct tmem_pool *pool,
-                                       struct tmem_oid *oid, uint32_t index)
-{
-       int ret = 0;
-
-       BUG_ON(preemptible());
-       BUG_ON(is_ephemeral(pool)); /* Fix later for shared pools? */
-       BUG_ON(pampd_is_remote(pampd));
-       if (raw)
-               zv_copy_from_pampd(data, bufsize, pampd);
-       else
-               zv_decompress(virt_to_page(data), pampd);
-       return ret;
-}
-
-static int zcache_pampd_get_data_and_free(char *data, size_t *bufsize, bool raw,
-                                       void *pampd, struct tmem_pool *pool,
-                                       struct tmem_oid *oid, uint32_t index)
-{
-       int ret = 0;
-       unsigned long flags;
-       struct zcache_client *cli = pool->client;
-
-       BUG_ON(preemptible());
-       BUG_ON(pampd_is_remote(pampd));
-       if (is_ephemeral(pool)) {
-               local_irq_save(flags);
-               if (raw)
-                       zbud_copy_from_pampd(data, bufsize, pampd);
-               else
-                       ret = zbud_decompress(virt_to_page(data), pampd);
-               zbud_free_and_delist((struct zbud_hdr *)pampd);
-               local_irq_restore(flags);
-               if (!is_local_client(cli)) {
-                       atomic_dec(&ramster_foreign_eph_pampd_count);
-                       WARN_ON_ONCE(atomic_read(&ramster_foreign_eph_pampd_count) < 0);
-               }
-               atomic_dec(&zcache_curr_eph_pampd_count);
-               WARN_ON_ONCE(atomic_read(&zcache_curr_eph_pampd_count) < 0);
-       } else {
-               if (is_local_client(cli))
-                       BUG();
-               if (raw)
-                       zv_copy_from_pampd(data, bufsize, pampd);
-               else
-                       zv_decompress(virt_to_page(data), pampd);
-               zv_free(cli->xvpool, pampd);
-               if (!is_local_client(cli)) {
-                       atomic_dec(&ramster_foreign_pers_pampd_count);
-                       WARN_ON_ONCE(atomic_read(&ramster_foreign_pers_pampd_count) < 0);
-               }
-               atomic_dec(&zcache_curr_pers_pampd_count);
-               WARN_ON_ONCE(atomic_read(&zcache_curr_pers_pampd_count) < 0);
-               ret = 0;
-       }
-       return ret;
-}
-
-static bool zcache_pampd_is_remote(void *pampd)
-{
-       return pampd_is_remote(pampd);
-}
-
-/*
- * free the pampd and remove it from any zcache lists
- * pampd must no longer be pointed to from any tmem data structures!
- */
-static void zcache_pampd_free(void *pampd, struct tmem_pool *pool,
-                             struct tmem_oid *oid, uint32_t index, bool acct)
-{
-       struct zcache_client *cli = pool->client;
-       bool eph = is_ephemeral(pool);
-       struct zv_hdr *zv;
-
-       BUG_ON(preemptible());
-       if (pampd_is_remote(pampd)) {
-               WARN_ON(acct == false);
-               if (oid == NULL) {
-                       /*
-                        * a NULL oid means to ignore this pampd free
-                        * as the remote freeing will be handled elsewhere
-                        */
-               } else if (eph) {
-                       /* FIXME remote flush optional but probably good idea */
-                       /* FIXME get these working properly again */
-                       atomic_dec(&zcache_curr_eph_pampd_count);
-                       WARN_ON_ONCE(atomic_read(&zcache_curr_eph_pampd_count) < 0);
-               } else if (pampd_is_intransit(pampd)) {
-                       /* did a pers remote get_and_free, so just free local */
-                       pampd = pampd_mask_intransit_and_remote(pampd);
-                       goto local_pers;
-               } else {
-                       struct flushlist_node *flnode =
-                               ramster_flnode_alloc(pool);
-
-                       flnode->xh.client_id = pampd_remote_node(pampd);
-                       flnode->xh.pool_id = pool->pool_id;
-                       flnode->xh.oid = *oid;
-                       flnode->xh.index = index;
-                       flnode->rem_op.op = RAMSTER_REMOTIFY_FLUSH_PAGE;
-                       spin_lock(&zcache_rem_op_list_lock);
-                       list_add(&flnode->rem_op.list, &zcache_rem_op_list);
-                       spin_unlock(&zcache_rem_op_list_lock);
-                       atomic_dec(&zcache_curr_pers_pampd_count);
-                       WARN_ON_ONCE(atomic_read(&zcache_curr_pers_pampd_count) < 0);
-                       atomic_dec(&ramster_remote_pers_pages);
-                       WARN_ON_ONCE(atomic_read(&ramster_remote_pers_pages) < 0);
-               }
-       } else if (eph) {
-               zbud_free_and_delist((struct zbud_hdr *)pampd);
-               if (!is_local_client(pool->client)) {
-                       atomic_dec(&ramster_foreign_eph_pampd_count);
-                       WARN_ON_ONCE(atomic_read(&ramster_foreign_eph_pampd_count) < 0);
-               }
-               if (acct)
-                       atomic_dec(&zcache_curr_eph_pampd_count);
-                       /* FIXME get these working properly again */
-                       WARN_ON_ONCE(atomic_read(&zcache_curr_eph_pampd_count) < 0);
-       } else {
-local_pers:
-               zv = (struct zv_hdr *)pampd;
-               if (!is_local_client(pool->client)) {
-                       atomic_dec(&ramster_foreign_pers_pampd_count);
-                       WARN_ON_ONCE(atomic_read(&ramster_foreign_pers_pampd_count) < 0);
-               }
-               zv_free(cli->xvpool, zv);
-               if (acct)
-                       atomic_dec(&zcache_curr_pers_pampd_count);
-               /* FIXME get these working properly again */
-               WARN_ON_ONCE(atomic_read(&zcache_curr_pers_pampd_count) < 0);
-       }
-}
-
-static void zcache_pampd_free_obj(struct tmem_pool *pool,
-                                       struct tmem_obj *obj)
-{
-       struct flushlist_node *flnode;
-
-       BUG_ON(preemptible());
-       if (obj->extra == NULL)
-               return;
-       BUG_ON(!pampd_is_remote(obj->extra));
-       flnode = ramster_flnode_alloc(pool);
-       flnode->xh.client_id = pampd_remote_node(obj->extra);
-       flnode->xh.pool_id = pool->pool_id;
-       flnode->xh.oid = obj->oid;
-       flnode->xh.index = FLUSH_ENTIRE_OBJECT;
-       flnode->rem_op.op = RAMSTER_REMOTIFY_FLUSH_OBJ;
-       spin_lock(&zcache_rem_op_list_lock);
-       list_add(&flnode->rem_op.list, &zcache_rem_op_list);
-       spin_unlock(&zcache_rem_op_list_lock);
-}
-
-void zcache_pampd_new_obj(struct tmem_obj *obj)
-{
-       obj->extra = NULL;
-}
-
-int zcache_pampd_replace_in_obj(void *new_pampd, struct tmem_obj *obj)
-{
-       int ret = -1;
-
-       if (new_pampd != NULL) {
-               if (obj->extra == NULL)
-                       obj->extra = new_pampd;
-               /* enforce that all remote pages in an object reside
-                * in the same node! */
-               else if (pampd_remote_node(new_pampd) !=
-                               pampd_remote_node((void *)(obj->extra)))
-                       BUG();
-               ret = 0;
-       }
-       return ret;
-}
-
-/*
- * Called by the message handler after a (still compressed) page has been
- * fetched from the remote machine in response to an "is_remote" tmem_get
- * or persistent tmem_localify.  For a tmem_get, "extra" is the address of
- * the page that is to be filled to succesfully resolve the tmem_get; for
- * a (persistent) tmem_localify, "extra" is NULL (as the data is placed only
- * in the local zcache).  "data" points to "size" bytes of (compressed) data
- * passed in the message.  In the case of a persistent remote get, if
- * pre-allocation was successful (see zcache_repatriate_preload), the page
- * is placed into both local zcache and at "extra".
- */
-int zcache_localify(int pool_id, struct tmem_oid *oidp,
-                       uint32_t index, char *data, size_t size,
-                       void *extra)
-{
-       int ret = -ENOENT;
-       unsigned long flags;
-       struct tmem_pool *pool;
-       bool ephemeral, delete = false;
-       size_t clen = PAGE_SIZE;
-       void *pampd, *saved_hb;
-       struct tmem_obj *obj;
-
-       pool = zcache_get_pool_by_id(LOCAL_CLIENT, pool_id);
-       if (unlikely(pool == NULL))
-               /* pool doesn't exist anymore */
-               goto out;
-       ephemeral = is_ephemeral(pool);
-       local_irq_save(flags);  /* FIXME: maybe only disable softirqs? */
-       pampd = tmem_localify_get_pampd(pool, oidp, index, &obj, &saved_hb);
-       if (pampd == NULL) {
-               /* hmmm... must have been a flush while waiting */
-#ifdef RAMSTER_TESTING
-               pr_err("UNTESTED pampd==NULL in zcache_localify\n");
-#endif
-               if (ephemeral)
-                       ramster_remote_eph_pages_unsucc_get++;
-               else
-                       ramster_remote_pers_pages_unsucc_get++;
-               obj = NULL;
-               goto finish;
-       } else if (unlikely(!pampd_is_remote(pampd))) {
-               /* hmmm... must have been a dup put while waiting */
-#ifdef RAMSTER_TESTING
-               pr_err("UNTESTED dup while waiting in zcache_localify\n");
-#endif
-               if (ephemeral)
-                       ramster_remote_eph_pages_unsucc_get++;
-               else
-                       ramster_remote_pers_pages_unsucc_get++;
-               obj = NULL;
-               pampd = NULL;
-               ret = -EEXIST;
-               goto finish;
-       } else if (size == 0) {
-               /* no remote data, delete the local is_remote pampd */
-               pampd = NULL;
-               if (ephemeral)
-                       ramster_remote_eph_pages_unsucc_get++;
-               else
-                       BUG();
-               delete = true;
-               goto finish;
-       }
-       if (!ephemeral && pampd_is_intransit(pampd)) {
-               /* localify to zcache */
-               pampd = pampd_mask_intransit_and_remote(pampd);
-               zv_copy_to_pampd(pampd, data, size);
-       } else {
-               pampd = NULL;
-               obj = NULL;
-       }
-       if (extra != NULL) {
-               /* decompress direct-to-memory to complete remotify */
-               ret = lzo1x_decompress_safe((char *)data, size,
-                                               (char *)extra, &clen);
-               BUG_ON(ret != LZO_E_OK);
-               BUG_ON(clen != PAGE_SIZE);
-       }
-       if (ephemeral)
-               ramster_remote_eph_pages_succ_get++;
-       else
-               ramster_remote_pers_pages_succ_get++;
-       ret = 0;
-finish:
-       tmem_localify_finish(obj, index, pampd, saved_hb, delete);
-       zcache_put_pool(pool);
-       local_irq_restore(flags);
-out:
-       return ret;
-}
-
-/*
- * Called on a remote persistent tmem_get to attempt to preallocate
- * local storage for the data contained in the remote persistent page.
- * If succesfully preallocated, returns the pampd, marked as remote and
- * in_transit.  Else returns NULL.  Note that the appropriate tmem data
- * structure must be locked.
- */
-static void *zcache_pampd_repatriate_preload(void *pampd,
-                                               struct tmem_pool *pool,
-                                               struct tmem_oid *oid,
-                                               uint32_t index,
-                                               bool *intransit)
-{
-       int clen = pampd_remote_size(pampd);
-       void *ret_pampd = NULL;
-       unsigned long flags;
-
-       if (!pampd_is_remote(pampd))
-               BUG();
-       if (is_ephemeral(pool))
-               BUG();
-       if (pampd_is_intransit(pampd)) {
-               /*
-                * to avoid multiple allocations (and maybe a memory leak)
-                * don't preallocate if already in the process of being
-                * repatriated
-                */
-               *intransit = true;
-               goto out;
-       }
-       *intransit = false;
-       local_irq_save(flags);
-       ret_pampd = (void *)zv_alloc(pool, oid, index, clen);
-       if (ret_pampd != NULL) {
-               /*
-                *  a pampd is marked intransit if it is remote and space has
-                *  been allocated for it locally (note, only happens for
-                *  persistent pages, in which case the remote copy is freed)
-                */
-               ret_pampd = pampd_mark_intransit(ret_pampd);
-               atomic_dec(&ramster_remote_pers_pages);
-               WARN_ON_ONCE(atomic_read(&ramster_remote_pers_pages) < 0);
-       } else
-               ramster_pers_pages_remote_nomem++;
-       local_irq_restore(flags);
-out:
-       return ret_pampd;
-}
-
-/*
- * Called on a remote tmem_get to invoke a message to fetch the page.
- * Might sleep so no tmem locks can be held.  "extra" is passed
- * all the way through the round-trip messaging to zcache_localify.
- */
-static int zcache_pampd_repatriate(void *fake_pampd, void *real_pampd,
-                                  struct tmem_pool *pool,
-                                  struct tmem_oid *oid, uint32_t index,
-                                  bool free, void *extra)
-{
-       struct tmem_xhandle xh;
-       int ret;
-
-       if (pampd_is_intransit(real_pampd))
-               /* have local space pre-reserved, so free remote copy */
-               free = true;
-       xh = tmem_xhandle_fill(LOCAL_CLIENT, pool, oid, index);
-       /* unreliable request/response for now */
-       ret = ramster_remote_async_get(&xh, free,
-                                       pampd_remote_node(fake_pampd),
-                                       pampd_remote_size(fake_pampd),
-                                       pampd_remote_cksum(fake_pampd),
-                                       extra);
-#ifdef RAMSTER_TESTING
-       if (ret != 0 && ret != -ENOENT)
-               pr_err("TESTING zcache_pampd_repatriate returns, ret=%d\n",
-                       ret);
-#endif
-       return ret;
-}
-
-static struct tmem_pamops zcache_pamops = {
-       .create = zcache_pampd_create,
-       .get_data = zcache_pampd_get_data,
-       .free = zcache_pampd_free,
-       .get_data_and_free = zcache_pampd_get_data_and_free,
-       .free_obj = zcache_pampd_free_obj,
-       .is_remote = zcache_pampd_is_remote,
-       .repatriate_preload = zcache_pampd_repatriate_preload,
-       .repatriate = zcache_pampd_repatriate,
-       .new_obj = zcache_pampd_new_obj,
-       .replace_in_obj = zcache_pampd_replace_in_obj,
-};
-
-/*
- * zcache compression/decompression and related per-cpu stuff
- */
-
-#define LZO_WORKMEM_BYTES LZO1X_1_MEM_COMPRESS
-#define LZO_DSTMEM_PAGE_ORDER 1
-static DEFINE_PER_CPU(unsigned char *, zcache_workmem);
-static DEFINE_PER_CPU(unsigned char *, zcache_dstmem);
-
-static int zcache_compress(struct page *from, void **out_va, size_t *out_len)
-{
-       int ret = 0;
-       unsigned char *dmem = __get_cpu_var(zcache_dstmem);
-       unsigned char *wmem = __get_cpu_var(zcache_workmem);
-       char *from_va;
-
-       BUG_ON(!irqs_disabled());
-       if (unlikely(dmem == NULL || wmem == NULL))
-               goto out;  /* no buffer, so can't compress */
-       from_va = kmap_atomic(from, KM_USER0);
-       mb();
-       ret = lzo1x_1_compress(from_va, PAGE_SIZE, dmem, out_len, wmem);
-       BUG_ON(ret != LZO_E_OK);
-       *out_va = dmem;
-       kunmap_atomic(from_va, KM_USER0);
-       ret = 1;
-out:
-       return ret;
-}
-
-
-static int zcache_cpu_notifier(struct notifier_block *nb,
-                               unsigned long action, void *pcpu)
-{
-       int cpu = (long)pcpu;
-       struct zcache_preload *kp;
-
-       switch (action) {
-       case CPU_UP_PREPARE:
-               per_cpu(zcache_dstmem, cpu) = (void *)__get_free_pages(
-                       GFP_KERNEL | __GFP_REPEAT,
-                       LZO_DSTMEM_PAGE_ORDER),
-               per_cpu(zcache_workmem, cpu) =
-                       kzalloc(LZO1X_MEM_COMPRESS,
-                               GFP_KERNEL | __GFP_REPEAT);
-               per_cpu(zcache_remoteputmem, cpu) =
-                       kzalloc(PAGE_SIZE, GFP_KERNEL | __GFP_REPEAT);
-               break;
-       case CPU_DEAD:
-       case CPU_UP_CANCELED:
-               kfree(per_cpu(zcache_remoteputmem, cpu));
-               per_cpu(zcache_remoteputmem, cpu) = NULL;
-               free_pages((unsigned long)per_cpu(zcache_dstmem, cpu),
-                               LZO_DSTMEM_PAGE_ORDER);
-               per_cpu(zcache_dstmem, cpu) = NULL;
-               kfree(per_cpu(zcache_workmem, cpu));
-               per_cpu(zcache_workmem, cpu) = NULL;
-               kp = &per_cpu(zcache_preloads, cpu);
-               while (kp->nr) {
-                       kmem_cache_free(zcache_objnode_cache,
-                                       kp->objnodes[kp->nr - 1]);
-                       kp->objnodes[kp->nr - 1] = NULL;
-                       kp->nr--;
-               }
-               if (kp->obj) {
-                       kmem_cache_free(zcache_obj_cache, kp->obj);
-                       kp->obj = NULL;
-               }
-               if (kp->flnode) {
-                       kmem_cache_free(ramster_flnode_cache, kp->flnode);
-                       kp->flnode = NULL;
-               }
-               if (kp->page) {
-                       free_page((unsigned long)kp->page);
-                       kp->page = NULL;
-               }
-               break;
-       default:
-               break;
-       }
-       return NOTIFY_OK;
-}
-
-static struct notifier_block zcache_cpu_notifier_block = {
-       .notifier_call = zcache_cpu_notifier
-};
-
-#ifdef CONFIG_SYSFS
-#define ZCACHE_SYSFS_RO(_name) \
-       static ssize_t zcache_##_name##_show(struct kobject *kobj, \
-                               struct kobj_attribute *attr, char *buf) \
-       { \
-               return sprintf(buf, "%lu\n", zcache_##_name); \
-       } \
-       static struct kobj_attribute zcache_##_name##_attr = { \
-               .attr = { .name = __stringify(_name), .mode = 0444 }, \
-               .show = zcache_##_name##_show, \
-       }
-
-#define ZCACHE_SYSFS_RO_ATOMIC(_name) \
-       static ssize_t zcache_##_name##_show(struct kobject *kobj, \
-                               struct kobj_attribute *attr, char *buf) \
-       { \
-           return sprintf(buf, "%d\n", atomic_read(&zcache_##_name)); \
-       } \
-       static struct kobj_attribute zcache_##_name##_attr = { \
-               .attr = { .name = __stringify(_name), .mode = 0444 }, \
-               .show = zcache_##_name##_show, \
-       }
-
-#define ZCACHE_SYSFS_RO_CUSTOM(_name, _func) \
-       static ssize_t zcache_##_name##_show(struct kobject *kobj, \
-                               struct kobj_attribute *attr, char *buf) \
-       { \
-           return _func(buf); \
-       } \
-       static struct kobj_attribute zcache_##_name##_attr = { \
-               .attr = { .name = __stringify(_name), .mode = 0444 }, \
-               .show = zcache_##_name##_show, \
-       }
-
-ZCACHE_SYSFS_RO(curr_obj_count_max);
-ZCACHE_SYSFS_RO(curr_objnode_count_max);
-ZCACHE_SYSFS_RO(flush_total);
-ZCACHE_SYSFS_RO(flush_found);
-ZCACHE_SYSFS_RO(flobj_total);
-ZCACHE_SYSFS_RO(flobj_found);
-ZCACHE_SYSFS_RO(failed_eph_puts);
-ZCACHE_SYSFS_RO(nonactive_puts);
-ZCACHE_SYSFS_RO(failed_pers_puts);
-ZCACHE_SYSFS_RO(zbud_curr_zbytes);
-ZCACHE_SYSFS_RO(zbud_cumul_zpages);
-ZCACHE_SYSFS_RO(zbud_cumul_zbytes);
-ZCACHE_SYSFS_RO(zbud_buddied_count);
-ZCACHE_SYSFS_RO(evicted_raw_pages);
-ZCACHE_SYSFS_RO(evicted_unbuddied_pages);
-ZCACHE_SYSFS_RO(evicted_buddied_pages);
-ZCACHE_SYSFS_RO(failed_get_free_pages);
-ZCACHE_SYSFS_RO(failed_alloc);
-ZCACHE_SYSFS_RO(put_to_flush);
-ZCACHE_SYSFS_RO(compress_poor);
-ZCACHE_SYSFS_RO(mean_compress_poor);
-ZCACHE_SYSFS_RO(policy_percent_exceeded);
-ZCACHE_SYSFS_RO_ATOMIC(zbud_curr_raw_pages);
-ZCACHE_SYSFS_RO_ATOMIC(zbud_curr_zpages);
-ZCACHE_SYSFS_RO_ATOMIC(curr_obj_count);
-ZCACHE_SYSFS_RO_ATOMIC(curr_objnode_count);
-ZCACHE_SYSFS_RO_CUSTOM(zbud_unbuddied_list_counts,
-                       zbud_show_unbuddied_list_counts);
-ZCACHE_SYSFS_RO_CUSTOM(zbud_cumul_chunk_counts,
-                       zbud_show_cumul_chunk_counts);
-ZCACHE_SYSFS_RO_CUSTOM(zv_curr_dist_counts,
-                       zv_curr_dist_counts_show);
-ZCACHE_SYSFS_RO_CUSTOM(zv_cumul_dist_counts,
-                       zv_cumul_dist_counts_show);
-
-static struct attribute *zcache_attrs[] = {
-       &zcache_curr_obj_count_attr.attr,
-       &zcache_curr_obj_count_max_attr.attr,
-       &zcache_curr_objnode_count_attr.attr,
-       &zcache_curr_objnode_count_max_attr.attr,
-       &zcache_flush_total_attr.attr,
-       &zcache_flobj_total_attr.attr,
-       &zcache_flush_found_attr.attr,
-       &zcache_flobj_found_attr.attr,
-       &zcache_failed_eph_puts_attr.attr,
-       &zcache_nonactive_puts_attr.attr,
-       &zcache_failed_pers_puts_attr.attr,
-       &zcache_policy_percent_exceeded_attr.attr,
-       &zcache_compress_poor_attr.attr,
-       &zcache_mean_compress_poor_attr.attr,
-       &zcache_zbud_curr_raw_pages_attr.attr,
-       &zcache_zbud_curr_zpages_attr.attr,
-       &zcache_zbud_curr_zbytes_attr.attr,
-       &zcache_zbud_cumul_zpages_attr.attr,
-       &zcache_zbud_cumul_zbytes_attr.attr,
-       &zcache_zbud_buddied_count_attr.attr,
-       &zcache_evicted_raw_pages_attr.attr,
-       &zcache_evicted_unbuddied_pages_attr.attr,
-       &zcache_evicted_buddied_pages_attr.attr,
-       &zcache_failed_get_free_pages_attr.attr,
-       &zcache_failed_alloc_attr.attr,
-       &zcache_put_to_flush_attr.attr,
-       &zcache_zbud_unbuddied_list_counts_attr.attr,
-       &zcache_zbud_cumul_chunk_counts_attr.attr,
-       &zcache_zv_curr_dist_counts_attr.attr,
-       &zcache_zv_cumul_dist_counts_attr.attr,
-       &zcache_zv_max_zsize_attr.attr,
-       &zcache_zv_max_mean_zsize_attr.attr,
-       &zcache_zv_page_count_policy_percent_attr.attr,
-       NULL,
-};
-
-static struct attribute_group zcache_attr_group = {
-       .attrs = zcache_attrs,
-       .name = "zcache",
-};
-
-#define RAMSTER_SYSFS_RO(_name) \
-       static ssize_t ramster_##_name##_show(struct kobject *kobj, \
-                               struct kobj_attribute *attr, char *buf) \
-       { \
-               return sprintf(buf, "%lu\n", ramster_##_name); \
-       } \
-       static struct kobj_attribute ramster_##_name##_attr = { \
-               .attr = { .name = __stringify(_name), .mode = 0444 }, \
-               .show = ramster_##_name##_show, \
-       }
-
-#define RAMSTER_SYSFS_RW(_name) \
-       static ssize_t ramster_##_name##_show(struct kobject *kobj, \
-                               struct kobj_attribute *attr, char *buf) \
-       { \
-               return sprintf(buf, "%lu\n", ramster_##_name); \
-       } \
-       static ssize_t ramster_##_name##_store(struct kobject *kobj, \
-               struct kobj_attribute *attr, const char *buf, size_t count) \
-       { \
-               int err; \
-               unsigned long enable; \
-               err = strict_strtoul(buf, 10, &enable); \
-               if (err) \
-                       return -EINVAL; \
-               ramster_##_name = enable; \
-               return count; \
-       } \
-       static struct kobj_attribute ramster_##_name##_attr = { \
-               .attr = { .name = __stringify(_name), .mode = 0644 }, \
-               .show = ramster_##_name##_show, \
-               .store = ramster_##_name##_store, \
-       }
-
-#define RAMSTER_SYSFS_RO_ATOMIC(_name) \
-       static ssize_t ramster_##_name##_show(struct kobject *kobj, \
-                               struct kobj_attribute *attr, char *buf) \
-       { \
-           return sprintf(buf, "%d\n", atomic_read(&ramster_##_name)); \
-       } \
-       static struct kobj_attribute ramster_##_name##_attr = { \
-               .attr = { .name = __stringify(_name), .mode = 0444 }, \
-               .show = ramster_##_name##_show, \
-       }
-
-RAMSTER_SYSFS_RO_ATOMIC(remote_pers_pages);
-RAMSTER_SYSFS_RW(pers_remotify_enable);
-RAMSTER_SYSFS_RW(eph_remotify_enable);
-RAMSTER_SYSFS_RO(eph_pages_remoted);
-RAMSTER_SYSFS_RO(eph_pages_remote_failed);
-RAMSTER_SYSFS_RO(pers_pages_remoted);
-RAMSTER_SYSFS_RO(pers_pages_remote_failed);
-RAMSTER_SYSFS_RO(pers_pages_remote_nomem);
-RAMSTER_SYSFS_RO(remote_pages_flushed);
-RAMSTER_SYSFS_RO(remote_page_flushes_failed);
-RAMSTER_SYSFS_RO(remote_objects_flushed);
-RAMSTER_SYSFS_RO(remote_object_flushes_failed);
-RAMSTER_SYSFS_RO(remote_eph_pages_succ_get);
-RAMSTER_SYSFS_RO(remote_eph_pages_unsucc_get);
-RAMSTER_SYSFS_RO(remote_pers_pages_succ_get);
-RAMSTER_SYSFS_RO(remote_pers_pages_unsucc_get);
-RAMSTER_SYSFS_RO_ATOMIC(foreign_eph_pampd_count);
-RAMSTER_SYSFS_RO(foreign_eph_pampd_count_max);
-RAMSTER_SYSFS_RO_ATOMIC(foreign_pers_pampd_count);
-RAMSTER_SYSFS_RO(foreign_pers_pampd_count_max);
-RAMSTER_SYSFS_RO_ATOMIC(curr_flnode_count);
-RAMSTER_SYSFS_RO(curr_flnode_count_max);
-
-#define MANUAL_NODES 8
-static bool ramster_nodes_manual_up[MANUAL_NODES];
-static ssize_t ramster_manual_node_up_show(struct kobject *kobj,
-                               struct kobj_attribute *attr, char *buf)
-{
-       int i;
-       char *p = buf;
-       for (i = 0; i < MANUAL_NODES; i++)
-               if (ramster_nodes_manual_up[i])
-                       p += sprintf(p, "%d ", i);
-       p += sprintf(p, "\n");
-       return p - buf;
-}
-
-static ssize_t ramster_manual_node_up_store(struct kobject *kobj,
-               struct kobj_attribute *attr, const char *buf, size_t count)
-{
-       int err;
-       unsigned long node_num;
-
-       err = strict_strtoul(buf, 10, &node_num);
-       if (err) {
-               pr_err("bad strtoul?\n");
-               return -EINVAL;
-       }
-       if (node_num >= MANUAL_NODES) {
-               pr_err("bad node_num=%lu?\n", node_num);
-               return -EINVAL;
-       }
-       if (ramster_nodes_manual_up[node_num]) {
-               pr_err("node %d already up, ignoring\n", (int)node_num);
-       } else {
-               ramster_nodes_manual_up[node_num] = true;
-               o2net_hb_node_up_manual((int)node_num);
-       }
-       return count;
-}
-
-static struct kobj_attribute ramster_manual_node_up_attr = {
-       .attr = { .name = "manual_node_up", .mode = 0644 },
-       .show = ramster_manual_node_up_show,
-       .store = ramster_manual_node_up_store,
-};
-
-static struct attribute *ramster_attrs[] = {
-       &ramster_pers_remotify_enable_attr.attr,
-       &ramster_eph_remotify_enable_attr.attr,
-       &ramster_remote_pers_pages_attr.attr,
-       &ramster_eph_pages_remoted_attr.attr,
-       &ramster_eph_pages_remote_failed_attr.attr,
-       &ramster_pers_pages_remoted_attr.attr,
-       &ramster_pers_pages_remote_failed_attr.attr,
-       &ramster_pers_pages_remote_nomem_attr.attr,
-       &ramster_remote_pages_flushed_attr.attr,
-       &ramster_remote_page_flushes_failed_attr.attr,
-       &ramster_remote_objects_flushed_attr.attr,
-       &ramster_remote_object_flushes_failed_attr.attr,
-       &ramster_remote_eph_pages_succ_get_attr.attr,
-       &ramster_remote_eph_pages_unsucc_get_attr.attr,
-       &ramster_remote_pers_pages_succ_get_attr.attr,
-       &ramster_remote_pers_pages_unsucc_get_attr.attr,
-       &ramster_foreign_eph_pampd_count_attr.attr,
-       &ramster_foreign_eph_pampd_count_max_attr.attr,
-       &ramster_foreign_pers_pampd_count_attr.attr,
-       &ramster_foreign_pers_pampd_count_max_attr.attr,
-       &ramster_curr_flnode_count_attr.attr,
-       &ramster_curr_flnode_count_max_attr.attr,
-       &ramster_manual_node_up_attr.attr,
-       NULL,
-};
-
-static struct attribute_group ramster_attr_group = {
-       .attrs = ramster_attrs,
-       .name = "ramster",
-};
-
-#endif /* CONFIG_SYSFS */
-/*
- * When zcache is disabled ("frozen"), pools can be created and destroyed,
- * but all puts (and thus all other operations that require memory allocation)
- * must fail.  If zcache is unfrozen, accepts puts, then frozen again,
- * data consistency requires all puts while frozen to be converted into
- * flushes.
- */
-static bool zcache_freeze;
-
-/*
- * zcache shrinker interface (only useful for ephemeral pages, so zbud only)
- */
-static int shrink_zcache_memory(struct shrinker *shrink,
-                               struct shrink_control *sc)
-{
-       int ret = -1;
-       int nr = sc->nr_to_scan;
-       gfp_t gfp_mask = sc->gfp_mask;
-
-       if (nr >= 0) {
-               if (!(gfp_mask & __GFP_FS))
-                       /* does this case really need to be skipped? */
-                       goto out;
-               zbud_evict_pages(nr);
-       }
-       ret = (int)atomic_read(&zcache_zbud_curr_raw_pages);
-out:
-       return ret;
-}
-
-static struct shrinker zcache_shrinker = {
-       .shrink = shrink_zcache_memory,
-       .seeks = DEFAULT_SEEKS,
-};
-
-/*
- * zcache shims between cleancache/frontswap ops and tmem
- */
-
-int zcache_put(int cli_id, int pool_id, struct tmem_oid *oidp,
-                       uint32_t index, char *data, size_t size,
-                       bool raw, int ephemeral)
-{
-       struct tmem_pool *pool;
-       int ret = -1;
-
-       BUG_ON(!irqs_disabled());
-       pool = zcache_get_pool_by_id(cli_id, pool_id);
-       if (unlikely(pool == NULL))
-               goto out;
-       if (!zcache_freeze && zcache_do_preload(pool) == 0) {
-               /* preload does preempt_disable on success */
-               ret = tmem_put(pool, oidp, index, data, size, raw, ephemeral);
-               if (ret < 0) {
-                       if (is_ephemeral(pool))
-                               zcache_failed_eph_puts++;
-                       else
-                               zcache_failed_pers_puts++;
-               }
-               zcache_put_pool(pool);
-               preempt_enable_no_resched();
-       } else {
-               zcache_put_to_flush++;
-               if (atomic_read(&pool->obj_count) > 0)
-                       /* the put fails whether the flush succeeds or not */
-                       (void)tmem_flush_page(pool, oidp, index);
-               zcache_put_pool(pool);
-       }
-out:
-       return ret;
-}
-
-int zcache_get(int cli_id, int pool_id, struct tmem_oid *oidp,
-                       uint32_t index, char *data, size_t *sizep,
-                       bool raw, int get_and_free)
-{
-       struct tmem_pool *pool;
-       int ret = -1;
-       bool eph;
-
-       if (!raw) {
-               BUG_ON(irqs_disabled());
-               BUG_ON(in_softirq());
-       }
-       pool = zcache_get_pool_by_id(cli_id, pool_id);
-       eph = is_ephemeral(pool);
-       if (likely(pool != NULL)) {
-               if (atomic_read(&pool->obj_count) > 0)
-                       ret = tmem_get(pool, oidp, index, data, sizep,
-                                       raw, get_and_free);
-               zcache_put_pool(pool);
-       }
-       WARN_ONCE((!eph && (ret != 0)), "zcache_get fails on persistent pool, "
-                         "bad things are very likely to happen soon\n");
-#ifdef RAMSTER_TESTING
-       if (ret != 0 && ret != -1 && !(ret == -EINVAL && is_ephemeral(pool)))
-               pr_err("TESTING zcache_get tmem_get returns ret=%d\n", ret);
-#endif
-       if (ret == -EAGAIN)
-               BUG(); /* FIXME... don't need this anymore??? let's ensure */
-       return ret;
-}
-
-int zcache_flush(int cli_id, int pool_id,
-                               struct tmem_oid *oidp, uint32_t index)
-{
-       struct tmem_pool *pool;
-       int ret = -1;
-       unsigned long flags;
-
-       local_irq_save(flags);
-       zcache_flush_total++;
-       pool = zcache_get_pool_by_id(cli_id, pool_id);
-       ramster_do_preload_flnode_only(pool);
-       if (likely(pool != NULL)) {
-               if (atomic_read(&pool->obj_count) > 0)
-                       ret = tmem_flush_page(pool, oidp, index);
-               zcache_put_pool(pool);
-       }
-       if (ret >= 0)
-               zcache_flush_found++;
-       local_irq_restore(flags);
-       return ret;
-}
-
-int zcache_flush_object(int cli_id, int pool_id, struct tmem_oid *oidp)
-{
-       struct tmem_pool *pool;
-       int ret = -1;
-       unsigned long flags;
-
-       local_irq_save(flags);
-       zcache_flobj_total++;
-       pool = zcache_get_pool_by_id(cli_id, pool_id);
-       ramster_do_preload_flnode_only(pool);
-       if (likely(pool != NULL)) {
-               if (atomic_read(&pool->obj_count) > 0)
-                       ret = tmem_flush_object(pool, oidp);
-               zcache_put_pool(pool);
-       }
-       if (ret >= 0)
-               zcache_flobj_found++;
-       local_irq_restore(flags);
-       return ret;
-}
-
-int zcache_client_destroy_pool(int cli_id, int pool_id)
-{
-       struct tmem_pool *pool = NULL;
-       struct zcache_client *cli = NULL;
-       int ret = -1;
-
-       if (pool_id < 0)
-               goto out;
-       if (cli_id == LOCAL_CLIENT)
-               cli = &zcache_host;
-       else if ((unsigned int)cli_id < MAX_CLIENTS)
-               cli = &zcache_clients[cli_id];
-       if (cli == NULL)
-               goto out;
-       atomic_inc(&cli->refcount);
-       pool = cli->tmem_pools[pool_id];
-       if (pool == NULL)
-               goto out;
-       cli->tmem_pools[pool_id] = NULL;
-       /* wait for pool activity on other cpus to quiesce */
-       while (atomic_read(&pool->refcount) != 0)
-               ;
-       atomic_dec(&cli->refcount);
-       local_bh_disable();
-       ret = tmem_destroy_pool(pool);
-       local_bh_enable();
-       kfree(pool);
-       pr_info("ramster: destroyed pool id=%d cli_id=%d\n", pool_id, cli_id);
-out:
-       return ret;
-}
-
-static int zcache_destroy_pool(int pool_id)
-{
-       return zcache_client_destroy_pool(LOCAL_CLIENT, pool_id);
-}
-
-int zcache_new_pool(uint16_t cli_id, uint32_t flags)
-{
-       int poolid = -1;
-       struct tmem_pool *pool;
-       struct zcache_client *cli = NULL;
-
-       if (cli_id == LOCAL_CLIENT)
-               cli = &zcache_host;
-       else if ((unsigned int)cli_id < MAX_CLIENTS)
-               cli = &zcache_clients[cli_id];
-       if (cli == NULL)
-               goto out;
-       atomic_inc(&cli->refcount);
-       pool = kmalloc(sizeof(struct tmem_pool), GFP_ATOMIC);
-       if (pool == NULL) {
-               pr_info("ramster: pool creation failed: out of memory\n");
-               goto out;
-       }
-
-       for (poolid = 0; poolid < MAX_POOLS_PER_CLIENT; poolid++)
-               if (cli->tmem_pools[poolid] == NULL)
-                       break;
-       if (poolid >= MAX_POOLS_PER_CLIENT) {
-               pr_info("ramster: pool creation failed: max exceeded\n");
-               kfree(pool);
-               poolid = -1;
-               goto out;
-       }
-       atomic_set(&pool->refcount, 0);
-       pool->client = cli;
-       pool->pool_id = poolid;
-       tmem_new_pool(pool, flags);
-       cli->tmem_pools[poolid] = pool;
-       pr_info("ramster: created %s tmem pool, id=%d, client=%d\n",
-               flags & TMEM_POOL_PERSIST ? "persistent" : "ephemeral",
-               poolid, cli_id);
-out:
-       if (cli != NULL)
-               atomic_dec(&cli->refcount);
-       return poolid;
-}
-
-static int zcache_local_new_pool(uint32_t flags)
-{
-       return zcache_new_pool(LOCAL_CLIENT, flags);
-}
-
-int zcache_autocreate_pool(int cli_id, int pool_id, bool ephemeral)
-{
-       struct tmem_pool *pool;
-       struct zcache_client *cli = NULL;
-       uint32_t flags = ephemeral ? 0 : TMEM_POOL_PERSIST;
-       int ret = -1;
-
-       if (cli_id == LOCAL_CLIENT)
-               goto out;
-       if (pool_id >= MAX_POOLS_PER_CLIENT)
-               goto out;
-       else if ((unsigned int)cli_id < MAX_CLIENTS)
-               cli = &zcache_clients[cli_id];
-       if ((ephemeral && !use_cleancache) || (!ephemeral && !use_frontswap))
-               BUG(); /* FIXME, handle more gracefully later */
-       if (!cli->allocated) {
-               if (zcache_new_client(cli_id))
-                       BUG(); /* FIXME, handle more gracefully later */
-               cli = &zcache_clients[cli_id];
-       }
-       atomic_inc(&cli->refcount);
-       pool = cli->tmem_pools[pool_id];
-       if (pool != NULL) {
-               if (pool->persistent && ephemeral) {
-                       pr_err("zcache_autocreate_pool: type mismatch\n");
-                       goto out;
-               }
-               ret = 0;
-               goto out;
-       }
-       pool = kmalloc(sizeof(struct tmem_pool), GFP_KERNEL);
-       if (pool == NULL) {
-               pr_info("ramster: pool creation failed: out of memory\n");
-               goto out;
-       }
-       atomic_set(&pool->refcount, 0);
-       pool->client = cli;
-       pool->pool_id = pool_id;
-       tmem_new_pool(pool, flags);
-       cli->tmem_pools[pool_id] = pool;
-       pr_info("ramster: AUTOcreated %s tmem poolid=%d, for remote client=%d\n",
-               flags & TMEM_POOL_PERSIST ? "persistent" : "ephemeral",
-               pool_id, cli_id);
-       ret = 0;
-out:
-       if (cli == NULL)
-               BUG(); /* FIXME, handle more gracefully later */
-               /* pr_err("zcache_autocreate_pool: failed\n"); */
-       if (cli != NULL)
-               atomic_dec(&cli->refcount);
-       return ret;
-}
-
-/**********
- * Two kernel functionalities currently can be layered on top of tmem.
- * These are "cleancache" which is used as a second-chance cache for clean
- * page cache pages; and "frontswap" which is used for swap pages
- * to avoid writes to disk.  A generic "shim" is provided here for each
- * to translate in-kernel semantics to zcache semantics.
- */
-
-#ifdef CONFIG_CLEANCACHE
-static void zcache_cleancache_put_page(int pool_id,
-                                       struct cleancache_filekey key,
-                                       pgoff_t index, struct page *page)
-{
-       u32 ind = (u32) index;
-       struct tmem_oid oid = *(struct tmem_oid *)&key;
-
-#ifdef __PG_WAS_ACTIVE
-       if (!PageWasActive(page)) {
-               zcache_nonactive_puts++;
-               return;
-       }
-#endif
-       if (likely(ind == index)) {
-               char *kva = page_address(page);
-
-               (void)zcache_put(LOCAL_CLIENT, pool_id, &oid, index,
-                       kva, PAGE_SIZE, 0, 1);
-       }
-}
-
-static int zcache_cleancache_get_page(int pool_id,
-                                       struct cleancache_filekey key,
-                                       pgoff_t index, struct page *page)
-{
-       u32 ind = (u32) index;
-       struct tmem_oid oid = *(struct tmem_oid *)&key;
-       int ret = -1;
-
-       preempt_disable();
-       if (likely(ind == index)) {
-               char *kva = page_address(page);
-               size_t size = PAGE_SIZE;
-
-               ret = zcache_get(LOCAL_CLIENT, pool_id, &oid, index,
-                       kva, &size, 0, 0);
-#ifdef __PG_WAS_ACTIVE
-               if (ret == 0)
-                       SetPageWasActive(page);
-#endif
-       }
-       preempt_enable();
-       return ret;
-}
-
-static void zcache_cleancache_flush_page(int pool_id,
-                                       struct cleancache_filekey key,
-                                       pgoff_t index)
-{
-       u32 ind = (u32) index;
-       struct tmem_oid oid = *(struct tmem_oid *)&key;
-
-       if (likely(ind == index))
-               (void)zcache_flush(LOCAL_CLIENT, pool_id, &oid, ind);
-}
-
-static void zcache_cleancache_flush_inode(int pool_id,
-                                       struct cleancache_filekey key)
-{
-       struct tmem_oid oid = *(struct tmem_oid *)&key;
-
-       (void)zcache_flush_object(LOCAL_CLIENT, pool_id, &oid);
-}
-
-static void zcache_cleancache_flush_fs(int pool_id)
-{
-       if (pool_id >= 0)
-               (void)zcache_destroy_pool(pool_id);
-}
-
-static int zcache_cleancache_init_fs(size_t pagesize)
-{
-       BUG_ON(sizeof(struct cleancache_filekey) !=
-                               sizeof(struct tmem_oid));
-       BUG_ON(pagesize != PAGE_SIZE);
-       return zcache_local_new_pool(0);
-}
-
-static int zcache_cleancache_init_shared_fs(char *uuid, size_t pagesize)
-{
-       /* shared pools are unsupported and map to private */
-       BUG_ON(sizeof(struct cleancache_filekey) !=
-                               sizeof(struct tmem_oid));
-       BUG_ON(pagesize != PAGE_SIZE);
-       return zcache_local_new_pool(0);
-}
-
-static struct cleancache_ops zcache_cleancache_ops = {
-       .put_page = zcache_cleancache_put_page,
-       .get_page = zcache_cleancache_get_page,
-       .invalidate_page = zcache_cleancache_flush_page,
-       .invalidate_inode = zcache_cleancache_flush_inode,
-       .invalidate_fs = zcache_cleancache_flush_fs,
-       .init_shared_fs = zcache_cleancache_init_shared_fs,
-       .init_fs = zcache_cleancache_init_fs
-};
-
-struct cleancache_ops zcache_cleancache_register_ops(void)
-{
-       struct cleancache_ops old_ops =
-               cleancache_register_ops(&zcache_cleancache_ops);
-
-       return old_ops;
-}
-#endif
-
-#ifdef CONFIG_FRONTSWAP
-/* a single tmem poolid is used for all frontswap "types" (swapfiles) */
-static int zcache_frontswap_poolid = -1;
-
-/*
- * Swizzling increases objects per swaptype, increasing tmem concurrency
- * for heavy swaploads.  Later, larger nr_cpus -> larger SWIZ_BITS
- */
-#define SWIZ_BITS              8
-#define SWIZ_MASK              ((1 << SWIZ_BITS) - 1)
-#define _oswiz(_type, _ind)    ((_type << SWIZ_BITS) | (_ind & SWIZ_MASK))
-#define iswiz(_ind)            (_ind >> SWIZ_BITS)
-
-static inline struct tmem_oid oswiz(unsigned type, u32 ind)
-{
-       struct tmem_oid oid = { .oid = { 0 } };
-       oid.oid[0] = _oswiz(type, ind);
-       return oid;
-}
-
-static int zcache_frontswap_put_page(unsigned type, pgoff_t offset,
-                                  struct page *page)
-{
-       u64 ind64 = (u64)offset;
-       u32 ind = (u32)offset;
-       struct tmem_oid oid = oswiz(type, ind);
-       int ret = -1;
-       unsigned long flags;
-       char *kva;
-
-       BUG_ON(!PageLocked(page));
-       if (likely(ind64 == ind)) {
-               local_irq_save(flags);
-               kva = page_address(page);
-               ret = zcache_put(LOCAL_CLIENT, zcache_frontswap_poolid,
-                               &oid, iswiz(ind), kva, PAGE_SIZE, 0, 0);
-               local_irq_restore(flags);
-       }
-       return ret;
-}
-
-/* returns 0 if the page was successfully gotten from frontswap, -1 if
- * was not present (should never happen!) */
-static int zcache_frontswap_get_page(unsigned type, pgoff_t offset,
-                                  struct page *page)
-{
-       u64 ind64 = (u64)offset;
-       u32 ind = (u32)offset;
-       struct tmem_oid oid = oswiz(type, ind);
-       int ret = -1;
-
-       preempt_disable(); /* FIXME, remove this? */
-       BUG_ON(!PageLocked(page));
-       if (likely(ind64 == ind)) {
-               char *kva = page_address(page);
-               size_t size = PAGE_SIZE;
-
-               ret = zcache_get(LOCAL_CLIENT, zcache_frontswap_poolid,
-                                       &oid, iswiz(ind), kva, &size, 0, -1);
-       }
-       preempt_enable(); /* FIXME, remove this? */
-       return ret;
-}
-
-/* flush a single page from frontswap */
-static void zcache_frontswap_flush_page(unsigned type, pgoff_t offset)
-{
-       u64 ind64 = (u64)offset;
-       u32 ind = (u32)offset;
-       struct tmem_oid oid = oswiz(type, ind);
-
-       if (likely(ind64 == ind))
-               (void)zcache_flush(LOCAL_CLIENT, zcache_frontswap_poolid,
-                                       &oid, iswiz(ind));
-}
-
-/* flush all pages from the passed swaptype */
-static void zcache_frontswap_flush_area(unsigned type)
-{
-       struct tmem_oid oid;
-       int ind;
-
-       for (ind = SWIZ_MASK; ind >= 0; ind--) {
-               oid = oswiz(type, ind);
-               (void)zcache_flush_object(LOCAL_CLIENT,
-                                               zcache_frontswap_poolid, &oid);
-       }
-}
-
-static void zcache_frontswap_init(unsigned ignored)
-{
-       /* a single tmem poolid is used for all frontswap "types" (swapfiles) */
-       if (zcache_frontswap_poolid < 0)
-               zcache_frontswap_poolid =
-                               zcache_local_new_pool(TMEM_POOL_PERSIST);
-}
-
-static struct frontswap_ops zcache_frontswap_ops = {
-       .put_page = zcache_frontswap_put_page,
-       .get_page = zcache_frontswap_get_page,
-       .invalidate_page = zcache_frontswap_flush_page,
-       .invalidate_area = zcache_frontswap_flush_area,
-       .init = zcache_frontswap_init
-};
-
-struct frontswap_ops zcache_frontswap_register_ops(void)
-{
-       struct frontswap_ops old_ops =
-               frontswap_register_ops(&zcache_frontswap_ops);
-
-       return old_ops;
-}
-#endif
-
-/*
- * frontswap selfshrinking
- */
-
-#ifdef CONFIG_FRONTSWAP
-/* In HZ, controls frequency of worker invocation. */
-static unsigned int selfshrink_interval __read_mostly = 5;
-
-static void selfshrink_process(struct work_struct *work);
-static DECLARE_DELAYED_WORK(selfshrink_worker, selfshrink_process);
-
-/* Enable/disable with sysfs. */
-static bool frontswap_selfshrinking __read_mostly;
-
-/* Enable/disable with kernel boot option. */
-static bool use_frontswap_selfshrink __initdata = true;
-
-/*
- * The default values for the following parameters were deemed reasonable
- * by experimentation, may be workload-dependent, and can all be
- * adjusted via sysfs.
- */
-
-/* Control rate for frontswap shrinking. Higher hysteresis is slower. */
-static unsigned int frontswap_hysteresis __read_mostly = 20;
-
-/*
- * Number of selfshrink worker invocations to wait before observing that
- * frontswap selfshrinking should commence. Note that selfshrinking does
- * not use a separate worker thread.
- */
-static unsigned int frontswap_inertia __read_mostly = 3;
-
-/* Countdown to next invocation of frontswap_shrink() */
-static unsigned long frontswap_inertia_counter;
-
-/*
- * Invoked by the selfshrink worker thread, uses current number of pages
- * in frontswap (frontswap_curr_pages()), previous status, and control
- * values (hysteresis and inertia) to determine if frontswap should be
- * shrunk and what the new frontswap size should be.  Note that
- * frontswap_shrink is essentially a partial swapoff that immediately
- * transfers pages from the "swap device" (frontswap) back into kernel
- * RAM; despite the name, frontswap "shrinking" is very different from
- * the "shrinker" interface used by the kernel MM subsystem to reclaim
- * memory.
- */
-static void frontswap_selfshrink(void)
-{
-       static unsigned long cur_frontswap_pages;
-       static unsigned long last_frontswap_pages;
-       static unsigned long tgt_frontswap_pages;
-
-       last_frontswap_pages = cur_frontswap_pages;
-       cur_frontswap_pages = frontswap_curr_pages();
-       if (!cur_frontswap_pages ||
-                       (cur_frontswap_pages > last_frontswap_pages)) {
-               frontswap_inertia_counter = frontswap_inertia;
-               return;
-       }
-       if (frontswap_inertia_counter && --frontswap_inertia_counter)
-               return;
-       if (cur_frontswap_pages <= frontswap_hysteresis)
-               tgt_frontswap_pages = 0;
-       else
-               tgt_frontswap_pages = cur_frontswap_pages -
-                       (cur_frontswap_pages / frontswap_hysteresis);
-       frontswap_shrink(tgt_frontswap_pages);
-}
-
-static int __init ramster_nofrontswap_selfshrink_setup(char *s)
-{
-       use_frontswap_selfshrink = false;
-       return 1;
-}
-
-__setup("noselfshrink", ramster_nofrontswap_selfshrink_setup);
-
-static void selfshrink_process(struct work_struct *work)
-{
-       if (frontswap_selfshrinking && frontswap_enabled) {
-               frontswap_selfshrink();
-               schedule_delayed_work(&selfshrink_worker,
-                       selfshrink_interval * HZ);
-       }
-}
-
-static int ramster_enabled;
-
-static int __init ramster_selfshrink_init(void)
-{
-       frontswap_selfshrinking = ramster_enabled && use_frontswap_selfshrink;
-       if (frontswap_selfshrinking)
-               pr_info("ramster: Initializing frontswap "
-                                       "selfshrinking driver.\n");
-       else
-               return -ENODEV;
-
-       schedule_delayed_work(&selfshrink_worker, selfshrink_interval * HZ);
-
-       return 0;
-}
-
-subsys_initcall(ramster_selfshrink_init);
-#endif
-
-/*
- * zcache initialization
- * NOTE FOR NOW ramster MUST BE PROVIDED AS A KERNEL BOOT PARAMETER OR
- * NOTHING HAPPENS!
- */
-
-static int ramster_enabled;
-
-static int __init enable_ramster(char *s)
-{
-       ramster_enabled = 1;
-       return 1;
-}
-__setup("ramster", enable_ramster);
-
-/* allow independent dynamic disabling of cleancache and frontswap */
-
-static int use_cleancache = 1;
-
-static int __init no_cleancache(char *s)
-{
-       pr_info("INIT no_cleancache called\n");
-       use_cleancache = 0;
-       return 1;
-}
-
-/*
- * FIXME: need to guarantee this gets checked before zcache_init is called
- * What is the correct way to achieve this?
- */
-early_param("nocleancache", no_cleancache);
-
-static int use_frontswap = 1;
-
-static int __init no_frontswap(char *s)
-{
-       pr_info("INIT no_frontswap called\n");
-       use_frontswap = 0;
-       return 1;
-}
-
-__setup("nofrontswap", no_frontswap);
-
-static int __init zcache_init(void)
-{
-       int ret = 0;
-
-#ifdef CONFIG_SYSFS
-       ret = sysfs_create_group(mm_kobj, &zcache_attr_group);
-       ret = sysfs_create_group(mm_kobj, &ramster_attr_group);
-       if (ret) {
-               pr_err("ramster: can't create sysfs\n");
-               goto out;
-       }
-#endif /* CONFIG_SYSFS */
-#if defined(CONFIG_CLEANCACHE) || defined(CONFIG_FRONTSWAP)
-       if (ramster_enabled) {
-               unsigned int cpu;
-
-               (void)ramster_o2net_register_handlers();
-               tmem_register_hostops(&zcache_hostops);
-               tmem_register_pamops(&zcache_pamops);
-               ret = register_cpu_notifier(&zcache_cpu_notifier_block);
-               if (ret) {
-                       pr_err("ramster: can't register cpu notifier\n");
-                       goto out;
-               }
-               for_each_online_cpu(cpu) {
-                       void *pcpu = (void *)(long)cpu;
-                       zcache_cpu_notifier(&zcache_cpu_notifier_block,
-                               CPU_UP_PREPARE, pcpu);
-               }
-       }
-       zcache_objnode_cache = kmem_cache_create("zcache_objnode",
-                               sizeof(struct tmem_objnode), 0, 0, NULL);
-       zcache_obj_cache = kmem_cache_create("zcache_obj",
-                               sizeof(struct tmem_obj), 0, 0, NULL);
-       ramster_flnode_cache = kmem_cache_create("ramster_flnode",
-                               sizeof(struct flushlist_node), 0, 0, NULL);
-#endif
-#ifdef CONFIG_CLEANCACHE
-       pr_info("INIT ramster_enabled=%d use_cleancache=%d\n",
-                                       ramster_enabled, use_cleancache);
-       if (ramster_enabled && use_cleancache) {
-               struct cleancache_ops old_ops;
-
-               zbud_init();
-               register_shrinker(&zcache_shrinker);
-               old_ops = zcache_cleancache_register_ops();
-               pr_info("ramster: cleancache enabled using kernel "
-                       "transcendent memory and compression buddies\n");
-               if (old_ops.init_fs != NULL)
-                       pr_warning("ramster: cleancache_ops overridden");
-       }
-#endif
-#ifdef CONFIG_FRONTSWAP
-       pr_info("INIT ramster_enabled=%d use_frontswap=%d\n",
-                                       ramster_enabled, use_frontswap);
-       if (ramster_enabled && use_frontswap) {
-               struct frontswap_ops old_ops;
-
-               zcache_new_client(LOCAL_CLIENT);
-               old_ops = zcache_frontswap_register_ops();
-               pr_info("ramster: frontswap enabled using kernel "
-                       "transcendent memory and xvmalloc\n");
-               if (old_ops.init != NULL)
-                       pr_warning("ramster: frontswap_ops overridden");
-       }
-       if (ramster_enabled && (use_frontswap || use_cleancache))
-               ramster_remotify_init();
-#endif
-out:
-       return ret;
-}
-
-module_init(zcache_init)
diff --git a/drivers/staging/ramster/zcache.h b/drivers/staging/ramster/zcache.h
deleted file mode 100644 (file)
index 250b121..0000000
+++ /dev/null
@@ -1,22 +0,0 @@
-/*
- * zcache.h
- *
- * External zcache functions
- *
- * Copyright (c) 2009-2012, Dan Magenheimer, Oracle Corp.
- */
-
-#ifndef _ZCACHE_H_
-#define _ZCACHE_H_
-
-extern int zcache_put(int, int, struct tmem_oid *, uint32_t,
-                       char *, size_t, bool, int);
-extern int zcache_autocreate_pool(int, int, bool);
-extern int zcache_get(int, int, struct tmem_oid *, uint32_t,
-                       char *, size_t *, bool, int);
-extern int zcache_flush(int, int, struct tmem_oid *, uint32_t);
-extern int zcache_flush_object(int, int, struct tmem_oid *);
-extern int zcache_localify(int, struct tmem_oid *, uint32_t,
-                       char *, size_t, void *);
-
-#endif /* _ZCACHE_H */