[funini.com] -> [kei@sodan] -> Kernel Reading

root/fs/xfs/support/ktrace.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. ktrace_init
  2. ktrace_uninit
  3. ktrace_alloc
  4. ktrace_free
  5. ktrace_enter
  6. ktrace_nentries
  7. ktrace_first
  8. ktrace_next
  9. ktrace_skip

/*
 * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
 * All Rights Reserved.
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public License as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it would be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write the Free Software Foundation,
 * Inc.,  51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA
 */
#include <xfs.h>

static kmem_zone_t *ktrace_hdr_zone;
static kmem_zone_t *ktrace_ent_zone;
static int          ktrace_zentries;

void __init
ktrace_init(int zentries)
{
        ktrace_zentries = roundup_pow_of_two(zentries);

        ktrace_hdr_zone = kmem_zone_init(sizeof(ktrace_t),
                                        "ktrace_hdr");
        ASSERT(ktrace_hdr_zone);

        ktrace_ent_zone = kmem_zone_init(ktrace_zentries
                                        * sizeof(ktrace_entry_t),
                                        "ktrace_ent");
        ASSERT(ktrace_ent_zone);
}

void __exit
ktrace_uninit(void)
{
        kmem_zone_destroy(ktrace_hdr_zone);
        kmem_zone_destroy(ktrace_ent_zone);
}

/*
 * ktrace_alloc()
 *
 * Allocate a ktrace header and enough buffering for the given
 * number of entries. Round the number of entries up to a
 * power of 2 so we can do fast masking to get the index from
 * the atomic index counter.
 */
ktrace_t *
ktrace_alloc(int nentries, unsigned int __nocast sleep)
{
        ktrace_t        *ktp;
        ktrace_entry_t  *ktep;
        int             entries;

        ktp = (ktrace_t*)kmem_zone_alloc(ktrace_hdr_zone, sleep);

        if (ktp == (ktrace_t*)NULL) {
                /*
                 * KM_SLEEP callers don't expect failure.
                 */
                if (sleep & KM_SLEEP)
                        panic("ktrace_alloc: NULL memory on KM_SLEEP request!");

                return NULL;
        }

        /*
         * Special treatment for buffers with the ktrace_zentries entries
         */
        entries = roundup_pow_of_two(nentries);
        if (entries == ktrace_zentries) {
                ktep = (ktrace_entry_t*)kmem_zone_zalloc(ktrace_ent_zone,
                                                            sleep);
        } else {
                ktep = (ktrace_entry_t*)kmem_zalloc((entries * sizeof(*ktep)),
                                                            sleep | KM_LARGE);
        }

        if (ktep == NULL) {
                /*
                 * KM_SLEEP callers don't expect failure.
                 */
                if (sleep & KM_SLEEP)
                        panic("ktrace_alloc: NULL memory on KM_SLEEP request!");

                kmem_free(ktp);

                return NULL;
        }

        ktp->kt_entries  = ktep;
        ktp->kt_nentries = entries;
        ASSERT(is_power_of_2(entries));
        ktp->kt_index_mask = entries - 1;
        atomic_set(&ktp->kt_index, 0);
        ktp->kt_rollover = 0;
        return ktp;
}


/*
 * ktrace_free()
 *
 * Free up the ktrace header and buffer.  It is up to the caller
 * to ensure that no-one is referencing it.
 */
void
ktrace_free(ktrace_t *ktp)
{
        int     entries_size;

        if (ktp == (ktrace_t *)NULL)
                return;

        /*
         * Special treatment for the Vnode trace buffer.
         */
        if (ktp->kt_nentries == ktrace_zentries) {
                kmem_zone_free(ktrace_ent_zone, ktp->kt_entries);
        } else {
                entries_size = (int)(ktp->kt_nentries * sizeof(ktrace_entry_t));

                kmem_free(ktp->kt_entries);
        }

        kmem_zone_free(ktrace_hdr_zone, ktp);
}


/*
 * Enter the given values into the "next" entry in the trace buffer.
 * kt_index is always the index of the next entry to be filled.
 */
void
ktrace_enter(
        ktrace_t        *ktp,
        void            *val0,
        void            *val1,
        void            *val2,
        void            *val3,
        void            *val4,
        void            *val5,
        void            *val6,
        void            *val7,
        void            *val8,
        void            *val9,
        void            *val10,
        void            *val11,
        void            *val12,
        void            *val13,
        void            *val14,
        void            *val15)
{
        int             index;
        ktrace_entry_t  *ktep;

        ASSERT(ktp != NULL);

        /*
         * Grab an entry by pushing the index up to the next one.
         */
        index = atomic_add_return(1, &ktp->kt_index);
        index = (index - 1) & ktp->kt_index_mask;
        if (!ktp->kt_rollover && index == ktp->kt_nentries - 1)
                ktp->kt_rollover = 1;

        ASSERT((index >= 0) && (index < ktp->kt_nentries));

        ktep = &(ktp->kt_entries[index]);

        ktep->val[0]  = val0;
        ktep->val[1]  = val1;
        ktep->val[2]  = val2;
        ktep->val[3]  = val3;
        ktep->val[4]  = val4;
        ktep->val[5]  = val5;
        ktep->val[6]  = val6;
        ktep->val[7]  = val7;
        ktep->val[8]  = val8;
        ktep->val[9]  = val9;
        ktep->val[10] = val10;
        ktep->val[11] = val11;
        ktep->val[12] = val12;
        ktep->val[13] = val13;
        ktep->val[14] = val14;
        ktep->val[15] = val15;
}

/*
 * Return the number of entries in the trace buffer.
 */
int
ktrace_nentries(
        ktrace_t        *ktp)
{
        int     index;
        if (ktp == NULL)
                return 0;

        index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask;
        return (ktp->kt_rollover ? ktp->kt_nentries : index);
}

/*
 * ktrace_first()
 *
 * This is used to find the start of the trace buffer.
 * In conjunction with ktrace_next() it can be used to
 * iterate through the entire trace buffer.  This code does
 * not do any locking because it is assumed that it is called
 * from the debugger.
 *
 * The caller must pass in a pointer to a ktrace_snap
 * structure in which we will keep some state used to
 * iterate through the buffer.  This state must not touched
 * by any code outside of this module.
 */
ktrace_entry_t *
ktrace_first(ktrace_t   *ktp, ktrace_snap_t     *ktsp)
{
        ktrace_entry_t  *ktep;
        int             index;
        int             nentries;

        if (ktp->kt_rollover)
                index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask;
        else
                index = 0;

        ktsp->ks_start = index;
        ktep = &(ktp->kt_entries[index]);

        nentries = ktrace_nentries(ktp);
        index++;
        if (index < nentries) {
                ktsp->ks_index = index;
        } else {
                ktsp->ks_index = 0;
                if (index > nentries)
                        ktep = NULL;
        }
        return ktep;
}

/*
 * ktrace_next()
 *
 * This is used to iterate through the entries of the given
 * trace buffer.  The caller must pass in the ktrace_snap_t
 * structure initialized by ktrace_first().  The return value
 * will be either a pointer to the next ktrace_entry or NULL
 * if all of the entries have been traversed.
 */
ktrace_entry_t *
ktrace_next(
        ktrace_t        *ktp,
        ktrace_snap_t   *ktsp)
{
        int             index;
        ktrace_entry_t  *ktep;

        index = ktsp->ks_index;
        if (index == ktsp->ks_start) {
                ktep = NULL;
        } else {
                ktep = &ktp->kt_entries[index];
        }

        index++;
        if (index == ktrace_nentries(ktp)) {
                ktsp->ks_index = 0;
        } else {
                ktsp->ks_index = index;
        }

        return ktep;
}

/*
 * ktrace_skip()
 *
 * Skip the next "count" entries and return the entry after that.
 * Return NULL if this causes us to iterate past the beginning again.
 */
ktrace_entry_t *
ktrace_skip(
        ktrace_t        *ktp,
        int             count,
        ktrace_snap_t   *ktsp)
{
        int             index;
        int             new_index;
        ktrace_entry_t  *ktep;
        int             nentries = ktrace_nentries(ktp);

        index = ktsp->ks_index;
        new_index = index + count;
        while (new_index >= nentries) {
                new_index -= nentries;
        }
        if (index == ktsp->ks_start) {
                /*
                 * We've iterated around to the start, so we're done.
                 */
                ktep = NULL;
        } else if ((new_index < index) && (index < ktsp->ks_index)) {
                /*
                 * We've skipped past the start again, so we're done.
                 */
                ktep = NULL;
                ktsp->ks_index = ktsp->ks_start;
        } else {
                ktep = &(ktp->kt_entries[new_index]);
                new_index++;
                if (new_index == nentries) {
                        ktsp->ks_index = 0;
                } else {
                        ktsp->ks_index = new_index;
                }
        }
        return ktep;
}

/* [<][>][^][v][top][bottom][index][help] */

[funini.com] -> [kei@sodan] -> Kernel Reading