mirror of
https://github.com/Fledge68/WiiFlow_Lite.git
synced 2024-11-28 06:04:16 +01:00
624 lines
15 KiB
C
624 lines
15 KiB
C
/*
|
|
* block.c --- iterate over all blocks in an inode
|
|
*
|
|
* Copyright (C) 1993, 1994, 1995, 1996 Theodore Ts'o.
|
|
*
|
|
* %Begin-Header%
|
|
* This file may be redistributed under the terms of the GNU Library
|
|
* General Public License, version 2.
|
|
* %End-Header%
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#if HAVE_UNISTD_H
|
|
#include <unistd.h>
|
|
#endif
|
|
|
|
#include "ext2_fs.h"
|
|
#include "ext2fs.h"
|
|
|
|
struct block_context {
|
|
ext2_filsys fs;
|
|
int (*func)(ext2_filsys fs,
|
|
blk64_t *blocknr,
|
|
e2_blkcnt_t bcount,
|
|
blk64_t ref_blk,
|
|
int ref_offset,
|
|
void *priv_data);
|
|
e2_blkcnt_t bcount;
|
|
int bsize;
|
|
int flags;
|
|
errcode_t errcode;
|
|
char *ind_buf;
|
|
char *dind_buf;
|
|
char *tind_buf;
|
|
void *priv_data;
|
|
};
|
|
|
|
#define check_for_ro_violation_return(ctx, ret) \
|
|
do { \
|
|
if (((ctx)->flags & BLOCK_FLAG_READ_ONLY) && \
|
|
((ret) & BLOCK_CHANGED)) { \
|
|
(ctx)->errcode = EXT2_ET_RO_BLOCK_ITERATE; \
|
|
ret |= BLOCK_ABORT | BLOCK_ERROR; \
|
|
return ret; \
|
|
} \
|
|
} while (0)
|
|
|
|
#define check_for_ro_violation_goto(ctx, ret, label) \
|
|
do { \
|
|
if (((ctx)->flags & BLOCK_FLAG_READ_ONLY) && \
|
|
((ret) & BLOCK_CHANGED)) { \
|
|
(ctx)->errcode = EXT2_ET_RO_BLOCK_ITERATE; \
|
|
ret |= BLOCK_ABORT | BLOCK_ERROR; \
|
|
goto label; \
|
|
} \
|
|
} while (0)
|
|
|
|
static int block_iterate_ind(blk_t *ind_block, blk_t ref_block,
|
|
int ref_offset, struct block_context *ctx)
|
|
{
|
|
int ret = 0, changed = 0;
|
|
int i, flags, limit, offset;
|
|
blk_t *block_nr;
|
|
blk64_t blk64;
|
|
|
|
limit = ctx->fs->blocksize >> 2;
|
|
if (!(ctx->flags & BLOCK_FLAG_DEPTH_TRAVERSE) &&
|
|
!(ctx->flags & BLOCK_FLAG_DATA_ONLY)) {
|
|
blk64 = *ind_block;
|
|
ret = (*ctx->func)(ctx->fs, &blk64,
|
|
BLOCK_COUNT_IND, ref_block,
|
|
ref_offset, ctx->priv_data);
|
|
*ind_block = blk64;
|
|
}
|
|
check_for_ro_violation_return(ctx, ret);
|
|
if (!*ind_block || (ret & BLOCK_ABORT)) {
|
|
ctx->bcount += limit;
|
|
return ret;
|
|
}
|
|
if (*ind_block >= ext2fs_blocks_count(ctx->fs->super) ||
|
|
*ind_block < ctx->fs->super->s_first_data_block) {
|
|
ctx->errcode = EXT2_ET_BAD_IND_BLOCK;
|
|
ret |= BLOCK_ERROR;
|
|
return ret;
|
|
}
|
|
ctx->errcode = ext2fs_read_ind_block(ctx->fs, *ind_block,
|
|
ctx->ind_buf);
|
|
if (ctx->errcode) {
|
|
ret |= BLOCK_ERROR;
|
|
return ret;
|
|
}
|
|
|
|
block_nr = (blk_t *) ctx->ind_buf;
|
|
offset = 0;
|
|
if (ctx->flags & BLOCK_FLAG_APPEND) {
|
|
for (i = 0; i < limit; i++, ctx->bcount++, block_nr++) {
|
|
blk64 = *block_nr;
|
|
flags = (*ctx->func)(ctx->fs, &blk64, ctx->bcount,
|
|
*ind_block, offset,
|
|
ctx->priv_data);
|
|
*block_nr = blk64;
|
|
changed |= flags;
|
|
if (flags & BLOCK_ABORT) {
|
|
ret |= BLOCK_ABORT;
|
|
break;
|
|
}
|
|
offset += sizeof(blk_t);
|
|
}
|
|
} else {
|
|
for (i = 0; i < limit; i++, ctx->bcount++, block_nr++) {
|
|
if (*block_nr == 0)
|
|
goto skip_sparse;
|
|
blk64 = *block_nr;
|
|
flags = (*ctx->func)(ctx->fs, &blk64, ctx->bcount,
|
|
*ind_block, offset,
|
|
ctx->priv_data);
|
|
*block_nr = blk64;
|
|
changed |= flags;
|
|
if (flags & BLOCK_ABORT) {
|
|
ret |= BLOCK_ABORT;
|
|
break;
|
|
}
|
|
skip_sparse:
|
|
offset += sizeof(blk_t);
|
|
}
|
|
}
|
|
check_for_ro_violation_return(ctx, changed);
|
|
if (changed & BLOCK_CHANGED) {
|
|
ctx->errcode = ext2fs_write_ind_block(ctx->fs, *ind_block,
|
|
ctx->ind_buf);
|
|
if (ctx->errcode)
|
|
ret |= BLOCK_ERROR | BLOCK_ABORT;
|
|
}
|
|
if ((ctx->flags & BLOCK_FLAG_DEPTH_TRAVERSE) &&
|
|
!(ctx->flags & BLOCK_FLAG_DATA_ONLY) &&
|
|
!(ret & BLOCK_ABORT)) {
|
|
blk64 = *ind_block;
|
|
ret |= (*ctx->func)(ctx->fs, &blk64,
|
|
BLOCK_COUNT_IND, ref_block,
|
|
ref_offset, ctx->priv_data);
|
|
*ind_block = blk64;
|
|
}
|
|
check_for_ro_violation_return(ctx, ret);
|
|
return ret;
|
|
}
|
|
|
|
static int block_iterate_dind(blk_t *dind_block, blk_t ref_block,
|
|
int ref_offset, struct block_context *ctx)
|
|
{
|
|
int ret = 0, changed = 0;
|
|
int i, flags, limit, offset;
|
|
blk_t *block_nr;
|
|
blk64_t blk64;
|
|
|
|
limit = ctx->fs->blocksize >> 2;
|
|
if (!(ctx->flags & (BLOCK_FLAG_DEPTH_TRAVERSE |
|
|
BLOCK_FLAG_DATA_ONLY))) {
|
|
blk64 = *dind_block;
|
|
ret = (*ctx->func)(ctx->fs, &blk64,
|
|
BLOCK_COUNT_DIND, ref_block,
|
|
ref_offset, ctx->priv_data);
|
|
*dind_block = blk64;
|
|
}
|
|
check_for_ro_violation_return(ctx, ret);
|
|
if (!*dind_block || (ret & BLOCK_ABORT)) {
|
|
ctx->bcount += limit*limit;
|
|
return ret;
|
|
}
|
|
if (*dind_block >= ext2fs_blocks_count(ctx->fs->super) ||
|
|
*dind_block < ctx->fs->super->s_first_data_block) {
|
|
ctx->errcode = EXT2_ET_BAD_DIND_BLOCK;
|
|
ret |= BLOCK_ERROR;
|
|
return ret;
|
|
}
|
|
ctx->errcode = ext2fs_read_ind_block(ctx->fs, *dind_block,
|
|
ctx->dind_buf);
|
|
if (ctx->errcode) {
|
|
ret |= BLOCK_ERROR;
|
|
return ret;
|
|
}
|
|
|
|
block_nr = (blk_t *) ctx->dind_buf;
|
|
offset = 0;
|
|
if (ctx->flags & BLOCK_FLAG_APPEND) {
|
|
for (i = 0; i < limit; i++, block_nr++) {
|
|
flags = block_iterate_ind(block_nr,
|
|
*dind_block, offset,
|
|
ctx);
|
|
changed |= flags;
|
|
if (flags & (BLOCK_ABORT | BLOCK_ERROR)) {
|
|
ret |= flags & (BLOCK_ABORT | BLOCK_ERROR);
|
|
break;
|
|
}
|
|
offset += sizeof(blk_t);
|
|
}
|
|
} else {
|
|
for (i = 0; i < limit; i++, block_nr++) {
|
|
if (*block_nr == 0) {
|
|
ctx->bcount += limit;
|
|
continue;
|
|
}
|
|
flags = block_iterate_ind(block_nr,
|
|
*dind_block, offset,
|
|
ctx);
|
|
changed |= flags;
|
|
if (flags & (BLOCK_ABORT | BLOCK_ERROR)) {
|
|
ret |= flags & (BLOCK_ABORT | BLOCK_ERROR);
|
|
break;
|
|
}
|
|
offset += sizeof(blk_t);
|
|
}
|
|
}
|
|
check_for_ro_violation_return(ctx, changed);
|
|
if (changed & BLOCK_CHANGED) {
|
|
ctx->errcode = ext2fs_write_ind_block(ctx->fs, *dind_block,
|
|
ctx->dind_buf);
|
|
if (ctx->errcode)
|
|
ret |= BLOCK_ERROR | BLOCK_ABORT;
|
|
}
|
|
if ((ctx->flags & BLOCK_FLAG_DEPTH_TRAVERSE) &&
|
|
!(ctx->flags & BLOCK_FLAG_DATA_ONLY) &&
|
|
!(ret & BLOCK_ABORT)) {
|
|
blk64 = *dind_block;
|
|
ret |= (*ctx->func)(ctx->fs, &blk64,
|
|
BLOCK_COUNT_DIND, ref_block,
|
|
ref_offset, ctx->priv_data);
|
|
*dind_block = blk64;
|
|
}
|
|
check_for_ro_violation_return(ctx, ret);
|
|
return ret;
|
|
}
|
|
|
|
static int block_iterate_tind(blk_t *tind_block, blk_t ref_block,
|
|
int ref_offset, struct block_context *ctx)
|
|
{
|
|
int ret = 0, changed = 0;
|
|
int i, flags, limit, offset;
|
|
blk_t *block_nr;
|
|
blk64_t blk64;
|
|
|
|
limit = ctx->fs->blocksize >> 2;
|
|
if (!(ctx->flags & (BLOCK_FLAG_DEPTH_TRAVERSE |
|
|
BLOCK_FLAG_DATA_ONLY))) {
|
|
blk64 = *tind_block;
|
|
ret = (*ctx->func)(ctx->fs, &blk64,
|
|
BLOCK_COUNT_TIND, ref_block,
|
|
ref_offset, ctx->priv_data);
|
|
*tind_block = blk64;
|
|
}
|
|
check_for_ro_violation_return(ctx, ret);
|
|
if (!*tind_block || (ret & BLOCK_ABORT)) {
|
|
ctx->bcount += limit*limit*limit;
|
|
return ret;
|
|
}
|
|
if (*tind_block >= ext2fs_blocks_count(ctx->fs->super) ||
|
|
*tind_block < ctx->fs->super->s_first_data_block) {
|
|
ctx->errcode = EXT2_ET_BAD_TIND_BLOCK;
|
|
ret |= BLOCK_ERROR;
|
|
return ret;
|
|
}
|
|
ctx->errcode = ext2fs_read_ind_block(ctx->fs, *tind_block,
|
|
ctx->tind_buf);
|
|
if (ctx->errcode) {
|
|
ret |= BLOCK_ERROR;
|
|
return ret;
|
|
}
|
|
|
|
block_nr = (blk_t *) ctx->tind_buf;
|
|
offset = 0;
|
|
if (ctx->flags & BLOCK_FLAG_APPEND) {
|
|
for (i = 0; i < limit; i++, block_nr++) {
|
|
flags = block_iterate_dind(block_nr,
|
|
*tind_block,
|
|
offset, ctx);
|
|
changed |= flags;
|
|
if (flags & (BLOCK_ABORT | BLOCK_ERROR)) {
|
|
ret |= flags & (BLOCK_ABORT | BLOCK_ERROR);
|
|
break;
|
|
}
|
|
offset += sizeof(blk_t);
|
|
}
|
|
} else {
|
|
for (i = 0; i < limit; i++, block_nr++) {
|
|
if (*block_nr == 0) {
|
|
ctx->bcount += limit*limit;
|
|
continue;
|
|
}
|
|
flags = block_iterate_dind(block_nr,
|
|
*tind_block,
|
|
offset, ctx);
|
|
changed |= flags;
|
|
if (flags & (BLOCK_ABORT | BLOCK_ERROR)) {
|
|
ret |= flags & (BLOCK_ABORT | BLOCK_ERROR);
|
|
break;
|
|
}
|
|
offset += sizeof(blk_t);
|
|
}
|
|
}
|
|
check_for_ro_violation_return(ctx, changed);
|
|
if (changed & BLOCK_CHANGED) {
|
|
ctx->errcode = ext2fs_write_ind_block(ctx->fs, *tind_block,
|
|
ctx->tind_buf);
|
|
if (ctx->errcode)
|
|
ret |= BLOCK_ERROR | BLOCK_ABORT;
|
|
}
|
|
if ((ctx->flags & BLOCK_FLAG_DEPTH_TRAVERSE) &&
|
|
!(ctx->flags & BLOCK_FLAG_DATA_ONLY) &&
|
|
!(ret & BLOCK_ABORT)) {
|
|
blk64 = *tind_block;
|
|
ret |= (*ctx->func)(ctx->fs, &blk64,
|
|
BLOCK_COUNT_TIND, ref_block,
|
|
ref_offset, ctx->priv_data);
|
|
*tind_block = blk64;
|
|
}
|
|
check_for_ro_violation_return(ctx, ret);
|
|
return ret;
|
|
}
|
|
|
|
errcode_t ext2fs_block_iterate3(ext2_filsys fs,
|
|
ext2_ino_t ino,
|
|
int flags,
|
|
char *block_buf,
|
|
int (*func)(ext2_filsys fs,
|
|
blk64_t *blocknr,
|
|
e2_blkcnt_t blockcnt,
|
|
blk64_t ref_blk,
|
|
int ref_offset,
|
|
void *priv_data),
|
|
void *priv_data)
|
|
{
|
|
int i;
|
|
int r, ret = 0;
|
|
struct ext2_inode inode;
|
|
errcode_t retval;
|
|
struct block_context ctx;
|
|
int limit;
|
|
blk64_t blk64;
|
|
|
|
EXT2_CHECK_MAGIC(fs, EXT2_ET_MAGIC_EXT2FS_FILSYS);
|
|
|
|
ctx.errcode = ext2fs_read_inode(fs, ino, &inode);
|
|
if (ctx.errcode)
|
|
return ctx.errcode;
|
|
|
|
/*
|
|
* Check to see if we need to limit large files
|
|
*/
|
|
if (flags & BLOCK_FLAG_NO_LARGE) {
|
|
if (!LINUX_S_ISDIR(inode.i_mode) &&
|
|
(inode.i_size_high != 0))
|
|
return EXT2_ET_FILE_TOO_BIG;
|
|
}
|
|
|
|
limit = fs->blocksize >> 2;
|
|
|
|
ctx.fs = fs;
|
|
ctx.func = func;
|
|
ctx.priv_data = priv_data;
|
|
ctx.flags = flags;
|
|
ctx.bcount = 0;
|
|
if (block_buf) {
|
|
ctx.ind_buf = block_buf;
|
|
} else {
|
|
retval = ext2fs_get_array(3, fs->blocksize, &ctx.ind_buf);
|
|
if (retval)
|
|
return retval;
|
|
}
|
|
ctx.dind_buf = ctx.ind_buf + fs->blocksize;
|
|
ctx.tind_buf = ctx.dind_buf + fs->blocksize;
|
|
|
|
/*
|
|
* Iterate over the HURD translator block (if present)
|
|
*/
|
|
if ((fs->super->s_creator_os == EXT2_OS_HURD) &&
|
|
!(flags & BLOCK_FLAG_DATA_ONLY)) {
|
|
if (inode.osd1.hurd1.h_i_translator) {
|
|
blk64 = inode.osd1.hurd1.h_i_translator;
|
|
ret |= (*ctx.func)(fs, &blk64,
|
|
BLOCK_COUNT_TRANSLATOR,
|
|
0, 0, priv_data);
|
|
inode.osd1.hurd1.h_i_translator = (blk_t) blk64;
|
|
if (ret & BLOCK_ABORT)
|
|
goto abort_exit;
|
|
check_for_ro_violation_goto(&ctx, ret, abort_exit);
|
|
}
|
|
}
|
|
|
|
if (inode.i_flags & EXT4_EXTENTS_FL) {
|
|
ext2_extent_handle_t handle;
|
|
struct ext2fs_extent extent;
|
|
e2_blkcnt_t blockcnt = 0;
|
|
blk64_t blk, new_blk;
|
|
int op = EXT2_EXTENT_ROOT;
|
|
int uninit;
|
|
unsigned int j;
|
|
|
|
ctx.errcode = ext2fs_extent_open2(fs, ino, &inode, &handle);
|
|
if (ctx.errcode)
|
|
goto abort_exit;
|
|
|
|
while (1) {
|
|
ctx.errcode = ext2fs_extent_get(handle, op, &extent);
|
|
if (ctx.errcode) {
|
|
if (ctx.errcode != EXT2_ET_EXTENT_NO_NEXT)
|
|
break;
|
|
ctx.errcode = 0;
|
|
if (!(flags & BLOCK_FLAG_APPEND))
|
|
break;
|
|
next_block_set:
|
|
blk = 0;
|
|
r = (*ctx.func)(fs, &blk, blockcnt,
|
|
0, 0, priv_data);
|
|
ret |= r;
|
|
check_for_ro_violation_goto(&ctx, ret,
|
|
extent_errout);
|
|
if (r & BLOCK_CHANGED) {
|
|
ctx.errcode =
|
|
ext2fs_extent_set_bmap(handle,
|
|
(blk64_t) blockcnt++,
|
|
(blk64_t) blk, 0);
|
|
if (ctx.errcode || (ret & BLOCK_ABORT))
|
|
break;
|
|
if (blk)
|
|
goto next_block_set;
|
|
}
|
|
break;
|
|
}
|
|
|
|
op = EXT2_EXTENT_NEXT;
|
|
blk = extent.e_pblk;
|
|
if (!(extent.e_flags & EXT2_EXTENT_FLAGS_LEAF)) {
|
|
if (ctx.flags & BLOCK_FLAG_DATA_ONLY)
|
|
continue;
|
|
if ((!(extent.e_flags &
|
|
EXT2_EXTENT_FLAGS_SECOND_VISIT) &&
|
|
!(ctx.flags & BLOCK_FLAG_DEPTH_TRAVERSE)) ||
|
|
((extent.e_flags &
|
|
EXT2_EXTENT_FLAGS_SECOND_VISIT) &&
|
|
(ctx.flags & BLOCK_FLAG_DEPTH_TRAVERSE))) {
|
|
ret |= (*ctx.func)(fs, &blk,
|
|
-1, 0, 0, priv_data);
|
|
if (ret & BLOCK_CHANGED) {
|
|
extent.e_pblk = blk;
|
|
ctx.errcode =
|
|
ext2fs_extent_replace(handle, 0, &extent);
|
|
if (ctx.errcode)
|
|
break;
|
|
}
|
|
}
|
|
continue;
|
|
}
|
|
uninit = 0;
|
|
if (extent.e_flags & EXT2_EXTENT_FLAGS_UNINIT)
|
|
uninit = EXT2_EXTENT_SET_BMAP_UNINIT;
|
|
for (blockcnt = extent.e_lblk, j = 0;
|
|
j < extent.e_len;
|
|
blk++, blockcnt++, j++) {
|
|
new_blk = blk;
|
|
r = (*ctx.func)(fs, &new_blk, blockcnt,
|
|
0, 0, priv_data);
|
|
ret |= r;
|
|
check_for_ro_violation_goto(&ctx, ret,
|
|
extent_errout);
|
|
if (r & BLOCK_CHANGED) {
|
|
ctx.errcode =
|
|
ext2fs_extent_set_bmap(handle,
|
|
(blk64_t) blockcnt,
|
|
new_blk, uninit);
|
|
if (ctx.errcode)
|
|
goto extent_errout;
|
|
}
|
|
if (ret & BLOCK_ABORT)
|
|
break;
|
|
}
|
|
}
|
|
|
|
extent_errout:
|
|
ext2fs_extent_free(handle);
|
|
ret |= BLOCK_ERROR | BLOCK_ABORT;
|
|
goto errout;
|
|
}
|
|
|
|
/*
|
|
* Iterate over normal data blocks
|
|
*/
|
|
for (i = 0; i < EXT2_NDIR_BLOCKS ; i++, ctx.bcount++) {
|
|
if (inode.i_block[i] || (flags & BLOCK_FLAG_APPEND)) {
|
|
blk64 = inode.i_block[i];
|
|
ret |= (*ctx.func)(fs, &blk64, ctx.bcount, 0, i,
|
|
priv_data);
|
|
inode.i_block[i] = (blk_t) blk64;
|
|
if (ret & BLOCK_ABORT)
|
|
goto abort_exit;
|
|
}
|
|
}
|
|
check_for_ro_violation_goto(&ctx, ret, abort_exit);
|
|
if (inode.i_block[EXT2_IND_BLOCK] || (flags & BLOCK_FLAG_APPEND)) {
|
|
ret |= block_iterate_ind(&inode.i_block[EXT2_IND_BLOCK],
|
|
0, EXT2_IND_BLOCK, &ctx);
|
|
if (ret & BLOCK_ABORT)
|
|
goto abort_exit;
|
|
} else
|
|
ctx.bcount += limit;
|
|
if (inode.i_block[EXT2_DIND_BLOCK] || (flags & BLOCK_FLAG_APPEND)) {
|
|
ret |= block_iterate_dind(&inode.i_block[EXT2_DIND_BLOCK],
|
|
0, EXT2_DIND_BLOCK, &ctx);
|
|
if (ret & BLOCK_ABORT)
|
|
goto abort_exit;
|
|
} else
|
|
ctx.bcount += limit * limit;
|
|
if (inode.i_block[EXT2_TIND_BLOCK] || (flags & BLOCK_FLAG_APPEND)) {
|
|
ret |= block_iterate_tind(&inode.i_block[EXT2_TIND_BLOCK],
|
|
0, EXT2_TIND_BLOCK, &ctx);
|
|
if (ret & BLOCK_ABORT)
|
|
goto abort_exit;
|
|
}
|
|
|
|
abort_exit:
|
|
if (ret & BLOCK_CHANGED) {
|
|
retval = ext2fs_write_inode(fs, ino, &inode);
|
|
if (retval) {
|
|
ret |= BLOCK_ERROR;
|
|
ctx.errcode = retval;
|
|
}
|
|
}
|
|
errout:
|
|
if (!block_buf)
|
|
ext2fs_free_mem(&ctx.ind_buf);
|
|
|
|
return (ret & BLOCK_ERROR) ? ctx.errcode : 0;
|
|
}
|
|
|
|
/*
|
|
* Emulate the old ext2fs_block_iterate function!
|
|
*/
|
|
|
|
struct xlate64 {
|
|
int (*func)(ext2_filsys fs,
|
|
blk_t *blocknr,
|
|
e2_blkcnt_t blockcnt,
|
|
blk_t ref_blk,
|
|
int ref_offset,
|
|
void *priv_data);
|
|
void *real_private;
|
|
};
|
|
|
|
static int xlate64_func(ext2_filsys fs, blk64_t *blocknr,
|
|
e2_blkcnt_t blockcnt, blk64_t ref_blk,
|
|
int ref_offset, void *priv_data)
|
|
{
|
|
struct xlate64 *xl = (struct xlate64 *) priv_data;
|
|
int ret;
|
|
blk_t block32 = *blocknr;
|
|
|
|
ret = (*xl->func)(fs, &block32, blockcnt, (blk_t) ref_blk, ref_offset,
|
|
xl->real_private);
|
|
*blocknr = block32;
|
|
return ret;
|
|
}
|
|
|
|
errcode_t ext2fs_block_iterate2(ext2_filsys fs,
|
|
ext2_ino_t ino,
|
|
int flags,
|
|
char *block_buf,
|
|
int (*func)(ext2_filsys fs,
|
|
blk_t *blocknr,
|
|
e2_blkcnt_t blockcnt,
|
|
blk_t ref_blk,
|
|
int ref_offset,
|
|
void *priv_data),
|
|
void *priv_data)
|
|
{
|
|
struct xlate64 xl;
|
|
|
|
xl.real_private = priv_data;
|
|
xl.func = func;
|
|
|
|
return ext2fs_block_iterate3(fs, ino, flags, block_buf,
|
|
xlate64_func, &xl);
|
|
}
|
|
|
|
|
|
struct xlate {
|
|
int (*func)(ext2_filsys fs,
|
|
blk_t *blocknr,
|
|
int bcount,
|
|
void *priv_data);
|
|
void *real_private;
|
|
};
|
|
|
|
#ifdef __TURBOC__
|
|
#pragma argsused
|
|
#endif
|
|
static int xlate_func(ext2_filsys fs, blk_t *blocknr, e2_blkcnt_t blockcnt,
|
|
blk_t ref_block EXT2FS_ATTR((unused)),
|
|
int ref_offset EXT2FS_ATTR((unused)),
|
|
void *priv_data)
|
|
{
|
|
struct xlate *xl = (struct xlate *) priv_data;
|
|
|
|
return (*xl->func)(fs, blocknr, (int) blockcnt, xl->real_private);
|
|
}
|
|
|
|
errcode_t ext2fs_block_iterate(ext2_filsys fs,
|
|
ext2_ino_t ino,
|
|
int flags,
|
|
char *block_buf,
|
|
int (*func)(ext2_filsys fs,
|
|
blk_t *blocknr,
|
|
int blockcnt,
|
|
void *priv_data),
|
|
void *priv_data)
|
|
{
|
|
struct xlate xl;
|
|
|
|
xl.real_private = priv_data;
|
|
xl.func = func;
|
|
|
|
return ext2fs_block_iterate2(fs, ino, BLOCK_FLAG_NO_LARGE | flags,
|
|
block_buf, xlate_func, &xl);
|
|
}
|
|
|