diff options
Diffstat (limited to 'fs/nfs/objlayout/objlayout.c')
| -rw-r--r-- | fs/nfs/objlayout/objlayout.c | 779 | 
1 files changed, 779 insertions, 0 deletions
diff --git a/fs/nfs/objlayout/objlayout.c b/fs/nfs/objlayout/objlayout.c new file mode 100644 index 00000000000..765d3f54e98 --- /dev/null +++ b/fs/nfs/objlayout/objlayout.c @@ -0,0 +1,779 @@ +/* + *  pNFS Objects layout driver high level definitions + * + *  Copyright (C) 2007 Panasas Inc. [year of first publication] + *  All rights reserved. + * + *  Benny Halevy <bhalevy@panasas.com> + *  Boaz Harrosh <bharrosh@panasas.com> + * + *  This program is free software; you can redistribute it and/or modify + *  it under the terms of the GNU General Public License version 2 + *  See the file COPYING included with this distribution for more details. + * + *  Redistribution and use in source and binary forms, with or without + *  modification, are permitted provided that the following conditions + *  are met: + * + *  1. Redistributions of source code must retain the above copyright + *     notice, this list of conditions and the following disclaimer. + *  2. Redistributions in binary form must reproduce the above copyright + *     notice, this list of conditions and the following disclaimer in the + *     documentation and/or other materials provided with the distribution. + *  3. Neither the name of the Panasas company nor the names of its + *     contributors may be used to endorse or promote products derived + *     from this software without specific prior written permission. + * + *  THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED + *  WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF + *  MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + *  DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE + *  FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR + *  BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF + *  LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING + *  NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS + *  SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include <linux/kmod.h> +#include <linux/moduleparam.h> +#include <linux/ratelimit.h> +#include <scsi/osd_initiator.h> +#include "objlayout.h" + +#define NFSDBG_FACILITY         NFSDBG_PNFS_LD +/* + * Create a objlayout layout structure for the given inode and return it. + */ +struct pnfs_layout_hdr * +objlayout_alloc_layout_hdr(struct inode *inode, gfp_t gfp_flags) +{ +	struct objlayout *objlay; + +	objlay = kzalloc(sizeof(struct objlayout), gfp_flags); +	if (!objlay) +		return NULL; +	spin_lock_init(&objlay->lock); +	INIT_LIST_HEAD(&objlay->err_list); +	dprintk("%s: Return %p\n", __func__, objlay); +	return &objlay->pnfs_layout; +} + +/* + * Free an objlayout layout structure + */ +void +objlayout_free_layout_hdr(struct pnfs_layout_hdr *lo) +{ +	struct objlayout *objlay = OBJLAYOUT(lo); + +	dprintk("%s: objlay %p\n", __func__, objlay); + +	WARN_ON(!list_empty(&objlay->err_list)); +	kfree(objlay); +} + +/* + * Unmarshall layout and store it in pnfslay. + */ +struct pnfs_layout_segment * +objlayout_alloc_lseg(struct pnfs_layout_hdr *pnfslay, +		     struct nfs4_layoutget_res *lgr, +		     gfp_t gfp_flags) +{ +	int status = -ENOMEM; +	struct xdr_stream stream; +	struct xdr_buf buf = { +		.pages =  lgr->layoutp->pages, +		.page_len =  lgr->layoutp->len, +		.buflen =  lgr->layoutp->len, +		.len = lgr->layoutp->len, +	}; +	struct page *scratch; +	struct pnfs_layout_segment *lseg; + +	dprintk("%s: Begin pnfslay %p\n", __func__, pnfslay); + +	scratch = alloc_page(gfp_flags); +	if (!scratch) +		goto err_nofree; + +	xdr_init_decode(&stream, &buf, NULL); +	xdr_set_scratch_buffer(&stream, page_address(scratch), PAGE_SIZE); + +	status = objio_alloc_lseg(&lseg, pnfslay, &lgr->range, &stream, gfp_flags); +	if (unlikely(status)) { +		dprintk("%s: objio_alloc_lseg Return err %d\n", __func__, +			status); +		goto err; +	} + +	__free_page(scratch); + +	dprintk("%s: Return %p\n", __func__, lseg); +	return lseg; + +err: +	__free_page(scratch); +err_nofree: +	dprintk("%s: Err Return=>%d\n", __func__, status); +	return ERR_PTR(status); +} + +/* + * Free a layout segement + */ +void +objlayout_free_lseg(struct pnfs_layout_segment *lseg) +{ +	dprintk("%s: freeing layout segment %p\n", __func__, lseg); + +	if (unlikely(!lseg)) +		return; + +	objio_free_lseg(lseg); +} + +/* + * I/O Operations + */ +static inline u64 +end_offset(u64 start, u64 len) +{ +	u64 end; + +	end = start + len; +	return end >= start ? end : NFS4_MAX_UINT64; +} + +static void _fix_verify_io_params(struct pnfs_layout_segment *lseg, +			   struct page ***p_pages, unsigned *p_pgbase, +			   u64 offset, unsigned long count) +{ +	u64 lseg_end_offset; + +	BUG_ON(offset < lseg->pls_range.offset); +	lseg_end_offset = end_offset(lseg->pls_range.offset, +				     lseg->pls_range.length); +	BUG_ON(offset >= lseg_end_offset); +	WARN_ON(offset + count > lseg_end_offset); + +	if (*p_pgbase > PAGE_SIZE) { +		dprintk("%s: pgbase(0x%x) > PAGE_SIZE\n", __func__, *p_pgbase); +		*p_pages += *p_pgbase >> PAGE_SHIFT; +		*p_pgbase &= ~PAGE_MASK; +	} +} + +/* + * I/O done common code + */ +static void +objlayout_iodone(struct objlayout_io_res *oir) +{ +	if (likely(oir->status >= 0)) { +		objio_free_result(oir); +	} else { +		struct objlayout *objlay = oir->objlay; + +		spin_lock(&objlay->lock); +		objlay->delta_space_valid = OBJ_DSU_INVALID; +		list_add(&objlay->err_list, &oir->err_list); +		spin_unlock(&objlay->lock); +	} +} + +/* + * objlayout_io_set_result - Set an osd_error code on a specific osd comp. + * + * The @index component IO failed (error returned from target). Register + * the error for later reporting at layout-return. + */ +void +objlayout_io_set_result(struct objlayout_io_res *oir, unsigned index, +			struct pnfs_osd_objid *pooid, int osd_error, +			u64 offset, u64 length, bool is_write) +{ +	struct pnfs_osd_ioerr *ioerr = &oir->ioerrs[index]; + +	BUG_ON(index >= oir->num_comps); +	if (osd_error) { +		ioerr->oer_component = *pooid; +		ioerr->oer_comp_offset = offset; +		ioerr->oer_comp_length = length; +		ioerr->oer_iswrite = is_write; +		ioerr->oer_errno = osd_error; + +		dprintk("%s: err[%d]: errno=%d is_write=%d dev(%llx:%llx) " +			"par=0x%llx obj=0x%llx offset=0x%llx length=0x%llx\n", +			__func__, index, ioerr->oer_errno, +			ioerr->oer_iswrite, +			_DEVID_LO(&ioerr->oer_component.oid_device_id), +			_DEVID_HI(&ioerr->oer_component.oid_device_id), +			ioerr->oer_component.oid_partition_id, +			ioerr->oer_component.oid_object_id, +			ioerr->oer_comp_offset, +			ioerr->oer_comp_length); +	} else { +		/* User need not call if no error is reported */ +		ioerr->oer_errno = 0; +	} +} + +/* Function scheduled on rpc workqueue to call ->nfs_readlist_complete(). + * This is because the osd completion is called with ints-off from + * the block layer + */ +static void _rpc_read_complete(struct work_struct *work) +{ +	struct rpc_task *task; +	struct nfs_pgio_data *rdata; + +	dprintk("%s enter\n", __func__); +	task = container_of(work, struct rpc_task, u.tk_work); +	rdata = container_of(task, struct nfs_pgio_data, task); + +	pnfs_ld_read_done(rdata); +} + +void +objlayout_read_done(struct objlayout_io_res *oir, ssize_t status, bool sync) +{ +	struct nfs_pgio_data *rdata = oir->rpcdata; + +	oir->status = rdata->task.tk_status = status; +	if (status >= 0) +		rdata->res.count = status; +	else +		rdata->header->pnfs_error = status; +	objlayout_iodone(oir); +	/* must not use oir after this point */ + +	dprintk("%s: Return status=%zd eof=%d sync=%d\n", __func__, +		status, rdata->res.eof, sync); + +	if (sync) +		pnfs_ld_read_done(rdata); +	else { +		INIT_WORK(&rdata->task.u.tk_work, _rpc_read_complete); +		schedule_work(&rdata->task.u.tk_work); +	} +} + +/* + * Perform sync or async reads. + */ +enum pnfs_try_status +objlayout_read_pagelist(struct nfs_pgio_data *rdata) +{ +	struct nfs_pgio_header *hdr = rdata->header; +	struct inode *inode = hdr->inode; +	loff_t offset = rdata->args.offset; +	size_t count = rdata->args.count; +	int err; +	loff_t eof; + +	eof = i_size_read(inode); +	if (unlikely(offset + count > eof)) { +		if (offset >= eof) { +			err = 0; +			rdata->res.count = 0; +			rdata->res.eof = 1; +			/*FIXME: do we need to call pnfs_ld_read_done() */ +			goto out; +		} +		count = eof - offset; +	} + +	rdata->res.eof = (offset + count) >= eof; +	_fix_verify_io_params(hdr->lseg, &rdata->args.pages, +			      &rdata->args.pgbase, +			      rdata->args.offset, rdata->args.count); + +	dprintk("%s: inode(%lx) offset 0x%llx count 0x%Zx eof=%d\n", +		__func__, inode->i_ino, offset, count, rdata->res.eof); + +	err = objio_read_pagelist(rdata); + out: +	if (unlikely(err)) { +		hdr->pnfs_error = err; +		dprintk("%s: Returned Error %d\n", __func__, err); +		return PNFS_NOT_ATTEMPTED; +	} +	return PNFS_ATTEMPTED; +} + +/* Function scheduled on rpc workqueue to call ->nfs_writelist_complete(). + * This is because the osd completion is called with ints-off from + * the block layer + */ +static void _rpc_write_complete(struct work_struct *work) +{ +	struct rpc_task *task; +	struct nfs_pgio_data *wdata; + +	dprintk("%s enter\n", __func__); +	task = container_of(work, struct rpc_task, u.tk_work); +	wdata = container_of(task, struct nfs_pgio_data, task); + +	pnfs_ld_write_done(wdata); +} + +void +objlayout_write_done(struct objlayout_io_res *oir, ssize_t status, bool sync) +{ +	struct nfs_pgio_data *wdata = oir->rpcdata; + +	oir->status = wdata->task.tk_status = status; +	if (status >= 0) { +		wdata->res.count = status; +		wdata->verf.committed = oir->committed; +	} else { +		wdata->header->pnfs_error = status; +	} +	objlayout_iodone(oir); +	/* must not use oir after this point */ + +	dprintk("%s: Return status %zd committed %d sync=%d\n", __func__, +		status, wdata->verf.committed, sync); + +	if (sync) +		pnfs_ld_write_done(wdata); +	else { +		INIT_WORK(&wdata->task.u.tk_work, _rpc_write_complete); +		schedule_work(&wdata->task.u.tk_work); +	} +} + +/* + * Perform sync or async writes. + */ +enum pnfs_try_status +objlayout_write_pagelist(struct nfs_pgio_data *wdata, +			 int how) +{ +	struct nfs_pgio_header *hdr = wdata->header; +	int err; + +	_fix_verify_io_params(hdr->lseg, &wdata->args.pages, +			      &wdata->args.pgbase, +			      wdata->args.offset, wdata->args.count); + +	err = objio_write_pagelist(wdata, how); +	if (unlikely(err)) { +		hdr->pnfs_error = err; +		dprintk("%s: Returned Error %d\n", __func__, err); +		return PNFS_NOT_ATTEMPTED; +	} +	return PNFS_ATTEMPTED; +} + +void +objlayout_encode_layoutcommit(struct pnfs_layout_hdr *pnfslay, +			      struct xdr_stream *xdr, +			      const struct nfs4_layoutcommit_args *args) +{ +	struct objlayout *objlay = OBJLAYOUT(pnfslay); +	struct pnfs_osd_layoutupdate lou; +	__be32 *start; + +	dprintk("%s: Begin\n", __func__); + +	spin_lock(&objlay->lock); +	lou.dsu_valid = (objlay->delta_space_valid == OBJ_DSU_VALID); +	lou.dsu_delta = objlay->delta_space_used; +	objlay->delta_space_used = 0; +	objlay->delta_space_valid = OBJ_DSU_INIT; +	lou.olu_ioerr_flag = !list_empty(&objlay->err_list); +	spin_unlock(&objlay->lock); + +	start = xdr_reserve_space(xdr, 4); + +	BUG_ON(pnfs_osd_xdr_encode_layoutupdate(xdr, &lou)); + +	*start = cpu_to_be32((xdr->p - start - 1) * 4); + +	dprintk("%s: Return delta_space_used %lld err %d\n", __func__, +		lou.dsu_delta, lou.olu_ioerr_flag); +} + +static int +err_prio(u32 oer_errno) +{ +	switch (oer_errno) { +	case 0: +		return 0; + +	case PNFS_OSD_ERR_RESOURCE: +		return OSD_ERR_PRI_RESOURCE; +	case PNFS_OSD_ERR_BAD_CRED: +		return OSD_ERR_PRI_BAD_CRED; +	case PNFS_OSD_ERR_NO_ACCESS: +		return OSD_ERR_PRI_NO_ACCESS; +	case PNFS_OSD_ERR_UNREACHABLE: +		return OSD_ERR_PRI_UNREACHABLE; +	case PNFS_OSD_ERR_NOT_FOUND: +		return OSD_ERR_PRI_NOT_FOUND; +	case PNFS_OSD_ERR_NO_SPACE: +		return OSD_ERR_PRI_NO_SPACE; +	default: +		WARN_ON(1); +		/* fallthrough */ +	case PNFS_OSD_ERR_EIO: +		return OSD_ERR_PRI_EIO; +	} +} + +static void +merge_ioerr(struct pnfs_osd_ioerr *dest_err, +	    const struct pnfs_osd_ioerr *src_err) +{ +	u64 dest_end, src_end; + +	if (!dest_err->oer_errno) { +		*dest_err = *src_err; +		/* accumulated device must be blank */ +		memset(&dest_err->oer_component.oid_device_id, 0, +			sizeof(dest_err->oer_component.oid_device_id)); + +		return; +	} + +	if (dest_err->oer_component.oid_partition_id != +				src_err->oer_component.oid_partition_id) +		dest_err->oer_component.oid_partition_id = 0; + +	if (dest_err->oer_component.oid_object_id != +				src_err->oer_component.oid_object_id) +		dest_err->oer_component.oid_object_id = 0; + +	if (dest_err->oer_comp_offset > src_err->oer_comp_offset) +		dest_err->oer_comp_offset = src_err->oer_comp_offset; + +	dest_end = end_offset(dest_err->oer_comp_offset, +			      dest_err->oer_comp_length); +	src_end =  end_offset(src_err->oer_comp_offset, +			      src_err->oer_comp_length); +	if (dest_end < src_end) +		dest_end = src_end; + +	dest_err->oer_comp_length = dest_end - dest_err->oer_comp_offset; + +	if ((src_err->oer_iswrite == dest_err->oer_iswrite) && +	    (err_prio(src_err->oer_errno) > err_prio(dest_err->oer_errno))) { +			dest_err->oer_errno = src_err->oer_errno; +	} else if (src_err->oer_iswrite) { +		dest_err->oer_iswrite = true; +		dest_err->oer_errno = src_err->oer_errno; +	} +} + +static void +encode_accumulated_error(struct objlayout *objlay, __be32 *p) +{ +	struct objlayout_io_res *oir, *tmp; +	struct pnfs_osd_ioerr accumulated_err = {.oer_errno = 0}; + +	list_for_each_entry_safe(oir, tmp, &objlay->err_list, err_list) { +		unsigned i; + +		for (i = 0; i < oir->num_comps; i++) { +			struct pnfs_osd_ioerr *ioerr = &oir->ioerrs[i]; + +			if (!ioerr->oer_errno) +				continue; + +			printk(KERN_ERR "NFS: %s: err[%d]: errno=%d " +				"is_write=%d dev(%llx:%llx) par=0x%llx " +				"obj=0x%llx offset=0x%llx length=0x%llx\n", +				__func__, i, ioerr->oer_errno, +				ioerr->oer_iswrite, +				_DEVID_LO(&ioerr->oer_component.oid_device_id), +				_DEVID_HI(&ioerr->oer_component.oid_device_id), +				ioerr->oer_component.oid_partition_id, +				ioerr->oer_component.oid_object_id, +				ioerr->oer_comp_offset, +				ioerr->oer_comp_length); + +			merge_ioerr(&accumulated_err, ioerr); +		} +		list_del(&oir->err_list); +		objio_free_result(oir); +	} + +	pnfs_osd_xdr_encode_ioerr(p, &accumulated_err); +} + +void +objlayout_encode_layoutreturn(struct pnfs_layout_hdr *pnfslay, +			      struct xdr_stream *xdr, +			      const struct nfs4_layoutreturn_args *args) +{ +	struct objlayout *objlay = OBJLAYOUT(pnfslay); +	struct objlayout_io_res *oir, *tmp; +	__be32 *start; + +	dprintk("%s: Begin\n", __func__); +	start = xdr_reserve_space(xdr, 4); +	BUG_ON(!start); + +	spin_lock(&objlay->lock); + +	list_for_each_entry_safe(oir, tmp, &objlay->err_list, err_list) { +		__be32 *last_xdr = NULL, *p; +		unsigned i; +		int res = 0; + +		for (i = 0; i < oir->num_comps; i++) { +			struct pnfs_osd_ioerr *ioerr = &oir->ioerrs[i]; + +			if (!ioerr->oer_errno) +				continue; + +			dprintk("%s: err[%d]: errno=%d is_write=%d " +				"dev(%llx:%llx) par=0x%llx obj=0x%llx " +				"offset=0x%llx length=0x%llx\n", +				__func__, i, ioerr->oer_errno, +				ioerr->oer_iswrite, +				_DEVID_LO(&ioerr->oer_component.oid_device_id), +				_DEVID_HI(&ioerr->oer_component.oid_device_id), +				ioerr->oer_component.oid_partition_id, +				ioerr->oer_component.oid_object_id, +				ioerr->oer_comp_offset, +				ioerr->oer_comp_length); + +			p = pnfs_osd_xdr_ioerr_reserve_space(xdr); +			if (unlikely(!p)) { +				res = -E2BIG; +				break; /* accumulated_error */ +			} + +			last_xdr = p; +			pnfs_osd_xdr_encode_ioerr(p, &oir->ioerrs[i]); +		} + +		/* TODO: use xdr_write_pages */ +		if (unlikely(res)) { +			/* no space for even one error descriptor */ +			BUG_ON(!last_xdr); + +			/* we've encountered a situation with lots and lots of +			 * errors and no space to encode them all. Use the last +			 * available slot to report the union of all the +			 * remaining errors. +			 */ +			encode_accumulated_error(objlay, last_xdr); +			goto loop_done; +		} +		list_del(&oir->err_list); +		objio_free_result(oir); +	} +loop_done: +	spin_unlock(&objlay->lock); + +	*start = cpu_to_be32((xdr->p - start - 1) * 4); +	dprintk("%s: Return\n", __func__); +} + + +/* + * Get Device Info API for io engines + */ +struct objlayout_deviceinfo { +	struct page *page; +	struct pnfs_osd_deviceaddr da; /* This must be last */ +}; + +/* Initialize and call nfs_getdeviceinfo, then decode and return a + * "struct pnfs_osd_deviceaddr *" Eventually objlayout_put_deviceinfo() + * should be called. + */ +int objlayout_get_deviceinfo(struct pnfs_layout_hdr *pnfslay, +	struct nfs4_deviceid *d_id, struct pnfs_osd_deviceaddr **deviceaddr, +	gfp_t gfp_flags) +{ +	struct objlayout_deviceinfo *odi; +	struct pnfs_device pd; +	struct page *page, **pages; +	u32 *p; +	int err; + +	page = alloc_page(gfp_flags); +	if (!page) +		return -ENOMEM; + +	pages = &page; +	pd.pages = pages; + +	memcpy(&pd.dev_id, d_id, sizeof(*d_id)); +	pd.layout_type = LAYOUT_OSD2_OBJECTS; +	pd.pages = &page; +	pd.pgbase = 0; +	pd.pglen = PAGE_SIZE; +	pd.mincount = 0; +	pd.maxcount = PAGE_SIZE; + +	err = nfs4_proc_getdeviceinfo(NFS_SERVER(pnfslay->plh_inode), &pd, +			pnfslay->plh_lc_cred); +	dprintk("%s nfs_getdeviceinfo returned %d\n", __func__, err); +	if (err) +		goto err_out; + +	p = page_address(page); +	odi = kzalloc(sizeof(*odi), gfp_flags); +	if (!odi) { +		err = -ENOMEM; +		goto err_out; +	} +	pnfs_osd_xdr_decode_deviceaddr(&odi->da, p); +	odi->page = page; +	*deviceaddr = &odi->da; +	return 0; + +err_out: +	__free_page(page); +	return err; +} + +void objlayout_put_deviceinfo(struct pnfs_osd_deviceaddr *deviceaddr) +{ +	struct objlayout_deviceinfo *odi = container_of(deviceaddr, +						struct objlayout_deviceinfo, +						da); + +	__free_page(odi->page); +	kfree(odi); +} + +enum { +	OBJLAYOUT_MAX_URI_LEN = 256, OBJLAYOUT_MAX_OSDNAME_LEN = 64, +	OBJLAYOUT_MAX_SYSID_HEX_LEN = OSD_SYSTEMID_LEN * 2 + 1, +	OSD_LOGIN_UPCALL_PATHLEN  = 256 +}; + +static char osd_login_prog[OSD_LOGIN_UPCALL_PATHLEN] = "/sbin/osd_login"; + +module_param_string(osd_login_prog, osd_login_prog, sizeof(osd_login_prog), +		    0600); +MODULE_PARM_DESC(osd_login_prog, "Path to the osd_login upcall program"); + +struct __auto_login { +	char uri[OBJLAYOUT_MAX_URI_LEN]; +	char osdname[OBJLAYOUT_MAX_OSDNAME_LEN]; +	char systemid_hex[OBJLAYOUT_MAX_SYSID_HEX_LEN]; +}; + +static int __objlayout_upcall(struct __auto_login *login) +{ +	static char *envp[] = { "HOME=/", +		"TERM=linux", +		"PATH=/sbin:/usr/sbin:/bin:/usr/bin", +		NULL +	}; +	char *argv[8]; +	int ret; + +	if (unlikely(!osd_login_prog[0])) { +		dprintk("%s: osd_login_prog is disabled\n", __func__); +		return -EACCES; +	} + +	dprintk("%s uri: %s\n", __func__, login->uri); +	dprintk("%s osdname %s\n", __func__, login->osdname); +	dprintk("%s systemid_hex %s\n", __func__, login->systemid_hex); + +	argv[0] = (char *)osd_login_prog; +	argv[1] = "-u"; +	argv[2] = login->uri; +	argv[3] = "-o"; +	argv[4] = login->osdname; +	argv[5] = "-s"; +	argv[6] = login->systemid_hex; +	argv[7] = NULL; + +	ret = call_usermodehelper(argv[0], argv, envp, UMH_WAIT_PROC); +	/* +	 * Disable the upcall mechanism if we're getting an ENOENT or +	 * EACCES error. The admin can re-enable it on the fly by using +	 * sysfs to set the objlayoutdriver.osd_login_prog module parameter once +	 * the problem has been fixed. +	 */ +	if (ret == -ENOENT || ret == -EACCES) { +		printk(KERN_ERR "PNFS-OBJ: %s was not found please set " +			"objlayoutdriver.osd_login_prog kernel parameter!\n", +			osd_login_prog); +		osd_login_prog[0] = '\0'; +	} +	dprintk("%s %s return value: %d\n", __func__, osd_login_prog, ret); + +	return ret; +} + +/* Assume dest is all zeros */ +static void __copy_nfsS_and_zero_terminate(struct nfs4_string s, +					   char *dest, int max_len, +					   const char *var_name) +{ +	if (!s.len) +		return; + +	if (s.len >= max_len) { +		pr_warn_ratelimited( +			"objlayout_autologin: %s: s.len(%d) >= max_len(%d)", +			var_name, s.len, max_len); +		s.len = max_len - 1; /* space for null terminator */ +	} + +	memcpy(dest, s.data, s.len); +} + +/* Assume sysid is all zeros */ +static void _sysid_2_hex(struct nfs4_string s, +		  char sysid[OBJLAYOUT_MAX_SYSID_HEX_LEN]) +{ +	int i; +	char *cur; + +	if (!s.len) +		return; + +	if (s.len != OSD_SYSTEMID_LEN) { +		pr_warn_ratelimited( +		    "objlayout_autologin: systemid_len(%d) != OSD_SYSTEMID_LEN", +		    s.len); +		if (s.len > OSD_SYSTEMID_LEN) +			s.len = OSD_SYSTEMID_LEN; +	} + +	cur = sysid; +	for (i = 0; i < s.len; i++) +		cur = hex_byte_pack(cur, s.data[i]); +} + +int objlayout_autologin(struct pnfs_osd_deviceaddr *deviceaddr) +{ +	int rc; +	struct __auto_login login; + +	if (!deviceaddr->oda_targetaddr.ota_netaddr.r_addr.len) +		return -ENODEV; + +	memset(&login, 0, sizeof(login)); +	__copy_nfsS_and_zero_terminate( +		deviceaddr->oda_targetaddr.ota_netaddr.r_addr, +		login.uri, sizeof(login.uri), "URI"); + +	__copy_nfsS_and_zero_terminate( +		deviceaddr->oda_osdname, +		login.osdname, sizeof(login.osdname), "OSDNAME"); + +	_sysid_2_hex(deviceaddr->oda_systemid, login.systemid_hex); + +	rc = __objlayout_upcall(&login); +	if (rc > 0) /* script returns positive values */ +		rc = -ENODEV; + +	return rc; +}  | 
