Linux Audio

Check our new training course

Loading...
v6.13.7
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
   4 * All Rights Reserved.
   5 */
   6#include "xfs.h"
   7#include "xfs_fs.h"
   8#include "xfs_shared.h"
   9#include "xfs_format.h"
  10#include "xfs_log_format.h"
  11#include "xfs_trans_resv.h"
  12#include "xfs_mount.h"
  13#include "xfs_inode.h"
  14#include "xfs_rtalloc.h"
  15#include "xfs_iwalk.h"
  16#include "xfs_itable.h"
  17#include "xfs_error.h"
  18#include "xfs_da_format.h"
  19#include "xfs_da_btree.h"
  20#include "xfs_attr.h"
  21#include "xfs_bmap.h"
  22#include "xfs_bmap_util.h"
  23#include "xfs_fsops.h"
  24#include "xfs_discard.h"
  25#include "xfs_quota.h"
 
  26#include "xfs_trace.h"
  27#include "xfs_icache.h"
  28#include "xfs_trans.h"
 
  29#include "xfs_btree.h"
  30#include <linux/fsmap.h>
  31#include "xfs_fsmap.h"
  32#include "scrub/xfs_scrub.h"
  33#include "xfs_sb.h"
  34#include "xfs_ag.h"
  35#include "xfs_health.h"
  36#include "xfs_reflink.h"
  37#include "xfs_ioctl.h"
  38#include "xfs_xattr.h"
  39#include "xfs_rtbitmap.h"
  40#include "xfs_file.h"
  41#include "xfs_exchrange.h"
  42#include "xfs_handle.h"
  43#include "xfs_rtgroup.h"
  44
  45#include <linux/mount.h>
  46#include <linux/fileattr.h>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  47
  48/* Return 0 on success or positive error */
  49int
  50xfs_fsbulkstat_one_fmt(
  51	struct xfs_ibulk		*breq,
  52	const struct xfs_bulkstat	*bstat)
  53{
  54	struct xfs_bstat		bs1;
  55
  56	xfs_bulkstat_to_bstat(breq->mp, &bs1, bstat);
  57	if (copy_to_user(breq->ubuffer, &bs1, sizeof(bs1)))
  58		return -EFAULT;
  59	return xfs_ibulk_advance(breq, sizeof(struct xfs_bstat));
  60}
  61
  62int
  63xfs_fsinumbers_fmt(
  64	struct xfs_ibulk		*breq,
  65	const struct xfs_inumbers	*igrp)
  66{
  67	struct xfs_inogrp		ig1;
  68
  69	xfs_inumbers_to_inogrp(&ig1, igrp);
  70	if (copy_to_user(breq->ubuffer, &ig1, sizeof(struct xfs_inogrp)))
  71		return -EFAULT;
  72	return xfs_ibulk_advance(breq, sizeof(struct xfs_inogrp));
  73}
  74
  75STATIC int
  76xfs_ioc_fsbulkstat(
  77	struct file		*file,
  78	unsigned int		cmd,
  79	void			__user *arg)
  80{
  81	struct xfs_mount	*mp = XFS_I(file_inode(file))->i_mount;
  82	struct xfs_fsop_bulkreq	bulkreq;
  83	struct xfs_ibulk	breq = {
  84		.mp		= mp,
  85		.idmap		= file_mnt_idmap(file),
  86		.ocount		= 0,
  87	};
  88	xfs_ino_t		lastino;
  89	int			error;
  90
  91	/* done = 1 if there are more stats to get and if bulkstat */
  92	/* should be called again (unused here, but used in dmapi) */
  93
  94	if (!capable(CAP_SYS_ADMIN))
  95		return -EPERM;
  96
  97	if (xfs_is_shutdown(mp))
  98		return -EIO;
  99
 100	if (copy_from_user(&bulkreq, arg, sizeof(struct xfs_fsop_bulkreq)))
 101		return -EFAULT;
 102
 103	if (copy_from_user(&lastino, bulkreq.lastip, sizeof(__s64)))
 104		return -EFAULT;
 105
 106	if (bulkreq.icount <= 0)
 107		return -EINVAL;
 108
 109	if (bulkreq.ubuffer == NULL)
 110		return -EINVAL;
 111
 112	breq.ubuffer = bulkreq.ubuffer;
 113	breq.icount = bulkreq.icount;
 114
 115	/*
 116	 * FSBULKSTAT_SINGLE expects that *lastip contains the inode number
 117	 * that we want to stat.  However, FSINUMBERS and FSBULKSTAT expect
 118	 * that *lastip contains either zero or the number of the last inode to
 119	 * be examined by the previous call and return results starting with
 120	 * the next inode after that.  The new bulk request back end functions
 121	 * take the inode to start with, so we have to compute the startino
 122	 * parameter from lastino to maintain correct function.  lastino == 0
 123	 * is a special case because it has traditionally meant "first inode
 124	 * in filesystem".
 125	 */
 126	if (cmd == XFS_IOC_FSINUMBERS) {
 127		breq.startino = lastino ? lastino + 1 : 0;
 128		error = xfs_inumbers(&breq, xfs_fsinumbers_fmt);
 129		lastino = breq.startino - 1;
 130	} else if (cmd == XFS_IOC_FSBULKSTAT_SINGLE) {
 131		breq.startino = lastino;
 132		breq.icount = 1;
 133		error = xfs_bulkstat_one(&breq, xfs_fsbulkstat_one_fmt);
 134	} else {	/* XFS_IOC_FSBULKSTAT */
 135		breq.startino = lastino ? lastino + 1 : 0;
 136		error = xfs_bulkstat(&breq, xfs_fsbulkstat_one_fmt);
 137		lastino = breq.startino - 1;
 138	}
 139
 140	if (error)
 141		return error;
 142
 143	if (bulkreq.lastip != NULL &&
 144	    copy_to_user(bulkreq.lastip, &lastino, sizeof(xfs_ino_t)))
 145		return -EFAULT;
 146
 147	if (bulkreq.ocount != NULL &&
 148	    copy_to_user(bulkreq.ocount, &breq.ocount, sizeof(__s32)))
 149		return -EFAULT;
 150
 151	return 0;
 152}
 153
 154/* Return 0 on success or positive error */
 155static int
 156xfs_bulkstat_fmt(
 157	struct xfs_ibulk		*breq,
 158	const struct xfs_bulkstat	*bstat)
 159{
 160	if (copy_to_user(breq->ubuffer, bstat, sizeof(struct xfs_bulkstat)))
 161		return -EFAULT;
 162	return xfs_ibulk_advance(breq, sizeof(struct xfs_bulkstat));
 163}
 164
 165/*
 166 * Check the incoming bulk request @hdr from userspace and initialize the
 167 * internal @breq bulk request appropriately.  Returns 0 if the bulk request
 168 * should proceed; -ECANCELED if there's nothing to do; or the usual
 169 * negative error code.
 170 */
 171static int
 172xfs_bulk_ireq_setup(
 173	struct xfs_mount	*mp,
 174	const struct xfs_bulk_ireq *hdr,
 175	struct xfs_ibulk	*breq,
 176	void __user		*ubuffer)
 177{
 178	if (hdr->icount == 0 ||
 179	    (hdr->flags & ~XFS_BULK_IREQ_FLAGS_ALL) ||
 180	    memchr_inv(hdr->reserved, 0, sizeof(hdr->reserved)))
 181		return -EINVAL;
 182
 183	breq->startino = hdr->ino;
 184	breq->ubuffer = ubuffer;
 185	breq->icount = hdr->icount;
 186	breq->ocount = 0;
 187	breq->flags = 0;
 188
 189	/*
 190	 * The @ino parameter is a special value, so we must look it up here.
 191	 * We're not allowed to have IREQ_AGNO, and we only return one inode
 192	 * worth of data.
 193	 */
 194	if (hdr->flags & XFS_BULK_IREQ_SPECIAL) {
 195		if (hdr->flags & XFS_BULK_IREQ_AGNO)
 196			return -EINVAL;
 197
 198		switch (hdr->ino) {
 199		case XFS_BULK_IREQ_SPECIAL_ROOT:
 200			breq->startino = mp->m_sb.sb_rootino;
 201			break;
 202		default:
 203			return -EINVAL;
 204		}
 205		breq->icount = 1;
 206	}
 207
 208	/*
 209	 * The IREQ_AGNO flag means that we only want results from a given AG.
 210	 * If @hdr->ino is zero, we start iterating in that AG.  If @hdr->ino is
 211	 * beyond the specified AG then we return no results.
 212	 */
 213	if (hdr->flags & XFS_BULK_IREQ_AGNO) {
 214		if (hdr->agno >= mp->m_sb.sb_agcount)
 215			return -EINVAL;
 216
 217		if (breq->startino == 0)
 218			breq->startino = XFS_AGINO_TO_INO(mp, hdr->agno, 0);
 219		else if (XFS_INO_TO_AGNO(mp, breq->startino) < hdr->agno)
 220			return -EINVAL;
 221
 222		breq->flags |= XFS_IBULK_SAME_AG;
 223
 224		/* Asking for an inode past the end of the AG?  We're done! */
 225		if (XFS_INO_TO_AGNO(mp, breq->startino) > hdr->agno)
 226			return -ECANCELED;
 227	} else if (hdr->agno)
 228		return -EINVAL;
 229
 230	/* Asking for an inode past the end of the FS?  We're done! */
 231	if (XFS_INO_TO_AGNO(mp, breq->startino) >= mp->m_sb.sb_agcount)
 232		return -ECANCELED;
 233
 234	if (hdr->flags & XFS_BULK_IREQ_NREXT64)
 235		breq->flags |= XFS_IBULK_NREXT64;
 236
 237	/* Caller wants to see metadata directories in bulkstat output. */
 238	if (hdr->flags & XFS_BULK_IREQ_METADIR)
 239		breq->flags |= XFS_IBULK_METADIR;
 240
 241	return 0;
 242}
 243
 244/*
 245 * Update the userspace bulk request @hdr to reflect the end state of the
 246 * internal bulk request @breq.
 247 */
 248static void
 249xfs_bulk_ireq_teardown(
 250	struct xfs_bulk_ireq	*hdr,
 251	struct xfs_ibulk	*breq)
 252{
 253	hdr->ino = breq->startino;
 254	hdr->ocount = breq->ocount;
 255}
 256
 257/* Handle the v5 bulkstat ioctl. */
 258STATIC int
 259xfs_ioc_bulkstat(
 260	struct file			*file,
 261	unsigned int			cmd,
 262	struct xfs_bulkstat_req __user	*arg)
 263{
 264	struct xfs_mount		*mp = XFS_I(file_inode(file))->i_mount;
 265	struct xfs_bulk_ireq		hdr;
 266	struct xfs_ibulk		breq = {
 267		.mp			= mp,
 268		.idmap			= file_mnt_idmap(file),
 269	};
 270	int				error;
 271
 272	if (!capable(CAP_SYS_ADMIN))
 273		return -EPERM;
 274
 275	if (xfs_is_shutdown(mp))
 276		return -EIO;
 277
 278	if (copy_from_user(&hdr, &arg->hdr, sizeof(hdr)))
 279		return -EFAULT;
 280
 281	error = xfs_bulk_ireq_setup(mp, &hdr, &breq, arg->bulkstat);
 282	if (error == -ECANCELED)
 283		goto out_teardown;
 284	if (error < 0)
 285		return error;
 286
 287	error = xfs_bulkstat(&breq, xfs_bulkstat_fmt);
 288	if (error)
 289		return error;
 290
 291out_teardown:
 292	xfs_bulk_ireq_teardown(&hdr, &breq);
 293	if (copy_to_user(&arg->hdr, &hdr, sizeof(hdr)))
 294		return -EFAULT;
 295
 296	return 0;
 297}
 298
 299STATIC int
 300xfs_inumbers_fmt(
 301	struct xfs_ibulk		*breq,
 302	const struct xfs_inumbers	*igrp)
 303{
 304	if (copy_to_user(breq->ubuffer, igrp, sizeof(struct xfs_inumbers)))
 305		return -EFAULT;
 306	return xfs_ibulk_advance(breq, sizeof(struct xfs_inumbers));
 307}
 308
 309/* Handle the v5 inumbers ioctl. */
 310STATIC int
 311xfs_ioc_inumbers(
 312	struct xfs_mount		*mp,
 313	unsigned int			cmd,
 314	struct xfs_inumbers_req __user	*arg)
 315{
 316	struct xfs_bulk_ireq		hdr;
 317	struct xfs_ibulk		breq = {
 318		.mp			= mp,
 319	};
 320	int				error;
 321
 322	if (!capable(CAP_SYS_ADMIN))
 323		return -EPERM;
 324
 325	if (xfs_is_shutdown(mp))
 326		return -EIO;
 327
 328	if (copy_from_user(&hdr, &arg->hdr, sizeof(hdr)))
 329		return -EFAULT;
 330
 331	if (hdr.flags & XFS_BULK_IREQ_METADIR)
 332		return -EINVAL;
 333
 334	error = xfs_bulk_ireq_setup(mp, &hdr, &breq, arg->inumbers);
 335	if (error == -ECANCELED)
 336		goto out_teardown;
 337	if (error < 0)
 338		return error;
 339
 340	error = xfs_inumbers(&breq, xfs_inumbers_fmt);
 341	if (error)
 342		return error;
 343
 344out_teardown:
 345	xfs_bulk_ireq_teardown(&hdr, &breq);
 346	if (copy_to_user(&arg->hdr, &hdr, sizeof(hdr)))
 347		return -EFAULT;
 348
 349	return 0;
 350}
 351
 352STATIC int
 353xfs_ioc_fsgeometry(
 354	struct xfs_mount	*mp,
 355	void			__user *arg,
 356	int			struct_version)
 357{
 358	struct xfs_fsop_geom	fsgeo;
 359	size_t			len;
 360
 361	xfs_fs_geometry(mp, &fsgeo, struct_version);
 362
 363	if (struct_version <= 3)
 364		len = sizeof(struct xfs_fsop_geom_v1);
 365	else if (struct_version == 4)
 366		len = sizeof(struct xfs_fsop_geom_v4);
 367	else {
 368		xfs_fsop_geom_health(mp, &fsgeo);
 369		len = sizeof(fsgeo);
 370	}
 371
 372	if (copy_to_user(arg, &fsgeo, len))
 373		return -EFAULT;
 374	return 0;
 375}
 376
 377STATIC int
 378xfs_ioc_ag_geometry(
 379	struct xfs_mount	*mp,
 380	void			__user *arg)
 381{
 382	struct xfs_perag	*pag;
 383	struct xfs_ag_geometry	ageo;
 384	int			error;
 385
 386	if (copy_from_user(&ageo, arg, sizeof(ageo)))
 387		return -EFAULT;
 388	if (ageo.ag_flags)
 389		return -EINVAL;
 390	if (memchr_inv(&ageo.ag_reserved, 0, sizeof(ageo.ag_reserved)))
 391		return -EINVAL;
 392
 393	pag = xfs_perag_get(mp, ageo.ag_number);
 394	if (!pag)
 395		return -EINVAL;
 396
 397	error = xfs_ag_get_geometry(pag, &ageo);
 398	xfs_perag_put(pag);
 399	if (error)
 400		return error;
 401
 402	if (copy_to_user(arg, &ageo, sizeof(ageo)))
 403		return -EFAULT;
 404	return 0;
 405}
 406
 407STATIC int
 408xfs_ioc_rtgroup_geometry(
 409	struct xfs_mount	*mp,
 410	void			__user *arg)
 411{
 412	struct xfs_rtgroup	*rtg;
 413	struct xfs_rtgroup_geometry rgeo;
 414	int			error;
 415
 416	if (copy_from_user(&rgeo, arg, sizeof(rgeo)))
 417		return -EFAULT;
 418	if (rgeo.rg_flags)
 419		return -EINVAL;
 420	if (memchr_inv(&rgeo.rg_reserved, 0, sizeof(rgeo.rg_reserved)))
 421		return -EINVAL;
 422	if (!xfs_has_rtgroups(mp))
 423		return -EINVAL;
 424
 425	rtg = xfs_rtgroup_get(mp, rgeo.rg_number);
 426	if (!rtg)
 427		return -EINVAL;
 
 
 
 428
 429	error = xfs_rtgroup_get_geometry(rtg, &rgeo);
 430	xfs_rtgroup_put(rtg);
 431	if (error)
 432		return error;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 433
 434	if (copy_to_user(arg, &rgeo, sizeof(rgeo)))
 435		return -EFAULT;
 436	return 0;
 437}
 438
 439/*
 440 * Linux extended inode flags interface.
 441 */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 442
 443static void
 444xfs_fill_fsxattr(
 445	struct xfs_inode	*ip,
 446	int			whichfork,
 447	struct fileattr		*fa)
 448{
 449	struct xfs_mount	*mp = ip->i_mount;
 450	struct xfs_ifork	*ifp = xfs_ifork_ptr(ip, whichfork);
 451
 452	fileattr_fill_xflags(fa, xfs_ip2xflags(ip));
 453
 454	if (ip->i_diflags & XFS_DIFLAG_EXTSIZE) {
 455		fa->fsx_extsize = XFS_FSB_TO_B(mp, ip->i_extsize);
 456	} else if (ip->i_diflags & XFS_DIFLAG_EXTSZINHERIT) {
 457		/*
 458		 * Don't let a misaligned extent size hint on a directory
 459		 * escape to userspace if it won't pass the setattr checks
 460		 * later.
 461		 */
 462		if ((ip->i_diflags & XFS_DIFLAG_RTINHERIT) &&
 463		    xfs_extlen_to_rtxmod(mp, ip->i_extsize) > 0) {
 464			fa->fsx_xflags &= ~(FS_XFLAG_EXTSIZE |
 465					    FS_XFLAG_EXTSZINHERIT);
 466			fa->fsx_extsize = 0;
 467		} else {
 468			fa->fsx_extsize = XFS_FSB_TO_B(mp, ip->i_extsize);
 469		}
 470	}
 471
 472	if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE)
 473		fa->fsx_cowextsize = XFS_FSB_TO_B(mp, ip->i_cowextsize);
 474	fa->fsx_projid = ip->i_projid;
 475	if (ifp && !xfs_need_iread_extents(ifp))
 476		fa->fsx_nextents = xfs_iext_count(ifp);
 477	else
 478		fa->fsx_nextents = xfs_ifork_nextents(ifp);
 479}
 480
 481STATIC int
 482xfs_ioc_fsgetxattra(
 483	xfs_inode_t		*ip,
 
 484	void			__user *arg)
 485{
 486	struct fileattr		fa;
 487
 488	xfs_ilock(ip, XFS_ILOCK_SHARED);
 489	xfs_fill_fsxattr(ip, XFS_ATTR_FORK, &fa);
 490	xfs_iunlock(ip, XFS_ILOCK_SHARED);
 491
 492	return copy_fsxattr_to_user(&fa, arg);
 
 
 493}
 494
 495int
 496xfs_fileattr_get(
 497	struct dentry		*dentry,
 498	struct fileattr		*fa)
 499{
 500	struct xfs_inode	*ip = XFS_I(d_inode(dentry));
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 501
 502	if (d_is_special(dentry))
 503		return -ENOTTY;
 504
 505	xfs_ilock(ip, XFS_ILOCK_SHARED);
 506	xfs_fill_fsxattr(ip, XFS_DATA_FORK, fa);
 507	xfs_iunlock(ip, XFS_ILOCK_SHARED);
 
 
 
 
 508
 509	return 0;
 
 
 
 
 
 510}
 511
 512static int
 513xfs_ioctl_setattr_xflags(
 514	struct xfs_trans	*tp,
 515	struct xfs_inode	*ip,
 516	struct fileattr		*fa)
 517{
 518	struct xfs_mount	*mp = ip->i_mount;
 519	bool			rtflag = (fa->fsx_xflags & FS_XFLAG_REALTIME);
 520	uint64_t		i_flags2;
 521
 522	if (rtflag != XFS_IS_REALTIME_INODE(ip)) {
 523		/* Can't change realtime flag if any extents are allocated. */
 524		if (xfs_inode_has_filedata(ip))
 525			return -EINVAL;
 526
 527		/*
 528		 * If S_DAX is enabled on this file, we can only switch the
 529		 * device if both support fsdax.  We can't update S_DAX because
 530		 * there might be other threads walking down the access paths.
 531		 */
 532		if (IS_DAX(VFS_I(ip)) &&
 533		    (mp->m_ddev_targp->bt_daxdev == NULL ||
 534		     (mp->m_rtdev_targp &&
 535		      mp->m_rtdev_targp->bt_daxdev == NULL)))
 536			return -EINVAL;
 537	}
 538
 539	if (rtflag) {
 540		/* If realtime flag is set then must have realtime device */
 541		if (mp->m_sb.sb_rblocks == 0 || mp->m_sb.sb_rextsize == 0 ||
 542		    xfs_extlen_to_rtxmod(mp, ip->i_extsize))
 543			return -EINVAL;
 544
 545		/* Clear reflink if we are actually able to set the rt flag. */
 546		if (xfs_is_reflink_inode(ip))
 547			ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
 548	}
 549
 550	/* diflags2 only valid for v3 inodes. */
 551	i_flags2 = xfs_flags2diflags2(ip, fa->fsx_xflags);
 552	if (i_flags2 && !xfs_has_v3inodes(mp))
 553		return -EINVAL;
 554
 555	ip->i_diflags = xfs_flags2diflags(ip, fa->fsx_xflags);
 556	ip->i_diflags2 = i_flags2;
 557
 558	xfs_diflags_to_iflags(ip, false);
 559
 560	/*
 561	 * Make the stable writes flag match that of the device the inode
 562	 * resides on when flipping the RT flag.
 563	 */
 564	if (rtflag != XFS_IS_REALTIME_INODE(ip) && S_ISREG(VFS_I(ip)->i_mode))
 565		xfs_update_stable_writes(ip);
 566
 567	xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_CHG);
 568	xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
 569	XFS_STATS_INC(mp, xs_ig_attrchg);
 570	return 0;
 571}
 572
 573static void
 574xfs_ioctl_setattr_prepare_dax(
 575	struct xfs_inode	*ip,
 576	struct fileattr		*fa)
 577{
 578	struct xfs_mount	*mp = ip->i_mount;
 579	struct inode            *inode = VFS_I(ip);
 580
 581	if (S_ISDIR(inode->i_mode))
 582		return;
 583
 584	if (xfs_has_dax_always(mp) || xfs_has_dax_never(mp))
 
 585		return;
 586
 587	if (((fa->fsx_xflags & FS_XFLAG_DAX) &&
 588	    !(ip->i_diflags2 & XFS_DIFLAG2_DAX)) ||
 589	    (!(fa->fsx_xflags & FS_XFLAG_DAX) &&
 590	     (ip->i_diflags2 & XFS_DIFLAG2_DAX)))
 591		d_mark_dontcache(inode);
 592}
 593
 594/*
 595 * Set up the transaction structure for the setattr operation, checking that we
 596 * have permission to do so. On success, return a clean transaction and the
 597 * inode locked exclusively ready for further operation specific checks. On
 598 * failure, return an error without modifying or locking the inode.
 599 */
 600static struct xfs_trans *
 601xfs_ioctl_setattr_get_trans(
 602	struct xfs_inode	*ip,
 603	struct xfs_dquot	*pdqp)
 604{
 605	struct xfs_mount	*mp = ip->i_mount;
 606	struct xfs_trans	*tp;
 607	int			error = -EROFS;
 608
 609	if (xfs_is_readonly(mp))
 610		goto out_error;
 611	error = -EIO;
 612	if (xfs_is_shutdown(mp))
 613		goto out_error;
 614
 615	error = xfs_trans_alloc_ichange(ip, NULL, NULL, pdqp,
 616			has_capability_noaudit(current, CAP_FOWNER), &tp);
 617	if (error)
 618		goto out_error;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 619
 620	if (xfs_has_wsync(mp))
 621		xfs_trans_set_sync(tp);
 622
 623	return tp;
 624
 625out_error:
 
 
 626	return ERR_PTR(error);
 627}
 628
 629/*
 630 * Validate a proposed extent size hint.  For regular files, the hint can only
 631 * be changed if no extents are allocated.
 
 
 
 
 
 
 
 
 
 
 
 
 632 */
 633static int
 634xfs_ioctl_setattr_check_extsize(
 635	struct xfs_inode	*ip,
 636	struct fileattr		*fa)
 637{
 638	struct xfs_mount	*mp = ip->i_mount;
 639	xfs_failaddr_t		failaddr;
 640	uint16_t		new_diflags;
 641
 642	if (!fa->fsx_valid)
 643		return 0;
 644
 645	if (S_ISREG(VFS_I(ip)->i_mode) && xfs_inode_has_filedata(ip) &&
 646	    XFS_FSB_TO_B(mp, ip->i_extsize) != fa->fsx_extsize)
 647		return -EINVAL;
 648
 649	if (fa->fsx_extsize & mp->m_blockmask)
 650		return -EINVAL;
 651
 652	new_diflags = xfs_flags2diflags(ip, fa->fsx_xflags);
 653
 654	/*
 655	 * Inode verifiers do not check that the extent size hint is an integer
 656	 * multiple of the rt extent size on a directory with both rtinherit
 657	 * and extszinherit flags set.  Don't let sysadmins misconfigure
 658	 * directories.
 659	 */
 660	if ((new_diflags & XFS_DIFLAG_RTINHERIT) &&
 661	    (new_diflags & XFS_DIFLAG_EXTSZINHERIT)) {
 662		unsigned int	rtextsize_bytes;
 663
 664		rtextsize_bytes = XFS_FSB_TO_B(mp, mp->m_sb.sb_rextsize);
 665		if (fa->fsx_extsize % rtextsize_bytes)
 
 
 
 
 666			return -EINVAL;
 667	}
 668
 669	failaddr = xfs_inode_validate_extsize(ip->i_mount,
 670			XFS_B_TO_FSB(mp, fa->fsx_extsize),
 671			VFS_I(ip)->i_mode, new_diflags);
 672	return failaddr != NULL ? -EINVAL : 0;
 673}
 674
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 675static int
 676xfs_ioctl_setattr_check_cowextsize(
 677	struct xfs_inode	*ip,
 678	struct fileattr		*fa)
 679{
 680	struct xfs_mount	*mp = ip->i_mount;
 681	xfs_failaddr_t		failaddr;
 682	uint64_t		new_diflags2;
 683	uint16_t		new_diflags;
 
 
 
 
 
 684
 685	if (!fa->fsx_valid)
 686		return 0;
 687
 688	if (fa->fsx_cowextsize & mp->m_blockmask)
 
 689		return -EINVAL;
 690
 691	new_diflags = xfs_flags2diflags(ip, fa->fsx_xflags);
 692	new_diflags2 = xfs_flags2diflags2(ip, fa->fsx_xflags);
 
 693
 694	failaddr = xfs_inode_validate_cowextsize(ip->i_mount,
 695			XFS_B_TO_FSB(mp, fa->fsx_cowextsize),
 696			VFS_I(ip)->i_mode, new_diflags, new_diflags2);
 697	return failaddr != NULL ? -EINVAL : 0;
 698}
 699
 700static int
 701xfs_ioctl_setattr_check_projid(
 702	struct xfs_inode	*ip,
 703	struct fileattr		*fa)
 704{
 705	if (!fa->fsx_valid)
 706		return 0;
 707
 708	/* Disallow 32bit project ids if 32bit IDs are not enabled. */
 709	if (fa->fsx_projid > (uint16_t)-1 &&
 710	    !xfs_has_projid32(ip->i_mount))
 711		return -EINVAL;
 712	return 0;
 713}
 714
 715int
 716xfs_fileattr_set(
 717	struct mnt_idmap	*idmap,
 718	struct dentry		*dentry,
 719	struct fileattr		*fa)
 720{
 721	struct xfs_inode	*ip = XFS_I(d_inode(dentry));
 722	struct xfs_mount	*mp = ip->i_mount;
 723	struct xfs_trans	*tp;
 724	struct xfs_dquot	*pdqp = NULL;
 725	struct xfs_dquot	*olddquot = NULL;
 726	int			error;
 727
 728	trace_xfs_ioctl_setattr(ip);
 729
 730	if (d_is_special(dentry))
 731		return -ENOTTY;
 732
 733	if (!fa->fsx_valid) {
 734		if (fa->flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL |
 735				  FS_NOATIME_FL | FS_NODUMP_FL |
 736				  FS_SYNC_FL | FS_DAX_FL | FS_PROJINHERIT_FL))
 737			return -EOPNOTSUPP;
 738	}
 739
 740	error = xfs_ioctl_setattr_check_projid(ip, fa);
 741	if (error)
 742		return error;
 743
 744	/*
 745	 * If disk quotas is on, we make sure that the dquots do exist on disk,
 746	 * before we start any other transactions. Trying to do this later
 747	 * is messy. We don't care to take a readlock to look at the ids
 748	 * in inode here, because we can't hold it across the trans_reserve.
 749	 * If the IDs do change before we take the ilock, we're covered
 750	 * because the i_*dquot fields will get updated anyway.
 751	 */
 752	if (fa->fsx_valid && XFS_IS_QUOTA_ON(mp)) {
 753		error = xfs_qm_vop_dqalloc(ip, VFS_I(ip)->i_uid,
 754				VFS_I(ip)->i_gid, fa->fsx_projid,
 755				XFS_QMOPT_PQUOTA, NULL, NULL, &pdqp);
 756		if (error)
 757			return error;
 758	}
 759
 760	xfs_ioctl_setattr_prepare_dax(ip, fa);
 761
 762	tp = xfs_ioctl_setattr_get_trans(ip, pdqp);
 763	if (IS_ERR(tp)) {
 764		error = PTR_ERR(tp);
 765		goto error_free_dquots;
 766	}
 767
 768	error = xfs_ioctl_setattr_check_extsize(ip, fa);
 769	if (error)
 
 
 
 
 
 
 
 
 
 770		goto error_trans_cancel;
 771
 772	error = xfs_ioctl_setattr_check_cowextsize(ip, fa);
 773	if (error)
 774		goto error_trans_cancel;
 775
 776	error = xfs_ioctl_setattr_xflags(tp, ip, fa);
 777	if (error)
 
 
 
 
 778		goto error_trans_cancel;
 779
 780	if (!fa->fsx_valid)
 781		goto skip_xattr;
 782	/*
 783	 * Change file ownership.  Must be the owner or privileged.  CAP_FSETID
 784	 * overrides the following restrictions:
 785	 *
 786	 * The set-user-ID and set-group-ID bits of a file will be cleared upon
 787	 * successful return from chown()
 788	 */
 789
 790	if ((VFS_I(ip)->i_mode & (S_ISUID|S_ISGID)) &&
 791	    !capable_wrt_inode_uidgid(idmap, VFS_I(ip), CAP_FSETID))
 792		VFS_I(ip)->i_mode &= ~(S_ISUID|S_ISGID);
 793
 794	/* Change the ownerships and register project quota modifications */
 795	if (ip->i_projid != fa->fsx_projid) {
 796		if (XFS_IS_PQUOTA_ON(mp)) {
 797			olddquot = xfs_qm_vop_chown(tp, ip,
 798						&ip->i_pdquot, pdqp);
 799		}
 800		ip->i_projid = fa->fsx_projid;
 801	}
 802
 803	/*
 804	 * Only set the extent size hint if we've already determined that the
 805	 * extent size hint should be set on the inode. If no extent size flags
 806	 * are set on the inode then unconditionally clear the extent size hint.
 807	 */
 808	if (ip->i_diflags & (XFS_DIFLAG_EXTSIZE | XFS_DIFLAG_EXTSZINHERIT))
 809		ip->i_extsize = XFS_B_TO_FSB(mp, fa->fsx_extsize);
 810	else
 811		ip->i_extsize = 0;
 812
 813	if (xfs_has_v3inodes(mp)) {
 814		if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE)
 815			ip->i_cowextsize = XFS_B_TO_FSB(mp, fa->fsx_cowextsize);
 816		else
 817			ip->i_cowextsize = 0;
 818	}
 819
 820skip_xattr:
 821	error = xfs_trans_commit(tp);
 822
 823	/*
 824	 * Release any dquot(s) the inode had kept before chown.
 825	 */
 826	xfs_qm_dqrele(olddquot);
 827	xfs_qm_dqrele(pdqp);
 828
 829	return error;
 830
 831error_trans_cancel:
 832	xfs_trans_cancel(tp);
 833error_free_dquots:
 834	xfs_qm_dqrele(pdqp);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 835	return error;
 836}
 837
 838static bool
 839xfs_getbmap_format(
 840	struct kgetbmap		*p,
 841	struct getbmapx __user	*u,
 842	size_t			recsize)
 843{
 844	if (put_user(p->bmv_offset, &u->bmv_offset) ||
 845	    put_user(p->bmv_block, &u->bmv_block) ||
 846	    put_user(p->bmv_length, &u->bmv_length) ||
 847	    put_user(0, &u->bmv_count) ||
 848	    put_user(0, &u->bmv_entries))
 849		return false;
 850	if (recsize < sizeof(struct getbmapx))
 851		return true;
 852	if (put_user(0, &u->bmv_iflags) ||
 853	    put_user(p->bmv_oflags, &u->bmv_oflags) ||
 854	    put_user(0, &u->bmv_unused1) ||
 855	    put_user(0, &u->bmv_unused2))
 856		return false;
 857	return true;
 858}
 859
 860STATIC int
 861xfs_ioc_getbmap(
 862	struct file		*file,
 863	unsigned int		cmd,
 864	void			__user *arg)
 865{
 866	struct getbmapx		bmx = { 0 };
 867	struct kgetbmap		*buf;
 868	size_t			recsize;
 869	int			error, i;
 870
 871	switch (cmd) {
 872	case XFS_IOC_GETBMAPA:
 873		bmx.bmv_iflags = BMV_IF_ATTRFORK;
 874		fallthrough;
 875	case XFS_IOC_GETBMAP:
 
 
 876		/* struct getbmap is a strict subset of struct getbmapx. */
 877		recsize = sizeof(struct getbmap);
 878		break;
 879	case XFS_IOC_GETBMAPX:
 880		recsize = sizeof(struct getbmapx);
 881		break;
 882	default:
 883		return -EINVAL;
 884	}
 885
 886	if (copy_from_user(&bmx, arg, recsize))
 887		return -EFAULT;
 888
 889	if (bmx.bmv_count < 2)
 890		return -EINVAL;
 891	if (bmx.bmv_count >= INT_MAX / recsize)
 892		return -ENOMEM;
 893
 894	buf = kvcalloc(bmx.bmv_count, sizeof(*buf), GFP_KERNEL);
 895	if (!buf)
 896		return -ENOMEM;
 897
 898	error = xfs_getbmap(XFS_I(file_inode(file)), &bmx, buf);
 899	if (error)
 900		goto out_free_buf;
 901
 902	error = -EFAULT;
 903	if (copy_to_user(arg, &bmx, recsize))
 904		goto out_free_buf;
 905	arg += recsize;
 906
 907	for (i = 0; i < bmx.bmv_entries; i++) {
 908		if (!xfs_getbmap_format(buf + i, arg, recsize))
 909			goto out_free_buf;
 910		arg += recsize;
 911	}
 912
 913	error = 0;
 914out_free_buf:
 915	kvfree(buf);
 916	return error;
 917}
 918
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 919int
 920xfs_ioc_swapext(
 921	xfs_swapext_t	*sxp)
 922{
 923	xfs_inode_t     *ip, *tip;
 
 
 924
 925	/* Pull information for the target fd */
 926	CLASS(fd, f)((int)sxp->sx_fdtarget);
 927	if (fd_empty(f))
 928		return -EINVAL;
 929
 930	if (!(fd_file(f)->f_mode & FMODE_WRITE) ||
 931	    !(fd_file(f)->f_mode & FMODE_READ) ||
 932	    (fd_file(f)->f_flags & O_APPEND))
 933		return -EBADF;
 934
 935	CLASS(fd, tmp)((int)sxp->sx_fdtmp);
 936	if (fd_empty(tmp))
 937		return -EINVAL;
 938
 939	if (!(fd_file(tmp)->f_mode & FMODE_WRITE) ||
 940	    !(fd_file(tmp)->f_mode & FMODE_READ) ||
 941	    (fd_file(tmp)->f_flags & O_APPEND))
 942		return -EBADF;
 943
 944	if (IS_SWAPFILE(file_inode(fd_file(f))) ||
 945	    IS_SWAPFILE(file_inode(fd_file(tmp))))
 946		return -EINVAL;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 947
 948	/*
 949	 * We need to ensure that the fds passed in point to XFS inodes
 950	 * before we cast and access them as XFS structures as we have no
 951	 * control over what the user passes us here.
 952	 */
 953	if (fd_file(f)->f_op != &xfs_file_operations ||
 954	    fd_file(tmp)->f_op != &xfs_file_operations)
 955		return -EINVAL;
 
 
 956
 957	ip = XFS_I(file_inode(fd_file(f)));
 958	tip = XFS_I(file_inode(fd_file(tmp)));
 959
 960	if (ip->i_mount != tip->i_mount)
 961		return -EINVAL;
 
 
 962
 963	if (ip->i_ino == tip->i_ino)
 964		return -EINVAL;
 
 
 965
 966	if (xfs_is_shutdown(ip->i_mount))
 967		return -EIO;
 
 
 968
 969	return xfs_swap_extents(ip, tip, sxp);
 
 
 
 
 
 
 
 970}
 971
 972static int
 973xfs_ioc_getlabel(
 974	struct xfs_mount	*mp,
 975	char			__user *user_label)
 976{
 977	struct xfs_sb		*sbp = &mp->m_sb;
 978	char			label[XFSLABEL_MAX + 1];
 979
 980	/* Paranoia */
 981	BUILD_BUG_ON(sizeof(sbp->sb_fname) > FSLABEL_MAX);
 982
 983	/* 1 larger than sb_fname, so this ensures a trailing NUL char */
 984	memset(label, 0, sizeof(label));
 985	spin_lock(&mp->m_sb_lock);
 986	strncpy(label, sbp->sb_fname, XFSLABEL_MAX);
 987	spin_unlock(&mp->m_sb_lock);
 988
 989	if (copy_to_user(user_label, label, sizeof(label)))
 990		return -EFAULT;
 991	return 0;
 992}
 993
 994static int
 995xfs_ioc_setlabel(
 996	struct file		*filp,
 997	struct xfs_mount	*mp,
 998	char			__user *newlabel)
 999{
1000	struct xfs_sb		*sbp = &mp->m_sb;
1001	char			label[XFSLABEL_MAX + 1];
1002	size_t			len;
1003	int			error;
1004
1005	if (!capable(CAP_SYS_ADMIN))
1006		return -EPERM;
1007	/*
1008	 * The generic ioctl allows up to FSLABEL_MAX chars, but XFS is much
1009	 * smaller, at 12 bytes.  We copy one more to be sure we find the
1010	 * (required) NULL character to test the incoming label length.
1011	 * NB: The on disk label doesn't need to be null terminated.
1012	 */
1013	if (copy_from_user(label, newlabel, XFSLABEL_MAX + 1))
1014		return -EFAULT;
1015	len = strnlen(label, XFSLABEL_MAX + 1);
1016	if (len > sizeof(sbp->sb_fname))
1017		return -EINVAL;
1018
1019	error = mnt_want_write_file(filp);
1020	if (error)
1021		return error;
1022
1023	spin_lock(&mp->m_sb_lock);
1024	memset(sbp->sb_fname, 0, sizeof(sbp->sb_fname));
1025	memcpy(sbp->sb_fname, label, len);
1026	spin_unlock(&mp->m_sb_lock);
1027
1028	/*
1029	 * Now we do several things to satisfy userspace.
1030	 * In addition to normal logging of the primary superblock, we also
1031	 * immediately write these changes to sector zero for the primary, then
1032	 * update all backup supers (as xfs_db does for a label change), then
1033	 * invalidate the block device page cache.  This is so that any prior
1034	 * buffered reads from userspace (i.e. from blkid) are invalidated,
1035	 * and userspace will see the newly-written label.
1036	 */
1037	error = xfs_sync_sb_buf(mp, true);
1038	if (error)
1039		goto out;
1040	/*
1041	 * growfs also updates backup supers so lock against that.
1042	 */
1043	mutex_lock(&mp->m_growlock);
1044	error = xfs_update_secondary_sbs(mp);
1045	mutex_unlock(&mp->m_growlock);
1046
1047	invalidate_bdev(mp->m_ddev_targp->bt_bdev);
1048	if (xfs_has_rtsb(mp) && mp->m_rtdev_targp)
1049		invalidate_bdev(mp->m_rtdev_targp->bt_bdev);
1050
1051out:
1052	mnt_drop_write_file(filp);
1053	return error;
1054}
1055
1056static inline int
1057xfs_fs_eofblocks_from_user(
1058	struct xfs_fs_eofblocks		*src,
1059	struct xfs_icwalk		*dst)
1060{
1061	if (src->eof_version != XFS_EOFBLOCKS_VERSION)
1062		return -EINVAL;
1063
1064	if (src->eof_flags & ~XFS_EOF_FLAGS_VALID)
1065		return -EINVAL;
1066
1067	if (memchr_inv(&src->pad32, 0, sizeof(src->pad32)) ||
1068	    memchr_inv(src->pad64, 0, sizeof(src->pad64)))
1069		return -EINVAL;
1070
1071	dst->icw_flags = 0;
1072	if (src->eof_flags & XFS_EOF_FLAGS_SYNC)
1073		dst->icw_flags |= XFS_ICWALK_FLAG_SYNC;
1074	if (src->eof_flags & XFS_EOF_FLAGS_UID)
1075		dst->icw_flags |= XFS_ICWALK_FLAG_UID;
1076	if (src->eof_flags & XFS_EOF_FLAGS_GID)
1077		dst->icw_flags |= XFS_ICWALK_FLAG_GID;
1078	if (src->eof_flags & XFS_EOF_FLAGS_PRID)
1079		dst->icw_flags |= XFS_ICWALK_FLAG_PRID;
1080	if (src->eof_flags & XFS_EOF_FLAGS_MINFILESIZE)
1081		dst->icw_flags |= XFS_ICWALK_FLAG_MINFILESIZE;
1082
1083	dst->icw_prid = src->eof_prid;
1084	dst->icw_min_file_size = src->eof_min_file_size;
1085
1086	dst->icw_uid = INVALID_UID;
1087	if (src->eof_flags & XFS_EOF_FLAGS_UID) {
1088		dst->icw_uid = make_kuid(current_user_ns(), src->eof_uid);
1089		if (!uid_valid(dst->icw_uid))
1090			return -EINVAL;
1091	}
1092
1093	dst->icw_gid = INVALID_GID;
1094	if (src->eof_flags & XFS_EOF_FLAGS_GID) {
1095		dst->icw_gid = make_kgid(current_user_ns(), src->eof_gid);
1096		if (!gid_valid(dst->icw_gid))
1097			return -EINVAL;
1098	}
1099	return 0;
1100}
1101
1102static int
1103xfs_ioctl_getset_resblocks(
1104	struct file		*filp,
1105	unsigned int		cmd,
1106	void __user		*arg)
1107{
1108	struct xfs_mount	*mp = XFS_I(file_inode(filp))->i_mount;
1109	struct xfs_fsop_resblks	fsop = { };
1110	int			error;
1111
1112	if (!capable(CAP_SYS_ADMIN))
1113		return -EPERM;
1114
1115	if (cmd == XFS_IOC_SET_RESBLKS) {
1116		if (xfs_is_readonly(mp))
1117			return -EROFS;
1118
1119		if (copy_from_user(&fsop, arg, sizeof(fsop)))
1120			return -EFAULT;
1121
1122		error = mnt_want_write_file(filp);
1123		if (error)
1124			return error;
1125		error = xfs_reserve_blocks(mp, fsop.resblks);
1126		mnt_drop_write_file(filp);
1127		if (error)
1128			return error;
1129	}
1130
1131	spin_lock(&mp->m_sb_lock);
1132	fsop.resblks = mp->m_resblks;
1133	fsop.resblks_avail = mp->m_resblks_avail;
1134	spin_unlock(&mp->m_sb_lock);
1135
1136	if (copy_to_user(arg, &fsop, sizeof(fsop)))
1137		return -EFAULT;
1138	return 0;
1139}
1140
1141static int
1142xfs_ioctl_fs_counts(
1143	struct xfs_mount	*mp,
1144	struct xfs_fsop_counts __user	*uarg)
1145{
1146	struct xfs_fsop_counts	out = {
1147		.allocino = percpu_counter_read_positive(&mp->m_icount),
1148		.freeino  = percpu_counter_read_positive(&mp->m_ifree),
1149		.freedata = percpu_counter_read_positive(&mp->m_fdblocks) -
1150				xfs_fdblocks_unavailable(mp),
1151		.freertx  = percpu_counter_read_positive(&mp->m_frextents),
1152	};
1153
1154	if (copy_to_user(uarg, &out, sizeof(out)))
1155		return -EFAULT;
1156	return 0;
1157}
1158
1159/*
1160 * These long-unused ioctls were removed from the official ioctl API in 5.17,
1161 * but retain these definitions so that we can log warnings about them.
1162 */
1163#define XFS_IOC_ALLOCSP		_IOW ('X', 10, struct xfs_flock64)
1164#define XFS_IOC_FREESP		_IOW ('X', 11, struct xfs_flock64)
1165#define XFS_IOC_ALLOCSP64	_IOW ('X', 36, struct xfs_flock64)
1166#define XFS_IOC_FREESP64	_IOW ('X', 37, struct xfs_flock64)
1167
1168/*
1169 * Note: some of the ioctl's return positive numbers as a
1170 * byte count indicating success, such as readlink_by_handle.
1171 * So we don't "sign flip" like most other routines.  This means
1172 * true errors need to be returned as a negative value.
1173 */
1174long
1175xfs_file_ioctl(
1176	struct file		*filp,
1177	unsigned int		cmd,
1178	unsigned long		p)
1179{
1180	struct inode		*inode = file_inode(filp);
1181	struct xfs_inode	*ip = XFS_I(inode);
1182	struct xfs_mount	*mp = ip->i_mount;
1183	void			__user *arg = (void __user *)p;
1184	int			error;
1185
1186	trace_xfs_file_ioctl(ip);
1187
1188	switch (cmd) {
1189	case FITRIM:
1190		return xfs_ioc_trim(mp, arg);
1191	case FS_IOC_GETFSLABEL:
1192		return xfs_ioc_getlabel(mp, arg);
1193	case FS_IOC_SETFSLABEL:
1194		return xfs_ioc_setlabel(filp, mp, arg);
1195	case XFS_IOC_ALLOCSP:
1196	case XFS_IOC_FREESP:
1197	case XFS_IOC_ALLOCSP64:
1198	case XFS_IOC_FREESP64:
1199		xfs_warn_once(mp,
1200	"%s should use fallocate; XFS_IOC_{ALLOC,FREE}SP ioctl unsupported",
1201				current->comm);
1202		return -ENOTTY;
 
 
1203	case XFS_IOC_DIOINFO: {
1204		struct xfs_buftarg	*target = xfs_inode_buftarg(ip);
1205		struct dioattr		da;
1206
1207		da.d_mem =  da.d_miniosz = target->bt_logical_sectorsize;
1208		da.d_maxiosz = INT_MAX & ~(da.d_miniosz - 1);
1209
1210		if (copy_to_user(arg, &da, sizeof(da)))
1211			return -EFAULT;
1212		return 0;
1213	}
1214
1215	case XFS_IOC_FSBULKSTAT_SINGLE:
1216	case XFS_IOC_FSBULKSTAT:
1217	case XFS_IOC_FSINUMBERS:
1218		return xfs_ioc_fsbulkstat(filp, cmd, arg);
1219
1220	case XFS_IOC_BULKSTAT:
1221		return xfs_ioc_bulkstat(filp, cmd, arg);
1222	case XFS_IOC_INUMBERS:
1223		return xfs_ioc_inumbers(mp, cmd, arg);
1224
1225	case XFS_IOC_FSGEOMETRY_V1:
1226		return xfs_ioc_fsgeometry(mp, arg, 3);
1227	case XFS_IOC_FSGEOMETRY_V4:
1228		return xfs_ioc_fsgeometry(mp, arg, 4);
1229	case XFS_IOC_FSGEOMETRY:
1230		return xfs_ioc_fsgeometry(mp, arg, 5);
1231
1232	case XFS_IOC_AG_GEOMETRY:
1233		return xfs_ioc_ag_geometry(mp, arg);
1234	case XFS_IOC_RTGROUP_GEOMETRY:
1235		return xfs_ioc_rtgroup_geometry(mp, arg);
1236
1237	case XFS_IOC_GETVERSION:
1238		return put_user(inode->i_generation, (int __user *)arg);
1239
 
 
1240	case XFS_IOC_FSGETXATTRA:
1241		return xfs_ioc_fsgetxattra(ip, arg);
1242	case XFS_IOC_GETPARENTS:
1243		return xfs_ioc_getparents(filp, arg);
1244	case XFS_IOC_GETPARENTS_BY_HANDLE:
1245		return xfs_ioc_getparents_by_handle(filp, arg);
 
 
 
1246	case XFS_IOC_GETBMAP:
1247	case XFS_IOC_GETBMAPA:
1248	case XFS_IOC_GETBMAPX:
1249		return xfs_ioc_getbmap(filp, cmd, arg);
1250
1251	case FS_IOC_GETFSMAP:
1252		return xfs_ioc_getfsmap(ip, arg);
1253
1254	case XFS_IOC_SCRUBV_METADATA:
1255		return xfs_ioc_scrubv_metadata(filp, arg);
1256	case XFS_IOC_SCRUB_METADATA:
1257		return xfs_ioc_scrub_metadata(filp, arg);
1258
1259	case XFS_IOC_FD_TO_HANDLE:
1260	case XFS_IOC_PATH_TO_HANDLE:
1261	case XFS_IOC_PATH_TO_FSHANDLE: {
1262		xfs_fsop_handlereq_t	hreq;
1263
1264		if (copy_from_user(&hreq, arg, sizeof(hreq)))
1265			return -EFAULT;
1266		return xfs_find_handle(cmd, &hreq);
1267	}
1268	case XFS_IOC_OPEN_BY_HANDLE: {
1269		xfs_fsop_handlereq_t	hreq;
1270
1271		if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t)))
1272			return -EFAULT;
1273		return xfs_open_by_handle(filp, &hreq);
1274	}
1275
1276	case XFS_IOC_READLINK_BY_HANDLE: {
1277		xfs_fsop_handlereq_t	hreq;
1278
1279		if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t)))
1280			return -EFAULT;
1281		return xfs_readlink_by_handle(filp, &hreq);
1282	}
1283	case XFS_IOC_ATTRLIST_BY_HANDLE:
1284		return xfs_attrlist_by_handle(filp, arg);
1285
1286	case XFS_IOC_ATTRMULTI_BY_HANDLE:
1287		return xfs_attrmulti_by_handle(filp, arg);
1288
1289	case XFS_IOC_SWAPEXT: {
1290		struct xfs_swapext	sxp;
1291
1292		if (copy_from_user(&sxp, arg, sizeof(xfs_swapext_t)))
1293			return -EFAULT;
1294		error = mnt_want_write_file(filp);
1295		if (error)
1296			return error;
1297		error = xfs_ioc_swapext(&sxp);
1298		mnt_drop_write_file(filp);
1299		return error;
1300	}
1301
1302	case XFS_IOC_FSCOUNTS:
1303		return xfs_ioctl_fs_counts(mp, arg);
1304
1305	case XFS_IOC_SET_RESBLKS:
1306	case XFS_IOC_GET_RESBLKS:
1307		return xfs_ioctl_getset_resblocks(filp, cmd, arg);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1308
1309	case XFS_IOC_FSGROWFSDATA: {
1310		struct xfs_growfs_data in;
1311
1312		if (copy_from_user(&in, arg, sizeof(in)))
1313			return -EFAULT;
1314
1315		error = mnt_want_write_file(filp);
1316		if (error)
1317			return error;
1318		error = xfs_growfs_data(mp, &in);
1319		mnt_drop_write_file(filp);
1320		return error;
1321	}
1322
1323	case XFS_IOC_FSGROWFSLOG: {
1324		struct xfs_growfs_log in;
1325
1326		if (copy_from_user(&in, arg, sizeof(in)))
1327			return -EFAULT;
1328
1329		error = mnt_want_write_file(filp);
1330		if (error)
1331			return error;
1332		error = xfs_growfs_log(mp, &in);
1333		mnt_drop_write_file(filp);
1334		return error;
1335	}
1336
1337	case XFS_IOC_FSGROWFSRT: {
1338		xfs_growfs_rt_t in;
1339
1340		if (copy_from_user(&in, arg, sizeof(in)))
1341			return -EFAULT;
1342
1343		error = mnt_want_write_file(filp);
1344		if (error)
1345			return error;
1346		error = xfs_growfs_rt(mp, &in);
1347		mnt_drop_write_file(filp);
1348		return error;
1349	}
1350
1351	case XFS_IOC_GOINGDOWN: {
1352		uint32_t in;
1353
1354		if (!capable(CAP_SYS_ADMIN))
1355			return -EPERM;
1356
1357		if (get_user(in, (uint32_t __user *)arg))
1358			return -EFAULT;
1359
1360		return xfs_fs_goingdown(mp, in);
1361	}
1362
1363	case XFS_IOC_ERROR_INJECTION: {
1364		xfs_error_injection_t in;
1365
1366		if (!capable(CAP_SYS_ADMIN))
1367			return -EPERM;
1368
1369		if (copy_from_user(&in, arg, sizeof(in)))
1370			return -EFAULT;
1371
1372		return xfs_errortag_add(mp, in.errtag);
1373	}
1374
1375	case XFS_IOC_ERROR_CLEARALL:
1376		if (!capable(CAP_SYS_ADMIN))
1377			return -EPERM;
1378
1379		return xfs_errortag_clearall(mp);
1380
1381	case XFS_IOC_FREE_EOFBLOCKS: {
1382		struct xfs_fs_eofblocks	eofb;
1383		struct xfs_icwalk	icw;
1384
1385		if (!capable(CAP_SYS_ADMIN))
1386			return -EPERM;
1387
1388		if (xfs_is_readonly(mp))
1389			return -EROFS;
1390
1391		if (copy_from_user(&eofb, arg, sizeof(eofb)))
1392			return -EFAULT;
1393
1394		error = xfs_fs_eofblocks_from_user(&eofb, &icw);
1395		if (error)
1396			return error;
1397
1398		trace_xfs_ioc_free_eofblocks(mp, &icw, _RET_IP_);
1399
1400		sb_start_write(mp->m_super);
1401		error = xfs_blockgc_free_space(mp, &icw);
1402		sb_end_write(mp->m_super);
1403		return error;
1404	}
1405
1406	case XFS_IOC_EXCHANGE_RANGE:
1407		return xfs_ioc_exchange_range(filp, arg);
1408	case XFS_IOC_START_COMMIT:
1409		return xfs_ioc_start_commit(filp, arg);
1410	case XFS_IOC_COMMIT_RANGE:
1411		return xfs_ioc_commit_range(filp, arg);
1412
1413	default:
1414		return -ENOTTY;
1415	}
1416}
v5.9
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
   4 * All Rights Reserved.
   5 */
   6#include "xfs.h"
   7#include "xfs_fs.h"
   8#include "xfs_shared.h"
   9#include "xfs_format.h"
  10#include "xfs_log_format.h"
  11#include "xfs_trans_resv.h"
  12#include "xfs_mount.h"
  13#include "xfs_inode.h"
  14#include "xfs_rtalloc.h"
  15#include "xfs_iwalk.h"
  16#include "xfs_itable.h"
  17#include "xfs_error.h"
 
 
  18#include "xfs_attr.h"
  19#include "xfs_bmap.h"
  20#include "xfs_bmap_util.h"
  21#include "xfs_fsops.h"
  22#include "xfs_discard.h"
  23#include "xfs_quota.h"
  24#include "xfs_export.h"
  25#include "xfs_trace.h"
  26#include "xfs_icache.h"
  27#include "xfs_trans.h"
  28#include "xfs_acl.h"
  29#include "xfs_btree.h"
  30#include <linux/fsmap.h>
  31#include "xfs_fsmap.h"
  32#include "scrub/xfs_scrub.h"
  33#include "xfs_sb.h"
  34#include "xfs_ag.h"
  35#include "xfs_health.h"
  36#include "xfs_reflink.h"
  37#include "xfs_ioctl.h"
  38#include "xfs_da_format.h"
  39#include "xfs_da_btree.h"
 
 
 
 
  40
  41#include <linux/mount.h>
  42#include <linux/namei.h>
  43
  44/*
  45 * xfs_find_handle maps from userspace xfs_fsop_handlereq structure to
  46 * a file or fs handle.
  47 *
  48 * XFS_IOC_PATH_TO_FSHANDLE
  49 *    returns fs handle for a mount point or path within that mount point
  50 * XFS_IOC_FD_TO_HANDLE
  51 *    returns full handle for a FD opened in user space
  52 * XFS_IOC_PATH_TO_HANDLE
  53 *    returns full handle for a path
  54 */
  55int
  56xfs_find_handle(
  57	unsigned int		cmd,
  58	xfs_fsop_handlereq_t	*hreq)
  59{
  60	int			hsize;
  61	xfs_handle_t		handle;
  62	struct inode		*inode;
  63	struct fd		f = {NULL};
  64	struct path		path;
  65	int			error;
  66	struct xfs_inode	*ip;
  67
  68	if (cmd == XFS_IOC_FD_TO_HANDLE) {
  69		f = fdget(hreq->fd);
  70		if (!f.file)
  71			return -EBADF;
  72		inode = file_inode(f.file);
  73	} else {
  74		error = user_path_at(AT_FDCWD, hreq->path, 0, &path);
  75		if (error)
  76			return error;
  77		inode = d_inode(path.dentry);
  78	}
  79	ip = XFS_I(inode);
  80
  81	/*
  82	 * We can only generate handles for inodes residing on a XFS filesystem,
  83	 * and only for regular files, directories or symbolic links.
  84	 */
  85	error = -EINVAL;
  86	if (inode->i_sb->s_magic != XFS_SB_MAGIC)
  87		goto out_put;
  88
  89	error = -EBADF;
  90	if (!S_ISREG(inode->i_mode) &&
  91	    !S_ISDIR(inode->i_mode) &&
  92	    !S_ISLNK(inode->i_mode))
  93		goto out_put;
  94
  95
  96	memcpy(&handle.ha_fsid, ip->i_mount->m_fixedfsid, sizeof(xfs_fsid_t));
  97
  98	if (cmd == XFS_IOC_PATH_TO_FSHANDLE) {
  99		/*
 100		 * This handle only contains an fsid, zero the rest.
 101		 */
 102		memset(&handle.ha_fid, 0, sizeof(handle.ha_fid));
 103		hsize = sizeof(xfs_fsid_t);
 104	} else {
 105		handle.ha_fid.fid_len = sizeof(xfs_fid_t) -
 106					sizeof(handle.ha_fid.fid_len);
 107		handle.ha_fid.fid_pad = 0;
 108		handle.ha_fid.fid_gen = inode->i_generation;
 109		handle.ha_fid.fid_ino = ip->i_ino;
 110		hsize = sizeof(xfs_handle_t);
 111	}
 112
 113	error = -EFAULT;
 114	if (copy_to_user(hreq->ohandle, &handle, hsize) ||
 115	    copy_to_user(hreq->ohandlen, &hsize, sizeof(__s32)))
 116		goto out_put;
 117
 118	error = 0;
 119
 120 out_put:
 121	if (cmd == XFS_IOC_FD_TO_HANDLE)
 122		fdput(f);
 123	else
 124		path_put(&path);
 125	return error;
 126}
 127
 128/*
 129 * No need to do permission checks on the various pathname components
 130 * as the handle operations are privileged.
 131 */
 132STATIC int
 133xfs_handle_acceptable(
 134	void			*context,
 135	struct dentry		*dentry)
 136{
 137	return 1;
 138}
 139
 140/*
 141 * Convert userspace handle data into a dentry.
 142 */
 143struct dentry *
 144xfs_handle_to_dentry(
 145	struct file		*parfilp,
 146	void __user		*uhandle,
 147	u32			hlen)
 148{
 149	xfs_handle_t		handle;
 150	struct xfs_fid64	fid;
 151
 152	/*
 153	 * Only allow handle opens under a directory.
 154	 */
 155	if (!S_ISDIR(file_inode(parfilp)->i_mode))
 156		return ERR_PTR(-ENOTDIR);
 157
 158	if (hlen != sizeof(xfs_handle_t))
 159		return ERR_PTR(-EINVAL);
 160	if (copy_from_user(&handle, uhandle, hlen))
 161		return ERR_PTR(-EFAULT);
 162	if (handle.ha_fid.fid_len !=
 163	    sizeof(handle.ha_fid) - sizeof(handle.ha_fid.fid_len))
 164		return ERR_PTR(-EINVAL);
 165
 166	memset(&fid, 0, sizeof(struct fid));
 167	fid.ino = handle.ha_fid.fid_ino;
 168	fid.gen = handle.ha_fid.fid_gen;
 169
 170	return exportfs_decode_fh(parfilp->f_path.mnt, (struct fid *)&fid, 3,
 171			FILEID_INO32_GEN | XFS_FILEID_TYPE_64FLAG,
 172			xfs_handle_acceptable, NULL);
 173}
 174
 175STATIC struct dentry *
 176xfs_handlereq_to_dentry(
 177	struct file		*parfilp,
 178	xfs_fsop_handlereq_t	*hreq)
 179{
 180	return xfs_handle_to_dentry(parfilp, hreq->ihandle, hreq->ihandlen);
 181}
 182
 183int
 184xfs_open_by_handle(
 185	struct file		*parfilp,
 186	xfs_fsop_handlereq_t	*hreq)
 187{
 188	const struct cred	*cred = current_cred();
 189	int			error;
 190	int			fd;
 191	int			permflag;
 192	struct file		*filp;
 193	struct inode		*inode;
 194	struct dentry		*dentry;
 195	fmode_t			fmode;
 196	struct path		path;
 197
 198	if (!capable(CAP_SYS_ADMIN))
 199		return -EPERM;
 200
 201	dentry = xfs_handlereq_to_dentry(parfilp, hreq);
 202	if (IS_ERR(dentry))
 203		return PTR_ERR(dentry);
 204	inode = d_inode(dentry);
 205
 206	/* Restrict xfs_open_by_handle to directories & regular files. */
 207	if (!(S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode))) {
 208		error = -EPERM;
 209		goto out_dput;
 210	}
 211
 212#if BITS_PER_LONG != 32
 213	hreq->oflags |= O_LARGEFILE;
 214#endif
 215
 216	permflag = hreq->oflags;
 217	fmode = OPEN_FMODE(permflag);
 218	if ((!(permflag & O_APPEND) || (permflag & O_TRUNC)) &&
 219	    (fmode & FMODE_WRITE) && IS_APPEND(inode)) {
 220		error = -EPERM;
 221		goto out_dput;
 222	}
 223
 224	if ((fmode & FMODE_WRITE) && IS_IMMUTABLE(inode)) {
 225		error = -EPERM;
 226		goto out_dput;
 227	}
 228
 229	/* Can't write directories. */
 230	if (S_ISDIR(inode->i_mode) && (fmode & FMODE_WRITE)) {
 231		error = -EISDIR;
 232		goto out_dput;
 233	}
 234
 235	fd = get_unused_fd_flags(0);
 236	if (fd < 0) {
 237		error = fd;
 238		goto out_dput;
 239	}
 240
 241	path.mnt = parfilp->f_path.mnt;
 242	path.dentry = dentry;
 243	filp = dentry_open(&path, hreq->oflags, cred);
 244	dput(dentry);
 245	if (IS_ERR(filp)) {
 246		put_unused_fd(fd);
 247		return PTR_ERR(filp);
 248	}
 249
 250	if (S_ISREG(inode->i_mode)) {
 251		filp->f_flags |= O_NOATIME;
 252		filp->f_mode |= FMODE_NOCMTIME;
 253	}
 254
 255	fd_install(fd, filp);
 256	return fd;
 257
 258 out_dput:
 259	dput(dentry);
 260	return error;
 261}
 262
 263int
 264xfs_readlink_by_handle(
 265	struct file		*parfilp,
 266	xfs_fsop_handlereq_t	*hreq)
 267{
 268	struct dentry		*dentry;
 269	__u32			olen;
 270	int			error;
 271
 272	if (!capable(CAP_SYS_ADMIN))
 273		return -EPERM;
 274
 275	dentry = xfs_handlereq_to_dentry(parfilp, hreq);
 276	if (IS_ERR(dentry))
 277		return PTR_ERR(dentry);
 278
 279	/* Restrict this handle operation to symlinks only. */
 280	if (!d_is_symlink(dentry)) {
 281		error = -EINVAL;
 282		goto out_dput;
 283	}
 284
 285	if (copy_from_user(&olen, hreq->ohandlen, sizeof(__u32))) {
 286		error = -EFAULT;
 287		goto out_dput;
 288	}
 289
 290	error = vfs_readlink(dentry, hreq->ohandle, olen);
 291
 292 out_dput:
 293	dput(dentry);
 294	return error;
 295}
 296
 297/*
 298 * Format an attribute and copy it out to the user's buffer.
 299 * Take care to check values and protect against them changing later,
 300 * we may be reading them directly out of a user buffer.
 301 */
 302static void
 303xfs_ioc_attr_put_listent(
 304	struct xfs_attr_list_context *context,
 305	int			flags,
 306	unsigned char		*name,
 307	int			namelen,
 308	int			valuelen)
 309{
 310	struct xfs_attrlist	*alist = context->buffer;
 311	struct xfs_attrlist_ent	*aep;
 312	int			arraytop;
 313
 314	ASSERT(!context->seen_enough);
 315	ASSERT(context->count >= 0);
 316	ASSERT(context->count < (ATTR_MAX_VALUELEN/8));
 317	ASSERT(context->firstu >= sizeof(*alist));
 318	ASSERT(context->firstu <= context->bufsize);
 319
 320	/*
 321	 * Only list entries in the right namespace.
 322	 */
 323	if (context->attr_filter != (flags & XFS_ATTR_NSP_ONDISK_MASK))
 324		return;
 325
 326	arraytop = sizeof(*alist) +
 327			context->count * sizeof(alist->al_offset[0]);
 328
 329	/* decrement by the actual bytes used by the attr */
 330	context->firstu -= round_up(offsetof(struct xfs_attrlist_ent, a_name) +
 331			namelen + 1, sizeof(uint32_t));
 332	if (context->firstu < arraytop) {
 333		trace_xfs_attr_list_full(context);
 334		alist->al_more = 1;
 335		context->seen_enough = 1;
 336		return;
 337	}
 338
 339	aep = context->buffer + context->firstu;
 340	aep->a_valuelen = valuelen;
 341	memcpy(aep->a_name, name, namelen);
 342	aep->a_name[namelen] = 0;
 343	alist->al_offset[context->count++] = context->firstu;
 344	alist->al_count = context->count;
 345	trace_xfs_attr_list_add(context);
 346}
 347
 348static unsigned int
 349xfs_attr_filter(
 350	u32			ioc_flags)
 351{
 352	if (ioc_flags & XFS_IOC_ATTR_ROOT)
 353		return XFS_ATTR_ROOT;
 354	if (ioc_flags & XFS_IOC_ATTR_SECURE)
 355		return XFS_ATTR_SECURE;
 356	return 0;
 357}
 358
 359static unsigned int
 360xfs_attr_flags(
 361	u32			ioc_flags)
 362{
 363	if (ioc_flags & XFS_IOC_ATTR_CREATE)
 364		return XATTR_CREATE;
 365	if (ioc_flags & XFS_IOC_ATTR_REPLACE)
 366		return XATTR_REPLACE;
 367	return 0;
 368}
 369
 370int
 371xfs_ioc_attr_list(
 372	struct xfs_inode		*dp,
 373	void __user			*ubuf,
 374	int				bufsize,
 375	int				flags,
 376	struct xfs_attrlist_cursor __user *ucursor)
 377{
 378	struct xfs_attr_list_context	context = { };
 379	struct xfs_attrlist		*alist;
 380	void				*buffer;
 381	int				error;
 382
 383	if (bufsize < sizeof(struct xfs_attrlist) ||
 384	    bufsize > XFS_XATTR_LIST_MAX)
 385		return -EINVAL;
 386
 387	/*
 388	 * Reject flags, only allow namespaces.
 389	 */
 390	if (flags & ~(XFS_IOC_ATTR_ROOT | XFS_IOC_ATTR_SECURE))
 391		return -EINVAL;
 392	if (flags == (XFS_IOC_ATTR_ROOT | XFS_IOC_ATTR_SECURE))
 393		return -EINVAL;
 394
 395	/*
 396	 * Validate the cursor.
 397	 */
 398	if (copy_from_user(&context.cursor, ucursor, sizeof(context.cursor)))
 399		return -EFAULT;
 400	if (context.cursor.pad1 || context.cursor.pad2)
 401		return -EINVAL;
 402	if (!context.cursor.initted &&
 403	    (context.cursor.hashval || context.cursor.blkno ||
 404	     context.cursor.offset))
 405		return -EINVAL;
 406
 407	buffer = kmem_zalloc_large(bufsize, 0);
 408	if (!buffer)
 409		return -ENOMEM;
 410
 411	/*
 412	 * Initialize the output buffer.
 413	 */
 414	context.dp = dp;
 415	context.resynch = 1;
 416	context.attr_filter = xfs_attr_filter(flags);
 417	context.buffer = buffer;
 418	context.bufsize = round_down(bufsize, sizeof(uint32_t));
 419	context.firstu = context.bufsize;
 420	context.put_listent = xfs_ioc_attr_put_listent;
 421
 422	alist = context.buffer;
 423	alist->al_count = 0;
 424	alist->al_more = 0;
 425	alist->al_offset[0] = context.bufsize;
 426
 427	error = xfs_attr_list(&context);
 428	if (error)
 429		goto out_free;
 430
 431	if (copy_to_user(ubuf, buffer, bufsize) ||
 432	    copy_to_user(ucursor, &context.cursor, sizeof(context.cursor)))
 433		error = -EFAULT;
 434out_free:
 435	kmem_free(buffer);
 436	return error;
 437}
 438
 439STATIC int
 440xfs_attrlist_by_handle(
 441	struct file		*parfilp,
 442	struct xfs_fsop_attrlist_handlereq __user *p)
 443{
 444	struct xfs_fsop_attrlist_handlereq al_hreq;
 445	struct dentry		*dentry;
 446	int			error = -ENOMEM;
 447
 448	if (!capable(CAP_SYS_ADMIN))
 449		return -EPERM;
 450	if (copy_from_user(&al_hreq, p, sizeof(al_hreq)))
 451		return -EFAULT;
 452
 453	dentry = xfs_handlereq_to_dentry(parfilp, &al_hreq.hreq);
 454	if (IS_ERR(dentry))
 455		return PTR_ERR(dentry);
 456
 457	error = xfs_ioc_attr_list(XFS_I(d_inode(dentry)), al_hreq.buffer,
 458				  al_hreq.buflen, al_hreq.flags, &p->pos);
 459	dput(dentry);
 460	return error;
 461}
 462
 463static int
 464xfs_attrmulti_attr_get(
 465	struct inode		*inode,
 466	unsigned char		*name,
 467	unsigned char		__user *ubuf,
 468	uint32_t		*len,
 469	uint32_t		flags)
 470{
 471	struct xfs_da_args	args = {
 472		.dp		= XFS_I(inode),
 473		.attr_filter	= xfs_attr_filter(flags),
 474		.attr_flags	= xfs_attr_flags(flags),
 475		.name		= name,
 476		.namelen	= strlen(name),
 477		.valuelen	= *len,
 478	};
 479	int			error;
 480
 481	if (*len > XFS_XATTR_SIZE_MAX)
 482		return -EINVAL;
 483
 484	error = xfs_attr_get(&args);
 485	if (error)
 486		goto out_kfree;
 487
 488	*len = args.valuelen;
 489	if (copy_to_user(ubuf, args.value, args.valuelen))
 490		error = -EFAULT;
 491
 492out_kfree:
 493	kmem_free(args.value);
 494	return error;
 495}
 496
 497static int
 498xfs_attrmulti_attr_set(
 499	struct inode		*inode,
 500	unsigned char		*name,
 501	const unsigned char	__user *ubuf,
 502	uint32_t		len,
 503	uint32_t		flags)
 504{
 505	struct xfs_da_args	args = {
 506		.dp		= XFS_I(inode),
 507		.attr_filter	= xfs_attr_filter(flags),
 508		.attr_flags	= xfs_attr_flags(flags),
 509		.name		= name,
 510		.namelen	= strlen(name),
 511	};
 512	int			error;
 513
 514	if (IS_IMMUTABLE(inode) || IS_APPEND(inode))
 515		return -EPERM;
 516
 517	if (ubuf) {
 518		if (len > XFS_XATTR_SIZE_MAX)
 519			return -EINVAL;
 520		args.value = memdup_user(ubuf, len);
 521		if (IS_ERR(args.value))
 522			return PTR_ERR(args.value);
 523		args.valuelen = len;
 524	}
 525
 526	error = xfs_attr_set(&args);
 527	if (!error && (flags & XFS_IOC_ATTR_ROOT))
 528		xfs_forget_acl(inode, name);
 529	kfree(args.value);
 530	return error;
 531}
 532
 533int
 534xfs_ioc_attrmulti_one(
 535	struct file		*parfilp,
 536	struct inode		*inode,
 537	uint32_t		opcode,
 538	void __user		*uname,
 539	void __user		*value,
 540	uint32_t		*len,
 541	uint32_t		flags)
 542{
 543	unsigned char		*name;
 544	int			error;
 545
 546	if ((flags & XFS_IOC_ATTR_ROOT) && (flags & XFS_IOC_ATTR_SECURE))
 547		return -EINVAL;
 548
 549	name = strndup_user(uname, MAXNAMELEN);
 550	if (IS_ERR(name))
 551		return PTR_ERR(name);
 552
 553	switch (opcode) {
 554	case ATTR_OP_GET:
 555		error = xfs_attrmulti_attr_get(inode, name, value, len, flags);
 556		break;
 557	case ATTR_OP_REMOVE:
 558		value = NULL;
 559		*len = 0;
 560		/* fall through */
 561	case ATTR_OP_SET:
 562		error = mnt_want_write_file(parfilp);
 563		if (error)
 564			break;
 565		error = xfs_attrmulti_attr_set(inode, name, value, *len, flags);
 566		mnt_drop_write_file(parfilp);
 567		break;
 568	default:
 569		error = -EINVAL;
 570		break;
 571	}
 572
 573	kfree(name);
 574	return error;
 575}
 576
 577STATIC int
 578xfs_attrmulti_by_handle(
 579	struct file		*parfilp,
 580	void			__user *arg)
 581{
 582	int			error;
 583	xfs_attr_multiop_t	*ops;
 584	xfs_fsop_attrmulti_handlereq_t am_hreq;
 585	struct dentry		*dentry;
 586	unsigned int		i, size;
 587
 588	if (!capable(CAP_SYS_ADMIN))
 589		return -EPERM;
 590	if (copy_from_user(&am_hreq, arg, sizeof(xfs_fsop_attrmulti_handlereq_t)))
 591		return -EFAULT;
 592
 593	/* overflow check */
 594	if (am_hreq.opcount >= INT_MAX / sizeof(xfs_attr_multiop_t))
 595		return -E2BIG;
 596
 597	dentry = xfs_handlereq_to_dentry(parfilp, &am_hreq.hreq);
 598	if (IS_ERR(dentry))
 599		return PTR_ERR(dentry);
 600
 601	error = -E2BIG;
 602	size = am_hreq.opcount * sizeof(xfs_attr_multiop_t);
 603	if (!size || size > 16 * PAGE_SIZE)
 604		goto out_dput;
 605
 606	ops = memdup_user(am_hreq.ops, size);
 607	if (IS_ERR(ops)) {
 608		error = PTR_ERR(ops);
 609		goto out_dput;
 610	}
 611
 612	error = 0;
 613	for (i = 0; i < am_hreq.opcount; i++) {
 614		ops[i].am_error = xfs_ioc_attrmulti_one(parfilp,
 615				d_inode(dentry), ops[i].am_opcode,
 616				ops[i].am_attrname, ops[i].am_attrvalue,
 617				&ops[i].am_length, ops[i].am_flags);
 618	}
 619
 620	if (copy_to_user(am_hreq.ops, ops, size))
 621		error = -EFAULT;
 622
 623	kfree(ops);
 624 out_dput:
 625	dput(dentry);
 626	return error;
 627}
 628
 629int
 630xfs_ioc_space(
 631	struct file		*filp,
 632	xfs_flock64_t		*bf)
 633{
 634	struct inode		*inode = file_inode(filp);
 635	struct xfs_inode	*ip = XFS_I(inode);
 636	struct iattr		iattr;
 637	enum xfs_prealloc_flags	flags = XFS_PREALLOC_CLEAR;
 638	uint			iolock = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL;
 639	int			error;
 640
 641	if (inode->i_flags & (S_IMMUTABLE|S_APPEND))
 642		return -EPERM;
 643
 644	if (!(filp->f_mode & FMODE_WRITE))
 645		return -EBADF;
 646
 647	if (!S_ISREG(inode->i_mode))
 648		return -EINVAL;
 649
 650	if (xfs_is_always_cow_inode(ip))
 651		return -EOPNOTSUPP;
 652
 653	if (filp->f_flags & O_DSYNC)
 654		flags |= XFS_PREALLOC_SYNC;
 655	if (filp->f_mode & FMODE_NOCMTIME)
 656		flags |= XFS_PREALLOC_INVISIBLE;
 657
 658	error = mnt_want_write_file(filp);
 659	if (error)
 660		return error;
 661
 662	xfs_ilock(ip, iolock);
 663	error = xfs_break_layouts(inode, &iolock, BREAK_UNMAP);
 664	if (error)
 665		goto out_unlock;
 666	inode_dio_wait(inode);
 667
 668	switch (bf->l_whence) {
 669	case 0: /*SEEK_SET*/
 670		break;
 671	case 1: /*SEEK_CUR*/
 672		bf->l_start += filp->f_pos;
 673		break;
 674	case 2: /*SEEK_END*/
 675		bf->l_start += XFS_ISIZE(ip);
 676		break;
 677	default:
 678		error = -EINVAL;
 679		goto out_unlock;
 680	}
 681
 682	if (bf->l_start < 0 || bf->l_start > inode->i_sb->s_maxbytes) {
 683		error = -EINVAL;
 684		goto out_unlock;
 685	}
 686
 687	if (bf->l_start > XFS_ISIZE(ip)) {
 688		error = xfs_alloc_file_space(ip, XFS_ISIZE(ip),
 689				bf->l_start - XFS_ISIZE(ip), 0);
 690		if (error)
 691			goto out_unlock;
 692	}
 693
 694	iattr.ia_valid = ATTR_SIZE;
 695	iattr.ia_size = bf->l_start;
 696	error = xfs_vn_setattr_size(file_dentry(filp), &iattr);
 697	if (error)
 698		goto out_unlock;
 699
 700	error = xfs_update_prealloc_flags(ip, flags);
 701
 702out_unlock:
 703	xfs_iunlock(ip, iolock);
 704	mnt_drop_write_file(filp);
 705	return error;
 706}
 707
 708/* Return 0 on success or positive error */
 709int
 710xfs_fsbulkstat_one_fmt(
 711	struct xfs_ibulk		*breq,
 712	const struct xfs_bulkstat	*bstat)
 713{
 714	struct xfs_bstat		bs1;
 715
 716	xfs_bulkstat_to_bstat(breq->mp, &bs1, bstat);
 717	if (copy_to_user(breq->ubuffer, &bs1, sizeof(bs1)))
 718		return -EFAULT;
 719	return xfs_ibulk_advance(breq, sizeof(struct xfs_bstat));
 720}
 721
 722int
 723xfs_fsinumbers_fmt(
 724	struct xfs_ibulk		*breq,
 725	const struct xfs_inumbers	*igrp)
 726{
 727	struct xfs_inogrp		ig1;
 728
 729	xfs_inumbers_to_inogrp(&ig1, igrp);
 730	if (copy_to_user(breq->ubuffer, &ig1, sizeof(struct xfs_inogrp)))
 731		return -EFAULT;
 732	return xfs_ibulk_advance(breq, sizeof(struct xfs_inogrp));
 733}
 734
 735STATIC int
 736xfs_ioc_fsbulkstat(
 737	xfs_mount_t		*mp,
 738	unsigned int		cmd,
 739	void			__user *arg)
 740{
 
 741	struct xfs_fsop_bulkreq	bulkreq;
 742	struct xfs_ibulk	breq = {
 743		.mp		= mp,
 
 744		.ocount		= 0,
 745	};
 746	xfs_ino_t		lastino;
 747	int			error;
 748
 749	/* done = 1 if there are more stats to get and if bulkstat */
 750	/* should be called again (unused here, but used in dmapi) */
 751
 752	if (!capable(CAP_SYS_ADMIN))
 753		return -EPERM;
 754
 755	if (XFS_FORCED_SHUTDOWN(mp))
 756		return -EIO;
 757
 758	if (copy_from_user(&bulkreq, arg, sizeof(struct xfs_fsop_bulkreq)))
 759		return -EFAULT;
 760
 761	if (copy_from_user(&lastino, bulkreq.lastip, sizeof(__s64)))
 762		return -EFAULT;
 763
 764	if (bulkreq.icount <= 0)
 765		return -EINVAL;
 766
 767	if (bulkreq.ubuffer == NULL)
 768		return -EINVAL;
 769
 770	breq.ubuffer = bulkreq.ubuffer;
 771	breq.icount = bulkreq.icount;
 772
 773	/*
 774	 * FSBULKSTAT_SINGLE expects that *lastip contains the inode number
 775	 * that we want to stat.  However, FSINUMBERS and FSBULKSTAT expect
 776	 * that *lastip contains either zero or the number of the last inode to
 777	 * be examined by the previous call and return results starting with
 778	 * the next inode after that.  The new bulk request back end functions
 779	 * take the inode to start with, so we have to compute the startino
 780	 * parameter from lastino to maintain correct function.  lastino == 0
 781	 * is a special case because it has traditionally meant "first inode
 782	 * in filesystem".
 783	 */
 784	if (cmd == XFS_IOC_FSINUMBERS) {
 785		breq.startino = lastino ? lastino + 1 : 0;
 786		error = xfs_inumbers(&breq, xfs_fsinumbers_fmt);
 787		lastino = breq.startino - 1;
 788	} else if (cmd == XFS_IOC_FSBULKSTAT_SINGLE) {
 789		breq.startino = lastino;
 790		breq.icount = 1;
 791		error = xfs_bulkstat_one(&breq, xfs_fsbulkstat_one_fmt);
 792	} else {	/* XFS_IOC_FSBULKSTAT */
 793		breq.startino = lastino ? lastino + 1 : 0;
 794		error = xfs_bulkstat(&breq, xfs_fsbulkstat_one_fmt);
 795		lastino = breq.startino - 1;
 796	}
 797
 798	if (error)
 799		return error;
 800
 801	if (bulkreq.lastip != NULL &&
 802	    copy_to_user(bulkreq.lastip, &lastino, sizeof(xfs_ino_t)))
 803		return -EFAULT;
 804
 805	if (bulkreq.ocount != NULL &&
 806	    copy_to_user(bulkreq.ocount, &breq.ocount, sizeof(__s32)))
 807		return -EFAULT;
 808
 809	return 0;
 810}
 811
 812/* Return 0 on success or positive error */
 813static int
 814xfs_bulkstat_fmt(
 815	struct xfs_ibulk		*breq,
 816	const struct xfs_bulkstat	*bstat)
 817{
 818	if (copy_to_user(breq->ubuffer, bstat, sizeof(struct xfs_bulkstat)))
 819		return -EFAULT;
 820	return xfs_ibulk_advance(breq, sizeof(struct xfs_bulkstat));
 821}
 822
 823/*
 824 * Check the incoming bulk request @hdr from userspace and initialize the
 825 * internal @breq bulk request appropriately.  Returns 0 if the bulk request
 826 * should proceed; -ECANCELED if there's nothing to do; or the usual
 827 * negative error code.
 828 */
 829static int
 830xfs_bulk_ireq_setup(
 831	struct xfs_mount	*mp,
 832	struct xfs_bulk_ireq	*hdr,
 833	struct xfs_ibulk	*breq,
 834	void __user		*ubuffer)
 835{
 836	if (hdr->icount == 0 ||
 837	    (hdr->flags & ~XFS_BULK_IREQ_FLAGS_ALL) ||
 838	    memchr_inv(hdr->reserved, 0, sizeof(hdr->reserved)))
 839		return -EINVAL;
 840
 841	breq->startino = hdr->ino;
 842	breq->ubuffer = ubuffer;
 843	breq->icount = hdr->icount;
 844	breq->ocount = 0;
 845	breq->flags = 0;
 846
 847	/*
 848	 * The @ino parameter is a special value, so we must look it up here.
 849	 * We're not allowed to have IREQ_AGNO, and we only return one inode
 850	 * worth of data.
 851	 */
 852	if (hdr->flags & XFS_BULK_IREQ_SPECIAL) {
 853		if (hdr->flags & XFS_BULK_IREQ_AGNO)
 854			return -EINVAL;
 855
 856		switch (hdr->ino) {
 857		case XFS_BULK_IREQ_SPECIAL_ROOT:
 858			hdr->ino = mp->m_sb.sb_rootino;
 859			break;
 860		default:
 861			return -EINVAL;
 862		}
 863		breq->icount = 1;
 864	}
 865
 866	/*
 867	 * The IREQ_AGNO flag means that we only want results from a given AG.
 868	 * If @hdr->ino is zero, we start iterating in that AG.  If @hdr->ino is
 869	 * beyond the specified AG then we return no results.
 870	 */
 871	if (hdr->flags & XFS_BULK_IREQ_AGNO) {
 872		if (hdr->agno >= mp->m_sb.sb_agcount)
 873			return -EINVAL;
 874
 875		if (breq->startino == 0)
 876			breq->startino = XFS_AGINO_TO_INO(mp, hdr->agno, 0);
 877		else if (XFS_INO_TO_AGNO(mp, breq->startino) < hdr->agno)
 878			return -EINVAL;
 879
 880		breq->flags |= XFS_IBULK_SAME_AG;
 881
 882		/* Asking for an inode past the end of the AG?  We're done! */
 883		if (XFS_INO_TO_AGNO(mp, breq->startino) > hdr->agno)
 884			return -ECANCELED;
 885	} else if (hdr->agno)
 886		return -EINVAL;
 887
 888	/* Asking for an inode past the end of the FS?  We're done! */
 889	if (XFS_INO_TO_AGNO(mp, breq->startino) >= mp->m_sb.sb_agcount)
 890		return -ECANCELED;
 891
 
 
 
 
 
 
 
 892	return 0;
 893}
 894
 895/*
 896 * Update the userspace bulk request @hdr to reflect the end state of the
 897 * internal bulk request @breq.
 898 */
 899static void
 900xfs_bulk_ireq_teardown(
 901	struct xfs_bulk_ireq	*hdr,
 902	struct xfs_ibulk	*breq)
 903{
 904	hdr->ino = breq->startino;
 905	hdr->ocount = breq->ocount;
 906}
 907
 908/* Handle the v5 bulkstat ioctl. */
 909STATIC int
 910xfs_ioc_bulkstat(
 911	struct xfs_mount		*mp,
 912	unsigned int			cmd,
 913	struct xfs_bulkstat_req __user	*arg)
 914{
 
 915	struct xfs_bulk_ireq		hdr;
 916	struct xfs_ibulk		breq = {
 917		.mp			= mp,
 
 918	};
 919	int				error;
 920
 921	if (!capable(CAP_SYS_ADMIN))
 922		return -EPERM;
 923
 924	if (XFS_FORCED_SHUTDOWN(mp))
 925		return -EIO;
 926
 927	if (copy_from_user(&hdr, &arg->hdr, sizeof(hdr)))
 928		return -EFAULT;
 929
 930	error = xfs_bulk_ireq_setup(mp, &hdr, &breq, arg->bulkstat);
 931	if (error == -ECANCELED)
 932		goto out_teardown;
 933	if (error < 0)
 934		return error;
 935
 936	error = xfs_bulkstat(&breq, xfs_bulkstat_fmt);
 937	if (error)
 938		return error;
 939
 940out_teardown:
 941	xfs_bulk_ireq_teardown(&hdr, &breq);
 942	if (copy_to_user(&arg->hdr, &hdr, sizeof(hdr)))
 943		return -EFAULT;
 944
 945	return 0;
 946}
 947
 948STATIC int
 949xfs_inumbers_fmt(
 950	struct xfs_ibulk		*breq,
 951	const struct xfs_inumbers	*igrp)
 952{
 953	if (copy_to_user(breq->ubuffer, igrp, sizeof(struct xfs_inumbers)))
 954		return -EFAULT;
 955	return xfs_ibulk_advance(breq, sizeof(struct xfs_inumbers));
 956}
 957
 958/* Handle the v5 inumbers ioctl. */
 959STATIC int
 960xfs_ioc_inumbers(
 961	struct xfs_mount		*mp,
 962	unsigned int			cmd,
 963	struct xfs_inumbers_req __user	*arg)
 964{
 965	struct xfs_bulk_ireq		hdr;
 966	struct xfs_ibulk		breq = {
 967		.mp			= mp,
 968	};
 969	int				error;
 970
 971	if (!capable(CAP_SYS_ADMIN))
 972		return -EPERM;
 973
 974	if (XFS_FORCED_SHUTDOWN(mp))
 975		return -EIO;
 976
 977	if (copy_from_user(&hdr, &arg->hdr, sizeof(hdr)))
 978		return -EFAULT;
 979
 
 
 
 980	error = xfs_bulk_ireq_setup(mp, &hdr, &breq, arg->inumbers);
 981	if (error == -ECANCELED)
 982		goto out_teardown;
 983	if (error < 0)
 984		return error;
 985
 986	error = xfs_inumbers(&breq, xfs_inumbers_fmt);
 987	if (error)
 988		return error;
 989
 990out_teardown:
 991	xfs_bulk_ireq_teardown(&hdr, &breq);
 992	if (copy_to_user(&arg->hdr, &hdr, sizeof(hdr)))
 993		return -EFAULT;
 994
 995	return 0;
 996}
 997
 998STATIC int
 999xfs_ioc_fsgeometry(
1000	struct xfs_mount	*mp,
1001	void			__user *arg,
1002	int			struct_version)
1003{
1004	struct xfs_fsop_geom	fsgeo;
1005	size_t			len;
1006
1007	xfs_fs_geometry(&mp->m_sb, &fsgeo, struct_version);
1008
1009	if (struct_version <= 3)
1010		len = sizeof(struct xfs_fsop_geom_v1);
1011	else if (struct_version == 4)
1012		len = sizeof(struct xfs_fsop_geom_v4);
1013	else {
1014		xfs_fsop_geom_health(mp, &fsgeo);
1015		len = sizeof(fsgeo);
1016	}
1017
1018	if (copy_to_user(arg, &fsgeo, len))
1019		return -EFAULT;
1020	return 0;
1021}
1022
1023STATIC int
1024xfs_ioc_ag_geometry(
1025	struct xfs_mount	*mp,
1026	void			__user *arg)
1027{
 
1028	struct xfs_ag_geometry	ageo;
1029	int			error;
1030
1031	if (copy_from_user(&ageo, arg, sizeof(ageo)))
1032		return -EFAULT;
1033	if (ageo.ag_flags)
1034		return -EINVAL;
1035	if (memchr_inv(&ageo.ag_reserved, 0, sizeof(ageo.ag_reserved)))
1036		return -EINVAL;
1037
1038	error = xfs_ag_get_geometry(mp, ageo.ag_number, &ageo);
 
 
 
 
 
1039	if (error)
1040		return error;
1041
1042	if (copy_to_user(arg, &ageo, sizeof(ageo)))
1043		return -EFAULT;
1044	return 0;
1045}
1046
1047/*
1048 * Linux extended inode flags interface.
1049 */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1050
1051STATIC unsigned int
1052xfs_merge_ioc_xflags(
1053	unsigned int	flags,
1054	unsigned int	start)
1055{
1056	unsigned int	xflags = start;
1057
1058	if (flags & FS_IMMUTABLE_FL)
1059		xflags |= FS_XFLAG_IMMUTABLE;
1060	else
1061		xflags &= ~FS_XFLAG_IMMUTABLE;
1062	if (flags & FS_APPEND_FL)
1063		xflags |= FS_XFLAG_APPEND;
1064	else
1065		xflags &= ~FS_XFLAG_APPEND;
1066	if (flags & FS_SYNC_FL)
1067		xflags |= FS_XFLAG_SYNC;
1068	else
1069		xflags &= ~FS_XFLAG_SYNC;
1070	if (flags & FS_NOATIME_FL)
1071		xflags |= FS_XFLAG_NOATIME;
1072	else
1073		xflags &= ~FS_XFLAG_NOATIME;
1074	if (flags & FS_NODUMP_FL)
1075		xflags |= FS_XFLAG_NODUMP;
1076	else
1077		xflags &= ~FS_XFLAG_NODUMP;
1078	if (flags & FS_DAX_FL)
1079		xflags |= FS_XFLAG_DAX;
1080	else
1081		xflags &= ~FS_XFLAG_DAX;
1082
1083	return xflags;
 
 
1084}
1085
1086STATIC unsigned int
1087xfs_di2lxflags(
1088	uint16_t	di_flags,
1089	uint64_t	di_flags2)
1090{
1091	unsigned int	flags = 0;
1092
1093	if (di_flags & XFS_DIFLAG_IMMUTABLE)
1094		flags |= FS_IMMUTABLE_FL;
1095	if (di_flags & XFS_DIFLAG_APPEND)
1096		flags |= FS_APPEND_FL;
1097	if (di_flags & XFS_DIFLAG_SYNC)
1098		flags |= FS_SYNC_FL;
1099	if (di_flags & XFS_DIFLAG_NOATIME)
1100		flags |= FS_NOATIME_FL;
1101	if (di_flags & XFS_DIFLAG_NODUMP)
1102		flags |= FS_NODUMP_FL;
1103	if (di_flags2 & XFS_DIFLAG2_DAX) {
1104		flags |= FS_DAX_FL;
1105	}
1106	return flags;
1107}
1108
1109static void
1110xfs_fill_fsxattr(
1111	struct xfs_inode	*ip,
1112	bool			attr,
1113	struct fsxattr		*fa)
1114{
1115	struct xfs_ifork	*ifp = attr ? ip->i_afp : &ip->i_df;
 
 
 
1116
1117	simple_fill_fsxattr(fa, xfs_ip2xflags(ip));
1118	fa->fsx_extsize = ip->i_d.di_extsize << ip->i_mount->m_sb.sb_blocklog;
1119	fa->fsx_cowextsize = ip->i_d.di_cowextsize <<
1120			ip->i_mount->m_sb.sb_blocklog;
1121	fa->fsx_projid = ip->i_d.di_projid;
1122	if (ifp && (ifp->if_flags & XFS_IFEXTENTS))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1123		fa->fsx_nextents = xfs_iext_count(ifp);
1124	else
1125		fa->fsx_nextents = xfs_ifork_nextents(ifp);
1126}
1127
1128STATIC int
1129xfs_ioc_fsgetxattr(
1130	xfs_inode_t		*ip,
1131	int			attr,
1132	void			__user *arg)
1133{
1134	struct fsxattr		fa;
1135
1136	xfs_ilock(ip, XFS_ILOCK_SHARED);
1137	xfs_fill_fsxattr(ip, attr, &fa);
1138	xfs_iunlock(ip, XFS_ILOCK_SHARED);
1139
1140	if (copy_to_user(arg, &fa, sizeof(fa)))
1141		return -EFAULT;
1142	return 0;
1143}
1144
1145STATIC uint16_t
1146xfs_flags2diflags(
1147	struct xfs_inode	*ip,
1148	unsigned int		xflags)
1149{
1150	/* can't set PREALLOC this way, just preserve it */
1151	uint16_t		di_flags =
1152		(ip->i_d.di_flags & XFS_DIFLAG_PREALLOC);
1153
1154	if (xflags & FS_XFLAG_IMMUTABLE)
1155		di_flags |= XFS_DIFLAG_IMMUTABLE;
1156	if (xflags & FS_XFLAG_APPEND)
1157		di_flags |= XFS_DIFLAG_APPEND;
1158	if (xflags & FS_XFLAG_SYNC)
1159		di_flags |= XFS_DIFLAG_SYNC;
1160	if (xflags & FS_XFLAG_NOATIME)
1161		di_flags |= XFS_DIFLAG_NOATIME;
1162	if (xflags & FS_XFLAG_NODUMP)
1163		di_flags |= XFS_DIFLAG_NODUMP;
1164	if (xflags & FS_XFLAG_NODEFRAG)
1165		di_flags |= XFS_DIFLAG_NODEFRAG;
1166	if (xflags & FS_XFLAG_FILESTREAM)
1167		di_flags |= XFS_DIFLAG_FILESTREAM;
1168	if (S_ISDIR(VFS_I(ip)->i_mode)) {
1169		if (xflags & FS_XFLAG_RTINHERIT)
1170			di_flags |= XFS_DIFLAG_RTINHERIT;
1171		if (xflags & FS_XFLAG_NOSYMLINKS)
1172			di_flags |= XFS_DIFLAG_NOSYMLINKS;
1173		if (xflags & FS_XFLAG_EXTSZINHERIT)
1174			di_flags |= XFS_DIFLAG_EXTSZINHERIT;
1175		if (xflags & FS_XFLAG_PROJINHERIT)
1176			di_flags |= XFS_DIFLAG_PROJINHERIT;
1177	} else if (S_ISREG(VFS_I(ip)->i_mode)) {
1178		if (xflags & FS_XFLAG_REALTIME)
1179			di_flags |= XFS_DIFLAG_REALTIME;
1180		if (xflags & FS_XFLAG_EXTSIZE)
1181			di_flags |= XFS_DIFLAG_EXTSIZE;
1182	}
1183
1184	return di_flags;
1185}
1186
1187STATIC uint64_t
1188xfs_flags2diflags2(
1189	struct xfs_inode	*ip,
1190	unsigned int		xflags)
1191{
1192	uint64_t		di_flags2 =
1193		(ip->i_d.di_flags2 & XFS_DIFLAG2_REFLINK);
1194
1195	if (xflags & FS_XFLAG_DAX)
1196		di_flags2 |= XFS_DIFLAG2_DAX;
1197	if (xflags & FS_XFLAG_COWEXTSIZE)
1198		di_flags2 |= XFS_DIFLAG2_COWEXTSIZE;
1199
1200	return di_flags2;
1201}
1202
1203static int
1204xfs_ioctl_setattr_xflags(
1205	struct xfs_trans	*tp,
1206	struct xfs_inode	*ip,
1207	struct fsxattr		*fa)
1208{
1209	struct xfs_mount	*mp = ip->i_mount;
1210	uint64_t		di_flags2;
 
1211
1212	/* Can't change realtime flag if any extents are allocated. */
1213	if ((ip->i_df.if_nextents || ip->i_delayed_blks) &&
1214	    XFS_IS_REALTIME_INODE(ip) != (fa->fsx_xflags & FS_XFLAG_REALTIME))
1215		return -EINVAL;
1216
1217	/* If realtime flag is set then must have realtime device */
1218	if (fa->fsx_xflags & FS_XFLAG_REALTIME) {
1219		if (mp->m_sb.sb_rblocks == 0 || mp->m_sb.sb_rextsize == 0 ||
1220		    (ip->i_d.di_extsize % mp->m_sb.sb_rextsize))
 
 
 
 
 
1221			return -EINVAL;
1222	}
1223
1224	/* Clear reflink if we are actually able to set the rt flag. */
1225	if ((fa->fsx_xflags & FS_XFLAG_REALTIME) && xfs_is_reflink_inode(ip))
1226		ip->i_d.di_flags2 &= ~XFS_DIFLAG2_REFLINK;
 
 
1227
1228	/* Don't allow us to set DAX mode for a reflinked file for now. */
1229	if ((fa->fsx_xflags & FS_XFLAG_DAX) && xfs_is_reflink_inode(ip))
1230		return -EINVAL;
 
1231
1232	/* diflags2 only valid for v3 inodes. */
1233	di_flags2 = xfs_flags2diflags2(ip, fa->fsx_xflags);
1234	if (di_flags2 && !xfs_sb_version_has_v3inode(&mp->m_sb))
1235		return -EINVAL;
1236
1237	ip->i_d.di_flags = xfs_flags2diflags(ip, fa->fsx_xflags);
1238	ip->i_d.di_flags2 = di_flags2;
1239
1240	xfs_diflags_to_iflags(ip, false);
 
 
 
 
 
 
 
 
1241	xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_CHG);
1242	xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
1243	XFS_STATS_INC(mp, xs_ig_attrchg);
1244	return 0;
1245}
1246
1247static void
1248xfs_ioctl_setattr_prepare_dax(
1249	struct xfs_inode	*ip,
1250	struct fsxattr		*fa)
1251{
1252	struct xfs_mount	*mp = ip->i_mount;
1253	struct inode            *inode = VFS_I(ip);
1254
1255	if (S_ISDIR(inode->i_mode))
1256		return;
1257
1258	if ((mp->m_flags & XFS_MOUNT_DAX_ALWAYS) ||
1259	    (mp->m_flags & XFS_MOUNT_DAX_NEVER))
1260		return;
1261
1262	if (((fa->fsx_xflags & FS_XFLAG_DAX) &&
1263	    !(ip->i_d.di_flags2 & XFS_DIFLAG2_DAX)) ||
1264	    (!(fa->fsx_xflags & FS_XFLAG_DAX) &&
1265	     (ip->i_d.di_flags2 & XFS_DIFLAG2_DAX)))
1266		d_mark_dontcache(inode);
1267}
1268
1269/*
1270 * Set up the transaction structure for the setattr operation, checking that we
1271 * have permission to do so. On success, return a clean transaction and the
1272 * inode locked exclusively ready for further operation specific checks. On
1273 * failure, return an error without modifying or locking the inode.
1274 */
1275static struct xfs_trans *
1276xfs_ioctl_setattr_get_trans(
1277	struct xfs_inode	*ip)
 
1278{
1279	struct xfs_mount	*mp = ip->i_mount;
1280	struct xfs_trans	*tp;
1281	int			error = -EROFS;
1282
1283	if (mp->m_flags & XFS_MOUNT_RDONLY)
1284		goto out_unlock;
1285	error = -EIO;
1286	if (XFS_FORCED_SHUTDOWN(mp))
1287		goto out_unlock;
1288
1289	error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp);
 
1290	if (error)
1291		goto out_unlock;
1292
1293	xfs_ilock(ip, XFS_ILOCK_EXCL);
1294	xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
1295
1296	/*
1297	 * CAP_FOWNER overrides the following restrictions:
1298	 *
1299	 * The user ID of the calling process must be equal to the file owner
1300	 * ID, except in cases where the CAP_FSETID capability is applicable.
1301	 */
1302	if (!inode_owner_or_capable(VFS_I(ip))) {
1303		error = -EPERM;
1304		goto out_cancel;
1305	}
1306
1307	if (mp->m_flags & XFS_MOUNT_WSYNC)
1308		xfs_trans_set_sync(tp);
1309
1310	return tp;
1311
1312out_cancel:
1313	xfs_trans_cancel(tp);
1314out_unlock:
1315	return ERR_PTR(error);
1316}
1317
1318/*
1319 * extent size hint validation is somewhat cumbersome. Rules are:
1320 *
1321 * 1. extent size hint is only valid for directories and regular files
1322 * 2. FS_XFLAG_EXTSIZE is only valid for regular files
1323 * 3. FS_XFLAG_EXTSZINHERIT is only valid for directories.
1324 * 4. can only be changed on regular files if no extents are allocated
1325 * 5. can be changed on directories at any time
1326 * 6. extsize hint of 0 turns off hints, clears inode flags.
1327 * 7. Extent size must be a multiple of the appropriate block size.
1328 * 8. for non-realtime files, the extent size hint must be limited
1329 *    to half the AG size to avoid alignment extending the extent beyond the
1330 *    limits of the AG.
1331 *
1332 * Please keep this function in sync with xfs_scrub_inode_extsize.
1333 */
1334static int
1335xfs_ioctl_setattr_check_extsize(
1336	struct xfs_inode	*ip,
1337	struct fsxattr		*fa)
1338{
1339	struct xfs_mount	*mp = ip->i_mount;
1340	xfs_extlen_t		size;
1341	xfs_fsblock_t		extsize_fsb;
1342
1343	if (S_ISREG(VFS_I(ip)->i_mode) && ip->i_df.if_nextents &&
1344	    ((ip->i_d.di_extsize << mp->m_sb.sb_blocklog) != fa->fsx_extsize))
 
 
 
1345		return -EINVAL;
1346
1347	if (fa->fsx_extsize == 0)
1348		return 0;
 
 
1349
1350	extsize_fsb = XFS_B_TO_FSB(mp, fa->fsx_extsize);
1351	if (extsize_fsb > MAXEXTLEN)
1352		return -EINVAL;
 
 
 
 
 
 
1353
1354	if (XFS_IS_REALTIME_INODE(ip) ||
1355	    (fa->fsx_xflags & FS_XFLAG_REALTIME)) {
1356		size = mp->m_sb.sb_rextsize << mp->m_sb.sb_blocklog;
1357	} else {
1358		size = mp->m_sb.sb_blocksize;
1359		if (extsize_fsb > mp->m_sb.sb_agblocks / 2)
1360			return -EINVAL;
1361	}
1362
1363	if (fa->fsx_extsize % size)
1364		return -EINVAL;
1365
1366	return 0;
1367}
1368
1369/*
1370 * CoW extent size hint validation rules are:
1371 *
1372 * 1. CoW extent size hint can only be set if reflink is enabled on the fs.
1373 *    The inode does not have to have any shared blocks, but it must be a v3.
1374 * 2. FS_XFLAG_COWEXTSIZE is only valid for directories and regular files;
1375 *    for a directory, the hint is propagated to new files.
1376 * 3. Can be changed on files & directories at any time.
1377 * 4. CoW extsize hint of 0 turns off hints, clears inode flags.
1378 * 5. Extent size must be a multiple of the appropriate block size.
1379 * 6. The extent size hint must be limited to half the AG size to avoid
1380 *    alignment extending the extent beyond the limits of the AG.
1381 *
1382 * Please keep this function in sync with xfs_scrub_inode_cowextsize.
1383 */
1384static int
1385xfs_ioctl_setattr_check_cowextsize(
1386	struct xfs_inode	*ip,
1387	struct fsxattr		*fa)
1388{
1389	struct xfs_mount	*mp = ip->i_mount;
1390	xfs_extlen_t		size;
1391	xfs_fsblock_t		cowextsize_fsb;
1392
1393	if (!(fa->fsx_xflags & FS_XFLAG_COWEXTSIZE))
1394		return 0;
1395
1396	if (!xfs_sb_version_hasreflink(&ip->i_mount->m_sb))
1397		return -EINVAL;
1398
1399	if (fa->fsx_cowextsize == 0)
1400		return 0;
1401
1402	cowextsize_fsb = XFS_B_TO_FSB(mp, fa->fsx_cowextsize);
1403	if (cowextsize_fsb > MAXEXTLEN)
1404		return -EINVAL;
1405
1406	size = mp->m_sb.sb_blocksize;
1407	if (cowextsize_fsb > mp->m_sb.sb_agblocks / 2)
1408		return -EINVAL;
1409
1410	if (fa->fsx_cowextsize % size)
1411		return -EINVAL;
1412
1413	return 0;
1414}
1415
1416static int
1417xfs_ioctl_setattr_check_projid(
1418	struct xfs_inode	*ip,
1419	struct fsxattr		*fa)
1420{
1421	/* Disallow 32bit project ids if projid32bit feature is not enabled. */
 
 
 
1422	if (fa->fsx_projid > (uint16_t)-1 &&
1423	    !xfs_sb_version_hasprojid32bit(&ip->i_mount->m_sb))
1424		return -EINVAL;
1425	return 0;
1426}
1427
1428STATIC int
1429xfs_ioctl_setattr(
1430	xfs_inode_t		*ip,
1431	struct fsxattr		*fa)
 
1432{
1433	struct fsxattr		old_fa;
1434	struct xfs_mount	*mp = ip->i_mount;
1435	struct xfs_trans	*tp;
1436	struct xfs_dquot	*pdqp = NULL;
1437	struct xfs_dquot	*olddquot = NULL;
1438	int			code;
1439
1440	trace_xfs_ioctl_setattr(ip);
1441
1442	code = xfs_ioctl_setattr_check_projid(ip, fa);
1443	if (code)
1444		return code;
 
 
 
 
 
 
 
 
 
 
1445
1446	/*
1447	 * If disk quotas is on, we make sure that the dquots do exist on disk,
1448	 * before we start any other transactions. Trying to do this later
1449	 * is messy. We don't care to take a readlock to look at the ids
1450	 * in inode here, because we can't hold it across the trans_reserve.
1451	 * If the IDs do change before we take the ilock, we're covered
1452	 * because the i_*dquot fields will get updated anyway.
1453	 */
1454	if (XFS_IS_QUOTA_ON(mp)) {
1455		code = xfs_qm_vop_dqalloc(ip, VFS_I(ip)->i_uid,
1456				VFS_I(ip)->i_gid, fa->fsx_projid,
1457				XFS_QMOPT_PQUOTA, NULL, NULL, &pdqp);
1458		if (code)
1459			return code;
1460	}
1461
1462	xfs_ioctl_setattr_prepare_dax(ip, fa);
1463
1464	tp = xfs_ioctl_setattr_get_trans(ip);
1465	if (IS_ERR(tp)) {
1466		code = PTR_ERR(tp);
1467		goto error_free_dquots;
1468	}
1469
1470	if (XFS_IS_QUOTA_RUNNING(mp) && XFS_IS_PQUOTA_ON(mp) &&
1471	    ip->i_d.di_projid != fa->fsx_projid) {
1472		code = xfs_qm_vop_chown_reserve(tp, ip, NULL, NULL, pdqp,
1473				capable(CAP_FOWNER) ?  XFS_QMOPT_FORCE_RES : 0);
1474		if (code)	/* out of quota */
1475			goto error_trans_cancel;
1476	}
1477
1478	xfs_fill_fsxattr(ip, false, &old_fa);
1479	code = vfs_ioc_fssetxattr_check(VFS_I(ip), &old_fa, fa);
1480	if (code)
1481		goto error_trans_cancel;
1482
1483	code = xfs_ioctl_setattr_check_extsize(ip, fa);
1484	if (code)
1485		goto error_trans_cancel;
1486
1487	code = xfs_ioctl_setattr_check_cowextsize(ip, fa);
1488	if (code)
1489		goto error_trans_cancel;
1490
1491	code = xfs_ioctl_setattr_xflags(tp, ip, fa);
1492	if (code)
1493		goto error_trans_cancel;
1494
 
 
1495	/*
1496	 * Change file ownership.  Must be the owner or privileged.  CAP_FSETID
1497	 * overrides the following restrictions:
1498	 *
1499	 * The set-user-ID and set-group-ID bits of a file will be cleared upon
1500	 * successful return from chown()
1501	 */
1502
1503	if ((VFS_I(ip)->i_mode & (S_ISUID|S_ISGID)) &&
1504	    !capable_wrt_inode_uidgid(VFS_I(ip), CAP_FSETID))
1505		VFS_I(ip)->i_mode &= ~(S_ISUID|S_ISGID);
1506
1507	/* Change the ownerships and register project quota modifications */
1508	if (ip->i_d.di_projid != fa->fsx_projid) {
1509		if (XFS_IS_QUOTA_RUNNING(mp) && XFS_IS_PQUOTA_ON(mp)) {
1510			olddquot = xfs_qm_vop_chown(tp, ip,
1511						&ip->i_pdquot, pdqp);
1512		}
1513		ip->i_d.di_projid = fa->fsx_projid;
1514	}
1515
1516	/*
1517	 * Only set the extent size hint if we've already determined that the
1518	 * extent size hint should be set on the inode. If no extent size flags
1519	 * are set on the inode then unconditionally clear the extent size hint.
1520	 */
1521	if (ip->i_d.di_flags & (XFS_DIFLAG_EXTSIZE | XFS_DIFLAG_EXTSZINHERIT))
1522		ip->i_d.di_extsize = fa->fsx_extsize >> mp->m_sb.sb_blocklog;
1523	else
1524		ip->i_d.di_extsize = 0;
1525	if (xfs_sb_version_has_v3inode(&mp->m_sb) &&
1526	    (ip->i_d.di_flags2 & XFS_DIFLAG2_COWEXTSIZE))
1527		ip->i_d.di_cowextsize = fa->fsx_cowextsize >>
1528				mp->m_sb.sb_blocklog;
1529	else
1530		ip->i_d.di_cowextsize = 0;
 
1531
1532	code = xfs_trans_commit(tp);
 
1533
1534	/*
1535	 * Release any dquot(s) the inode had kept before chown.
1536	 */
1537	xfs_qm_dqrele(olddquot);
1538	xfs_qm_dqrele(pdqp);
1539
1540	return code;
1541
1542error_trans_cancel:
1543	xfs_trans_cancel(tp);
1544error_free_dquots:
1545	xfs_qm_dqrele(pdqp);
1546	return code;
1547}
1548
1549STATIC int
1550xfs_ioc_fssetxattr(
1551	xfs_inode_t		*ip,
1552	struct file		*filp,
1553	void			__user *arg)
1554{
1555	struct fsxattr		fa;
1556	int error;
1557
1558	if (copy_from_user(&fa, arg, sizeof(fa)))
1559		return -EFAULT;
1560
1561	error = mnt_want_write_file(filp);
1562	if (error)
1563		return error;
1564	error = xfs_ioctl_setattr(ip, &fa);
1565	mnt_drop_write_file(filp);
1566	return error;
1567}
1568
1569STATIC int
1570xfs_ioc_getxflags(
1571	xfs_inode_t		*ip,
1572	void			__user *arg)
1573{
1574	unsigned int		flags;
1575
1576	flags = xfs_di2lxflags(ip->i_d.di_flags, ip->i_d.di_flags2);
1577	if (copy_to_user(arg, &flags, sizeof(flags)))
1578		return -EFAULT;
1579	return 0;
1580}
1581
1582STATIC int
1583xfs_ioc_setxflags(
1584	struct xfs_inode	*ip,
1585	struct file		*filp,
1586	void			__user *arg)
1587{
1588	struct xfs_trans	*tp;
1589	struct fsxattr		fa;
1590	struct fsxattr		old_fa;
1591	unsigned int		flags;
1592	int			error;
1593
1594	if (copy_from_user(&flags, arg, sizeof(flags)))
1595		return -EFAULT;
1596
1597	if (flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL | \
1598		      FS_NOATIME_FL | FS_NODUMP_FL | \
1599		      FS_SYNC_FL | FS_DAX_FL))
1600		return -EOPNOTSUPP;
1601
1602	fa.fsx_xflags = xfs_merge_ioc_xflags(flags, xfs_ip2xflags(ip));
1603
1604	error = mnt_want_write_file(filp);
1605	if (error)
1606		return error;
1607
1608	xfs_ioctl_setattr_prepare_dax(ip, &fa);
1609
1610	tp = xfs_ioctl_setattr_get_trans(ip);
1611	if (IS_ERR(tp)) {
1612		error = PTR_ERR(tp);
1613		goto out_drop_write;
1614	}
1615
1616	xfs_fill_fsxattr(ip, false, &old_fa);
1617	error = vfs_ioc_fssetxattr_check(VFS_I(ip), &old_fa, &fa);
1618	if (error) {
1619		xfs_trans_cancel(tp);
1620		goto out_drop_write;
1621	}
1622
1623	error = xfs_ioctl_setattr_xflags(tp, ip, &fa);
1624	if (error) {
1625		xfs_trans_cancel(tp);
1626		goto out_drop_write;
1627	}
1628
1629	error = xfs_trans_commit(tp);
1630out_drop_write:
1631	mnt_drop_write_file(filp);
1632	return error;
1633}
1634
1635static bool
1636xfs_getbmap_format(
1637	struct kgetbmap		*p,
1638	struct getbmapx __user	*u,
1639	size_t			recsize)
1640{
1641	if (put_user(p->bmv_offset, &u->bmv_offset) ||
1642	    put_user(p->bmv_block, &u->bmv_block) ||
1643	    put_user(p->bmv_length, &u->bmv_length) ||
1644	    put_user(0, &u->bmv_count) ||
1645	    put_user(0, &u->bmv_entries))
1646		return false;
1647	if (recsize < sizeof(struct getbmapx))
1648		return true;
1649	if (put_user(0, &u->bmv_iflags) ||
1650	    put_user(p->bmv_oflags, &u->bmv_oflags) ||
1651	    put_user(0, &u->bmv_unused1) ||
1652	    put_user(0, &u->bmv_unused2))
1653		return false;
1654	return true;
1655}
1656
1657STATIC int
1658xfs_ioc_getbmap(
1659	struct file		*file,
1660	unsigned int		cmd,
1661	void			__user *arg)
1662{
1663	struct getbmapx		bmx = { 0 };
1664	struct kgetbmap		*buf;
1665	size_t			recsize;
1666	int			error, i;
1667
1668	switch (cmd) {
1669	case XFS_IOC_GETBMAPA:
1670		bmx.bmv_iflags = BMV_IF_ATTRFORK;
1671		/*FALLTHRU*/
1672	case XFS_IOC_GETBMAP:
1673		if (file->f_mode & FMODE_NOCMTIME)
1674			bmx.bmv_iflags |= BMV_IF_NO_DMAPI_READ;
1675		/* struct getbmap is a strict subset of struct getbmapx. */
1676		recsize = sizeof(struct getbmap);
1677		break;
1678	case XFS_IOC_GETBMAPX:
1679		recsize = sizeof(struct getbmapx);
1680		break;
1681	default:
1682		return -EINVAL;
1683	}
1684
1685	if (copy_from_user(&bmx, arg, recsize))
1686		return -EFAULT;
1687
1688	if (bmx.bmv_count < 2)
1689		return -EINVAL;
1690	if (bmx.bmv_count > ULONG_MAX / recsize)
1691		return -ENOMEM;
1692
1693	buf = kmem_zalloc_large(bmx.bmv_count * sizeof(*buf), 0);
1694	if (!buf)
1695		return -ENOMEM;
1696
1697	error = xfs_getbmap(XFS_I(file_inode(file)), &bmx, buf);
1698	if (error)
1699		goto out_free_buf;
1700
1701	error = -EFAULT;
1702	if (copy_to_user(arg, &bmx, recsize))
1703		goto out_free_buf;
1704	arg += recsize;
1705
1706	for (i = 0; i < bmx.bmv_entries; i++) {
1707		if (!xfs_getbmap_format(buf + i, arg, recsize))
1708			goto out_free_buf;
1709		arg += recsize;
1710	}
1711
1712	error = 0;
1713out_free_buf:
1714	kmem_free(buf);
1715	return error;
1716}
1717
1718struct getfsmap_info {
1719	struct xfs_mount	*mp;
1720	struct fsmap_head __user *data;
1721	unsigned int		idx;
1722	__u32			last_flags;
1723};
1724
1725STATIC int
1726xfs_getfsmap_format(struct xfs_fsmap *xfm, void *priv)
1727{
1728	struct getfsmap_info	*info = priv;
1729	struct fsmap		fm;
1730
1731	trace_xfs_getfsmap_mapping(info->mp, xfm);
1732
1733	info->last_flags = xfm->fmr_flags;
1734	xfs_fsmap_from_internal(&fm, xfm);
1735	if (copy_to_user(&info->data->fmh_recs[info->idx++], &fm,
1736			sizeof(struct fsmap)))
1737		return -EFAULT;
1738
1739	return 0;
1740}
1741
1742STATIC int
1743xfs_ioc_getfsmap(
1744	struct xfs_inode	*ip,
1745	struct fsmap_head	__user *arg)
1746{
1747	struct getfsmap_info	info = { NULL };
1748	struct xfs_fsmap_head	xhead = {0};
1749	struct fsmap_head	head;
1750	bool			aborted = false;
1751	int			error;
1752
1753	if (copy_from_user(&head, arg, sizeof(struct fsmap_head)))
1754		return -EFAULT;
1755	if (memchr_inv(head.fmh_reserved, 0, sizeof(head.fmh_reserved)) ||
1756	    memchr_inv(head.fmh_keys[0].fmr_reserved, 0,
1757		       sizeof(head.fmh_keys[0].fmr_reserved)) ||
1758	    memchr_inv(head.fmh_keys[1].fmr_reserved, 0,
1759		       sizeof(head.fmh_keys[1].fmr_reserved)))
1760		return -EINVAL;
1761
1762	xhead.fmh_iflags = head.fmh_iflags;
1763	xhead.fmh_count = head.fmh_count;
1764	xfs_fsmap_to_internal(&xhead.fmh_keys[0], &head.fmh_keys[0]);
1765	xfs_fsmap_to_internal(&xhead.fmh_keys[1], &head.fmh_keys[1]);
1766
1767	trace_xfs_getfsmap_low_key(ip->i_mount, &xhead.fmh_keys[0]);
1768	trace_xfs_getfsmap_high_key(ip->i_mount, &xhead.fmh_keys[1]);
1769
1770	info.mp = ip->i_mount;
1771	info.data = arg;
1772	error = xfs_getfsmap(ip->i_mount, &xhead, xfs_getfsmap_format, &info);
1773	if (error == -ECANCELED) {
1774		error = 0;
1775		aborted = true;
1776	} else if (error)
1777		return error;
1778
1779	/* If we didn't abort, set the "last" flag in the last fmx */
1780	if (!aborted && info.idx) {
1781		info.last_flags |= FMR_OF_LAST;
1782		if (copy_to_user(&info.data->fmh_recs[info.idx - 1].fmr_flags,
1783				&info.last_flags, sizeof(info.last_flags)))
1784			return -EFAULT;
1785	}
1786
1787	/* copy back header */
1788	head.fmh_entries = xhead.fmh_entries;
1789	head.fmh_oflags = xhead.fmh_oflags;
1790	if (copy_to_user(arg, &head, sizeof(struct fsmap_head)))
1791		return -EFAULT;
1792
1793	return 0;
1794}
1795
1796STATIC int
1797xfs_ioc_scrub_metadata(
1798	struct xfs_inode		*ip,
1799	void				__user *arg)
1800{
1801	struct xfs_scrub_metadata	scrub;
1802	int				error;
1803
1804	if (!capable(CAP_SYS_ADMIN))
1805		return -EPERM;
1806
1807	if (copy_from_user(&scrub, arg, sizeof(scrub)))
1808		return -EFAULT;
1809
1810	error = xfs_scrub_metadata(ip, &scrub);
1811	if (error)
1812		return error;
1813
1814	if (copy_to_user(arg, &scrub, sizeof(scrub)))
1815		return -EFAULT;
1816
1817	return 0;
1818}
1819
1820int
1821xfs_ioc_swapext(
1822	xfs_swapext_t	*sxp)
1823{
1824	xfs_inode_t     *ip, *tip;
1825	struct fd	f, tmp;
1826	int		error = 0;
1827
1828	/* Pull information for the target fd */
1829	f = fdget((int)sxp->sx_fdtarget);
1830	if (!f.file) {
1831		error = -EINVAL;
1832		goto out;
1833	}
 
 
 
 
 
 
 
 
 
 
 
 
1834
1835	if (!(f.file->f_mode & FMODE_WRITE) ||
1836	    !(f.file->f_mode & FMODE_READ) ||
1837	    (f.file->f_flags & O_APPEND)) {
1838		error = -EBADF;
1839		goto out_put_file;
1840	}
1841
1842	tmp = fdget((int)sxp->sx_fdtmp);
1843	if (!tmp.file) {
1844		error = -EINVAL;
1845		goto out_put_file;
1846	}
1847
1848	if (!(tmp.file->f_mode & FMODE_WRITE) ||
1849	    !(tmp.file->f_mode & FMODE_READ) ||
1850	    (tmp.file->f_flags & O_APPEND)) {
1851		error = -EBADF;
1852		goto out_put_tmp_file;
1853	}
1854
1855	if (IS_SWAPFILE(file_inode(f.file)) ||
1856	    IS_SWAPFILE(file_inode(tmp.file))) {
1857		error = -EINVAL;
1858		goto out_put_tmp_file;
1859	}
1860
1861	/*
1862	 * We need to ensure that the fds passed in point to XFS inodes
1863	 * before we cast and access them as XFS structures as we have no
1864	 * control over what the user passes us here.
1865	 */
1866	if (f.file->f_op != &xfs_file_operations ||
1867	    tmp.file->f_op != &xfs_file_operations) {
1868		error = -EINVAL;
1869		goto out_put_tmp_file;
1870	}
1871
1872	ip = XFS_I(file_inode(f.file));
1873	tip = XFS_I(file_inode(tmp.file));
1874
1875	if (ip->i_mount != tip->i_mount) {
1876		error = -EINVAL;
1877		goto out_put_tmp_file;
1878	}
1879
1880	if (ip->i_ino == tip->i_ino) {
1881		error = -EINVAL;
1882		goto out_put_tmp_file;
1883	}
1884
1885	if (XFS_FORCED_SHUTDOWN(ip->i_mount)) {
1886		error = -EIO;
1887		goto out_put_tmp_file;
1888	}
1889
1890	error = xfs_swap_extents(ip, tip, sxp);
1891
1892 out_put_tmp_file:
1893	fdput(tmp);
1894 out_put_file:
1895	fdput(f);
1896 out:
1897	return error;
1898}
1899
1900static int
1901xfs_ioc_getlabel(
1902	struct xfs_mount	*mp,
1903	char			__user *user_label)
1904{
1905	struct xfs_sb		*sbp = &mp->m_sb;
1906	char			label[XFSLABEL_MAX + 1];
1907
1908	/* Paranoia */
1909	BUILD_BUG_ON(sizeof(sbp->sb_fname) > FSLABEL_MAX);
1910
1911	/* 1 larger than sb_fname, so this ensures a trailing NUL char */
1912	memset(label, 0, sizeof(label));
1913	spin_lock(&mp->m_sb_lock);
1914	strncpy(label, sbp->sb_fname, XFSLABEL_MAX);
1915	spin_unlock(&mp->m_sb_lock);
1916
1917	if (copy_to_user(user_label, label, sizeof(label)))
1918		return -EFAULT;
1919	return 0;
1920}
1921
1922static int
1923xfs_ioc_setlabel(
1924	struct file		*filp,
1925	struct xfs_mount	*mp,
1926	char			__user *newlabel)
1927{
1928	struct xfs_sb		*sbp = &mp->m_sb;
1929	char			label[XFSLABEL_MAX + 1];
1930	size_t			len;
1931	int			error;
1932
1933	if (!capable(CAP_SYS_ADMIN))
1934		return -EPERM;
1935	/*
1936	 * The generic ioctl allows up to FSLABEL_MAX chars, but XFS is much
1937	 * smaller, at 12 bytes.  We copy one more to be sure we find the
1938	 * (required) NULL character to test the incoming label length.
1939	 * NB: The on disk label doesn't need to be null terminated.
1940	 */
1941	if (copy_from_user(label, newlabel, XFSLABEL_MAX + 1))
1942		return -EFAULT;
1943	len = strnlen(label, XFSLABEL_MAX + 1);
1944	if (len > sizeof(sbp->sb_fname))
1945		return -EINVAL;
1946
1947	error = mnt_want_write_file(filp);
1948	if (error)
1949		return error;
1950
1951	spin_lock(&mp->m_sb_lock);
1952	memset(sbp->sb_fname, 0, sizeof(sbp->sb_fname));
1953	memcpy(sbp->sb_fname, label, len);
1954	spin_unlock(&mp->m_sb_lock);
1955
1956	/*
1957	 * Now we do several things to satisfy userspace.
1958	 * In addition to normal logging of the primary superblock, we also
1959	 * immediately write these changes to sector zero for the primary, then
1960	 * update all backup supers (as xfs_db does for a label change), then
1961	 * invalidate the block device page cache.  This is so that any prior
1962	 * buffered reads from userspace (i.e. from blkid) are invalidated,
1963	 * and userspace will see the newly-written label.
1964	 */
1965	error = xfs_sync_sb_buf(mp);
1966	if (error)
1967		goto out;
1968	/*
1969	 * growfs also updates backup supers so lock against that.
1970	 */
1971	mutex_lock(&mp->m_growlock);
1972	error = xfs_update_secondary_sbs(mp);
1973	mutex_unlock(&mp->m_growlock);
1974
1975	invalidate_bdev(mp->m_ddev_targp->bt_bdev);
 
 
1976
1977out:
1978	mnt_drop_write_file(filp);
1979	return error;
1980}
1981
1982static inline int
1983xfs_fs_eofblocks_from_user(
1984	struct xfs_fs_eofblocks		*src,
1985	struct xfs_eofblocks		*dst)
1986{
1987	if (src->eof_version != XFS_EOFBLOCKS_VERSION)
1988		return -EINVAL;
1989
1990	if (src->eof_flags & ~XFS_EOF_FLAGS_VALID)
1991		return -EINVAL;
1992
1993	if (memchr_inv(&src->pad32, 0, sizeof(src->pad32)) ||
1994	    memchr_inv(src->pad64, 0, sizeof(src->pad64)))
1995		return -EINVAL;
1996
1997	dst->eof_flags = src->eof_flags;
1998	dst->eof_prid = src->eof_prid;
1999	dst->eof_min_file_size = src->eof_min_file_size;
 
 
 
 
 
 
 
 
2000
2001	dst->eof_uid = INVALID_UID;
 
 
 
2002	if (src->eof_flags & XFS_EOF_FLAGS_UID) {
2003		dst->eof_uid = make_kuid(current_user_ns(), src->eof_uid);
2004		if (!uid_valid(dst->eof_uid))
2005			return -EINVAL;
2006	}
2007
2008	dst->eof_gid = INVALID_GID;
2009	if (src->eof_flags & XFS_EOF_FLAGS_GID) {
2010		dst->eof_gid = make_kgid(current_user_ns(), src->eof_gid);
2011		if (!gid_valid(dst->eof_gid))
2012			return -EINVAL;
2013	}
2014	return 0;
2015}
2016
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2017/*
2018 * Note: some of the ioctl's return positive numbers as a
2019 * byte count indicating success, such as readlink_by_handle.
2020 * So we don't "sign flip" like most other routines.  This means
2021 * true errors need to be returned as a negative value.
2022 */
2023long
2024xfs_file_ioctl(
2025	struct file		*filp,
2026	unsigned int		cmd,
2027	unsigned long		p)
2028{
2029	struct inode		*inode = file_inode(filp);
2030	struct xfs_inode	*ip = XFS_I(inode);
2031	struct xfs_mount	*mp = ip->i_mount;
2032	void			__user *arg = (void __user *)p;
2033	int			error;
2034
2035	trace_xfs_file_ioctl(ip);
2036
2037	switch (cmd) {
2038	case FITRIM:
2039		return xfs_ioc_trim(mp, arg);
2040	case FS_IOC_GETFSLABEL:
2041		return xfs_ioc_getlabel(mp, arg);
2042	case FS_IOC_SETFSLABEL:
2043		return xfs_ioc_setlabel(filp, mp, arg);
2044	case XFS_IOC_ALLOCSP:
2045	case XFS_IOC_FREESP:
2046	case XFS_IOC_ALLOCSP64:
2047	case XFS_IOC_FREESP64: {
2048		xfs_flock64_t		bf;
2049
2050		if (copy_from_user(&bf, arg, sizeof(bf)))
2051			return -EFAULT;
2052		return xfs_ioc_space(filp, &bf);
2053	}
2054	case XFS_IOC_DIOINFO: {
2055		struct xfs_buftarg	*target = xfs_inode_buftarg(ip);
2056		struct dioattr		da;
2057
2058		da.d_mem =  da.d_miniosz = target->bt_logical_sectorsize;
2059		da.d_maxiosz = INT_MAX & ~(da.d_miniosz - 1);
2060
2061		if (copy_to_user(arg, &da, sizeof(da)))
2062			return -EFAULT;
2063		return 0;
2064	}
2065
2066	case XFS_IOC_FSBULKSTAT_SINGLE:
2067	case XFS_IOC_FSBULKSTAT:
2068	case XFS_IOC_FSINUMBERS:
2069		return xfs_ioc_fsbulkstat(mp, cmd, arg);
2070
2071	case XFS_IOC_BULKSTAT:
2072		return xfs_ioc_bulkstat(mp, cmd, arg);
2073	case XFS_IOC_INUMBERS:
2074		return xfs_ioc_inumbers(mp, cmd, arg);
2075
2076	case XFS_IOC_FSGEOMETRY_V1:
2077		return xfs_ioc_fsgeometry(mp, arg, 3);
2078	case XFS_IOC_FSGEOMETRY_V4:
2079		return xfs_ioc_fsgeometry(mp, arg, 4);
2080	case XFS_IOC_FSGEOMETRY:
2081		return xfs_ioc_fsgeometry(mp, arg, 5);
2082
2083	case XFS_IOC_AG_GEOMETRY:
2084		return xfs_ioc_ag_geometry(mp, arg);
 
 
2085
2086	case XFS_IOC_GETVERSION:
2087		return put_user(inode->i_generation, (int __user *)arg);
2088
2089	case XFS_IOC_FSGETXATTR:
2090		return xfs_ioc_fsgetxattr(ip, 0, arg);
2091	case XFS_IOC_FSGETXATTRA:
2092		return xfs_ioc_fsgetxattr(ip, 1, arg);
2093	case XFS_IOC_FSSETXATTR:
2094		return xfs_ioc_fssetxattr(ip, filp, arg);
2095	case XFS_IOC_GETXFLAGS:
2096		return xfs_ioc_getxflags(ip, arg);
2097	case XFS_IOC_SETXFLAGS:
2098		return xfs_ioc_setxflags(ip, filp, arg);
2099
2100	case XFS_IOC_GETBMAP:
2101	case XFS_IOC_GETBMAPA:
2102	case XFS_IOC_GETBMAPX:
2103		return xfs_ioc_getbmap(filp, cmd, arg);
2104
2105	case FS_IOC_GETFSMAP:
2106		return xfs_ioc_getfsmap(ip, arg);
2107
 
 
2108	case XFS_IOC_SCRUB_METADATA:
2109		return xfs_ioc_scrub_metadata(ip, arg);
2110
2111	case XFS_IOC_FD_TO_HANDLE:
2112	case XFS_IOC_PATH_TO_HANDLE:
2113	case XFS_IOC_PATH_TO_FSHANDLE: {
2114		xfs_fsop_handlereq_t	hreq;
2115
2116		if (copy_from_user(&hreq, arg, sizeof(hreq)))
2117			return -EFAULT;
2118		return xfs_find_handle(cmd, &hreq);
2119	}
2120	case XFS_IOC_OPEN_BY_HANDLE: {
2121		xfs_fsop_handlereq_t	hreq;
2122
2123		if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t)))
2124			return -EFAULT;
2125		return xfs_open_by_handle(filp, &hreq);
2126	}
2127
2128	case XFS_IOC_READLINK_BY_HANDLE: {
2129		xfs_fsop_handlereq_t	hreq;
2130
2131		if (copy_from_user(&hreq, arg, sizeof(xfs_fsop_handlereq_t)))
2132			return -EFAULT;
2133		return xfs_readlink_by_handle(filp, &hreq);
2134	}
2135	case XFS_IOC_ATTRLIST_BY_HANDLE:
2136		return xfs_attrlist_by_handle(filp, arg);
2137
2138	case XFS_IOC_ATTRMULTI_BY_HANDLE:
2139		return xfs_attrmulti_by_handle(filp, arg);
2140
2141	case XFS_IOC_SWAPEXT: {
2142		struct xfs_swapext	sxp;
2143
2144		if (copy_from_user(&sxp, arg, sizeof(xfs_swapext_t)))
2145			return -EFAULT;
2146		error = mnt_want_write_file(filp);
2147		if (error)
2148			return error;
2149		error = xfs_ioc_swapext(&sxp);
2150		mnt_drop_write_file(filp);
2151		return error;
2152	}
2153
2154	case XFS_IOC_FSCOUNTS: {
2155		xfs_fsop_counts_t out;
2156
2157		xfs_fs_counts(mp, &out);
2158
2159		if (copy_to_user(arg, &out, sizeof(out)))
2160			return -EFAULT;
2161		return 0;
2162	}
2163
2164	case XFS_IOC_SET_RESBLKS: {
2165		xfs_fsop_resblks_t inout;
2166		uint64_t	   in;
2167
2168		if (!capable(CAP_SYS_ADMIN))
2169			return -EPERM;
2170
2171		if (mp->m_flags & XFS_MOUNT_RDONLY)
2172			return -EROFS;
2173
2174		if (copy_from_user(&inout, arg, sizeof(inout)))
2175			return -EFAULT;
2176
2177		error = mnt_want_write_file(filp);
2178		if (error)
2179			return error;
2180
2181		/* input parameter is passed in resblks field of structure */
2182		in = inout.resblks;
2183		error = xfs_reserve_blocks(mp, &in, &inout);
2184		mnt_drop_write_file(filp);
2185		if (error)
2186			return error;
2187
2188		if (copy_to_user(arg, &inout, sizeof(inout)))
2189			return -EFAULT;
2190		return 0;
2191	}
2192
2193	case XFS_IOC_GET_RESBLKS: {
2194		xfs_fsop_resblks_t out;
2195
2196		if (!capable(CAP_SYS_ADMIN))
2197			return -EPERM;
2198
2199		error = xfs_reserve_blocks(mp, NULL, &out);
2200		if (error)
2201			return error;
2202
2203		if (copy_to_user(arg, &out, sizeof(out)))
2204			return -EFAULT;
2205
2206		return 0;
2207	}
2208
2209	case XFS_IOC_FSGROWFSDATA: {
2210		xfs_growfs_data_t in;
2211
2212		if (copy_from_user(&in, arg, sizeof(in)))
2213			return -EFAULT;
2214
2215		error = mnt_want_write_file(filp);
2216		if (error)
2217			return error;
2218		error = xfs_growfs_data(mp, &in);
2219		mnt_drop_write_file(filp);
2220		return error;
2221	}
2222
2223	case XFS_IOC_FSGROWFSLOG: {
2224		xfs_growfs_log_t in;
2225
2226		if (copy_from_user(&in, arg, sizeof(in)))
2227			return -EFAULT;
2228
2229		error = mnt_want_write_file(filp);
2230		if (error)
2231			return error;
2232		error = xfs_growfs_log(mp, &in);
2233		mnt_drop_write_file(filp);
2234		return error;
2235	}
2236
2237	case XFS_IOC_FSGROWFSRT: {
2238		xfs_growfs_rt_t in;
2239
2240		if (copy_from_user(&in, arg, sizeof(in)))
2241			return -EFAULT;
2242
2243		error = mnt_want_write_file(filp);
2244		if (error)
2245			return error;
2246		error = xfs_growfs_rt(mp, &in);
2247		mnt_drop_write_file(filp);
2248		return error;
2249	}
2250
2251	case XFS_IOC_GOINGDOWN: {
2252		uint32_t in;
2253
2254		if (!capable(CAP_SYS_ADMIN))
2255			return -EPERM;
2256
2257		if (get_user(in, (uint32_t __user *)arg))
2258			return -EFAULT;
2259
2260		return xfs_fs_goingdown(mp, in);
2261	}
2262
2263	case XFS_IOC_ERROR_INJECTION: {
2264		xfs_error_injection_t in;
2265
2266		if (!capable(CAP_SYS_ADMIN))
2267			return -EPERM;
2268
2269		if (copy_from_user(&in, arg, sizeof(in)))
2270			return -EFAULT;
2271
2272		return xfs_errortag_add(mp, in.errtag);
2273	}
2274
2275	case XFS_IOC_ERROR_CLEARALL:
2276		if (!capable(CAP_SYS_ADMIN))
2277			return -EPERM;
2278
2279		return xfs_errortag_clearall(mp);
2280
2281	case XFS_IOC_FREE_EOFBLOCKS: {
2282		struct xfs_fs_eofblocks eofb;
2283		struct xfs_eofblocks keofb;
2284
2285		if (!capable(CAP_SYS_ADMIN))
2286			return -EPERM;
2287
2288		if (mp->m_flags & XFS_MOUNT_RDONLY)
2289			return -EROFS;
2290
2291		if (copy_from_user(&eofb, arg, sizeof(eofb)))
2292			return -EFAULT;
2293
2294		error = xfs_fs_eofblocks_from_user(&eofb, &keofb);
2295		if (error)
2296			return error;
2297
 
 
2298		sb_start_write(mp->m_super);
2299		error = xfs_icache_free_eofblocks(mp, &keofb);
2300		sb_end_write(mp->m_super);
2301		return error;
2302	}
 
 
 
 
 
 
 
2303
2304	default:
2305		return -ENOTTY;
2306	}
2307}