2019-05-22 10:51:42 +03:00
/* SPDX-License-Identifier: GPL-2.0-or-later */
2005-04-17 02:20:36 +04:00
/**
* aops . h - Defines for NTFS kernel address space operations and page cache
* handling . Part of the Linux - NTFS project .
*
* Copyright ( c ) 2001 - 2004 Anton Altaparmakov
* Copyright ( c ) 2002 Richard Russon
*/
# ifndef _LINUX_NTFS_AOPS_H
# define _LINUX_NTFS_AOPS_H
# include <linux/mm.h>
# include <linux/highmem.h>
# include <linux/pagemap.h>
# include <linux/fs.h>
# include "inode.h"
/**
* ntfs_unmap_page - release a page that was mapped using ntfs_map_page ( )
* @ page : the page to release
*
* Unpin , unmap and release a page that was obtained from ntfs_map_page ( ) .
*/
static inline void ntfs_unmap_page ( struct page * page )
{
kunmap ( page ) ;
mm, fs: get rid of PAGE_CACHE_* and page_cache_{get,release} macros
PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} macros were introduced *long* time
ago with promise that one day it will be possible to implement page
cache with bigger chunks than PAGE_SIZE.
This promise never materialized. And unlikely will.
We have many places where PAGE_CACHE_SIZE assumed to be equal to
PAGE_SIZE. And it's constant source of confusion on whether
PAGE_CACHE_* or PAGE_* constant should be used in a particular case,
especially on the border between fs and mm.
Global switching to PAGE_CACHE_SIZE != PAGE_SIZE would cause to much
breakage to be doable.
Let's stop pretending that pages in page cache are special. They are
not.
The changes are pretty straight-forward:
- <foo> << (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- <foo> >> (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} -> PAGE_{SIZE,SHIFT,MASK,ALIGN};
- page_cache_get() -> get_page();
- page_cache_release() -> put_page();
This patch contains automated changes generated with coccinelle using
script below. For some reason, coccinelle doesn't patch header files.
I've called spatch for them manually.
The only adjustment after coccinelle is revert of changes to
PAGE_CAHCE_ALIGN definition: we are going to drop it later.
There are few places in the code where coccinelle didn't reach. I'll
fix them manually in a separate patch. Comments and documentation also
will be addressed with the separate patch.
virtual patch
@@
expression E;
@@
- E << (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
expression E;
@@
- E >> (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
@@
- PAGE_CACHE_SHIFT
+ PAGE_SHIFT
@@
@@
- PAGE_CACHE_SIZE
+ PAGE_SIZE
@@
@@
- PAGE_CACHE_MASK
+ PAGE_MASK
@@
expression E;
@@
- PAGE_CACHE_ALIGN(E)
+ PAGE_ALIGN(E)
@@
expression E;
@@
- page_cache_get(E)
+ get_page(E)
@@
expression E;
@@
- page_cache_release(E)
+ put_page(E)
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2016-04-01 15:29:47 +03:00
put_page ( page ) ;
2005-04-17 02:20:36 +04:00
}
/**
* ntfs_map_page - map a page into accessible memory , reading it if necessary
* @ mapping : address space for which to obtain the page
* @ index : index into the page cache for @ mapping of the page to map
*
* Read a page from the page cache of the address space @ mapping at position
2016-04-01 15:29:48 +03:00
* @ index , where @ index is in units of PAGE_SIZE , and not in bytes .
2005-04-17 02:20:36 +04:00
*
* If the page is not in memory it is loaded from disk first using the readpage
* method defined in the address space operations of @ mapping and the page is
* added to the page cache of @ mapping in the process .
*
* If the page belongs to an mst protected attribute and it is marked as such
* in its ntfs inode ( NInoMstProtected ( ) ) the mst fixups are applied but no
* error checking is performed . This means the caller has to verify whether
* the ntfs record ( s ) contained in the page are valid or not using one of the
* ntfs_is_XXXX_record { , p } ( ) macros , where XXXX is the record type you are
* expecting to see . ( For details of the macros , see fs / ntfs / layout . h . )
*
* If the page is in high memory it is mapped into memory directly addressible
* by the kernel .
*
* Finally the page count is incremented , thus pinning the page into place .
*
* The above means that page_address ( page ) can be used on all pages obtained
* with ntfs_map_page ( ) to get the kernel virtual address of the page .
*
* When finished with the page , the caller has to call ntfs_unmap_page ( ) to
* unpin , unmap and release the page .
*
* Note this does not grant exclusive access . If such is desired , the caller
* must provide it independently of the ntfs_ { un } map_page ( ) calls by using
* a { rw_ } semaphore or other means of serialization . A spin lock cannot be
* used as ntfs_map_page ( ) can block .
*
* The unlocked and uptodate page is returned on success or an encoded error
* on failure . Caller has to test for error using the IS_ERR ( ) macro on the
2006-10-01 10:27:12 +04:00
* return value . If that evaluates to ' true ' , the negative error code can be
2005-04-17 02:20:36 +04:00
* obtained using PTR_ERR ( ) on the return value of ntfs_map_page ( ) .
*/
static inline struct page * ntfs_map_page ( struct address_space * mapping ,
unsigned long index )
{
2006-06-23 13:05:08 +04:00
struct page * page = read_mapping_page ( mapping , index , NULL ) ;
2005-04-17 02:20:36 +04:00
if ( ! IS_ERR ( page ) ) {
kmap ( page ) ;
2007-05-07 01:49:04 +04:00
if ( ! PageError ( page ) )
2005-04-17 02:20:36 +04:00
return page ;
ntfs_unmap_page ( page ) ;
return ERR_PTR ( - EIO ) ;
}
return page ;
}
# ifdef NTFS_RW
extern void mark_ntfs_record_dirty ( struct page * page , const unsigned int ofs ) ;
# endif /* NTFS_RW */
# endif /* _LINUX_NTFS_AOPS_H */