Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
b5f5914c
Commit
b5f5914c
authored
Aug 18, 2015
by
Al Viro
Browse files
Options
Browse Files
Download
Plain Diff
Merge branch 'ufs' into for-next
parents
15cf3b7a
4e317ce7
Changes
6
Hide whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
644 additions
and
882 deletions
+644
-882
fs/ufs/Makefile
fs/ufs/Makefile
+1
-1
fs/ufs/balloc.c
fs/ufs/balloc.c
+4
-0
fs/ufs/inode.c
fs/ufs/inode.c
+632
-316
fs/ufs/super.c
fs/ufs/super.c
+4
-32
fs/ufs/truncate.c
fs/ufs/truncate.c
+0
-523
fs/ufs/ufs.h
fs/ufs/ufs.h
+3
-10
No files found.
fs/ufs/Makefile
View file @
b5f5914c
...
@@ -5,5 +5,5 @@
...
@@ -5,5 +5,5 @@
obj-$(CONFIG_UFS_FS)
+=
ufs.o
obj-$(CONFIG_UFS_FS)
+=
ufs.o
ufs-objs
:=
balloc.o cylinder.o dir.o file.o ialloc.o inode.o
\
ufs-objs
:=
balloc.o cylinder.o dir.o file.o ialloc.o inode.o
\
namei.o super.o symlink.o
truncate.o
util.o
namei.o super.o symlink.o util.o
ccflags-$(CONFIG_UFS_DEBUG)
+=
-DDEBUG
ccflags-$(CONFIG_UFS_DEBUG)
+=
-DDEBUG
fs/ufs/balloc.c
View file @
b5f5914c
...
@@ -417,7 +417,9 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
...
@@ -417,7 +417,9 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
if
(
oldcount
==
0
)
{
if
(
oldcount
==
0
)
{
result
=
ufs_alloc_fragments
(
inode
,
cgno
,
goal
,
count
,
err
);
result
=
ufs_alloc_fragments
(
inode
,
cgno
,
goal
,
count
,
err
);
if
(
result
)
{
if
(
result
)
{
write_seqlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
ufs_cpu_to_data_ptr
(
sb
,
p
,
result
);
ufs_cpu_to_data_ptr
(
sb
,
p
,
result
);
write_sequnlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
*
err
=
0
;
*
err
=
0
;
UFS_I
(
inode
)
->
i_lastfrag
=
UFS_I
(
inode
)
->
i_lastfrag
=
max
(
UFS_I
(
inode
)
->
i_lastfrag
,
fragment
+
count
);
max
(
UFS_I
(
inode
)
->
i_lastfrag
,
fragment
+
count
);
...
@@ -473,7 +475,9 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
...
@@ -473,7 +475,9 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
ufs_change_blocknr
(
inode
,
fragment
-
oldcount
,
oldcount
,
ufs_change_blocknr
(
inode
,
fragment
-
oldcount
,
oldcount
,
uspi
->
s_sbbase
+
tmp
,
uspi
->
s_sbbase
+
tmp
,
uspi
->
s_sbbase
+
result
,
locked_page
);
uspi
->
s_sbbase
+
result
,
locked_page
);
write_seqlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
ufs_cpu_to_data_ptr
(
sb
,
p
,
result
);
ufs_cpu_to_data_ptr
(
sb
,
p
,
result
);
write_sequnlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
*
err
=
0
;
*
err
=
0
;
UFS_I
(
inode
)
->
i_lastfrag
=
max
(
UFS_I
(
inode
)
->
i_lastfrag
,
UFS_I
(
inode
)
->
i_lastfrag
=
max
(
UFS_I
(
inode
)
->
i_lastfrag
,
fragment
+
count
);
fragment
+
count
);
...
...
fs/ufs/inode.c
View file @
b5f5914c
...
@@ -41,9 +41,7 @@
...
@@ -41,9 +41,7 @@
#include "swab.h"
#include "swab.h"
#include "util.h"
#include "util.h"
static
u64
ufs_frag_map
(
struct
inode
*
inode
,
sector_t
frag
,
bool
needs_lock
);
static
int
ufs_block_to_path
(
struct
inode
*
inode
,
sector_t
i_block
,
unsigned
offsets
[
4
])
static
int
ufs_block_to_path
(
struct
inode
*
inode
,
sector_t
i_block
,
sector_t
offsets
[
4
])
{
{
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
inode
->
i_sb
)
->
s_uspi
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
inode
->
i_sb
)
->
s_uspi
;
int
ptrs
=
uspi
->
s_apb
;
int
ptrs
=
uspi
->
s_apb
;
...
@@ -75,227 +73,232 @@ static int ufs_block_to_path(struct inode *inode, sector_t i_block, sector_t off
...
@@ -75,227 +73,232 @@ static int ufs_block_to_path(struct inode *inode, sector_t i_block, sector_t off
return
n
;
return
n
;
}
}
typedef
struct
{
void
*
p
;
union
{
__fs32
key32
;
__fs64
key64
;
};
struct
buffer_head
*
bh
;
}
Indirect
;
static
inline
int
grow_chain32
(
struct
ufs_inode_info
*
ufsi
,
struct
buffer_head
*
bh
,
__fs32
*
v
,
Indirect
*
from
,
Indirect
*
to
)
{
Indirect
*
p
;
unsigned
seq
;
to
->
bh
=
bh
;
do
{
seq
=
read_seqbegin
(
&
ufsi
->
meta_lock
);
to
->
key32
=
*
(
__fs32
*
)(
to
->
p
=
v
);
for
(
p
=
from
;
p
<=
to
&&
p
->
key32
==
*
(
__fs32
*
)
p
->
p
;
p
++
)
;
}
while
(
read_seqretry
(
&
ufsi
->
meta_lock
,
seq
));
return
(
p
>
to
);
}
static
inline
int
grow_chain64
(
struct
ufs_inode_info
*
ufsi
,
struct
buffer_head
*
bh
,
__fs64
*
v
,
Indirect
*
from
,
Indirect
*
to
)
{
Indirect
*
p
;
unsigned
seq
;
to
->
bh
=
bh
;
do
{
seq
=
read_seqbegin
(
&
ufsi
->
meta_lock
);
to
->
key64
=
*
(
__fs64
*
)(
to
->
p
=
v
);
for
(
p
=
from
;
p
<=
to
&&
p
->
key64
==
*
(
__fs64
*
)
p
->
p
;
p
++
)
;
}
while
(
read_seqretry
(
&
ufsi
->
meta_lock
,
seq
));
return
(
p
>
to
);
}
/*
/*
* Returns the location of the fragment from
* Returns the location of the fragment from
* the beginning of the filesystem.
* the beginning of the filesystem.
*/
*/
static
u64
ufs_frag_map
(
struct
inode
*
inode
,
sector_t
frag
,
bool
needs_lock
)
static
u64
ufs_frag_map
(
struct
inode
*
inode
,
unsigned
offsets
[
4
],
int
depth
)
{
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
u64
mask
=
(
u64
)
uspi
->
s_apbmask
>>
uspi
->
s_fpbshift
;
u64
mask
=
(
u64
)
uspi
->
s_apbmask
>>
uspi
->
s_fpbshift
;
int
shift
=
uspi
->
s_apbshift
-
uspi
->
s_fpbshift
;
int
shift
=
uspi
->
s_apbshift
-
uspi
->
s_fpbshift
;
sector_t
offsets
[
4
],
*
p
;
Indirect
chain
[
4
],
*
q
=
chain
;
int
depth
=
ufs_block_to_path
(
inode
,
frag
>>
uspi
->
s_fpbshift
,
offsets
);
unsigned
*
p
;
u64
ret
=
0L
;
__fs32
block
;
__fs64
u2_block
=
0L
;
unsigned
flags
=
UFS_SB
(
sb
)
->
s_flags
;
unsigned
flags
=
UFS_SB
(
sb
)
->
s_flags
;
u64
temp
=
0L
;
u64
res
=
0
;
UFSD
(
": frag = %llu depth = %d
\n
"
,
(
unsigned
long
long
)
frag
,
depth
);
UFSD
(
": uspi->s_fpbshift = %d ,uspi->s_apbmask = %x, mask=%llx
\n
"
,
UFSD
(
": uspi->s_fpbshift = %d ,uspi->s_apbmask = %x, mask=%llx
\n
"
,
uspi
->
s_fpbshift
,
uspi
->
s_apbmask
,
uspi
->
s_fpbshift
,
uspi
->
s_apbmask
,
(
unsigned
long
long
)
mask
);
(
unsigned
long
long
)
mask
);
if
(
depth
==
0
)
if
(
depth
==
0
)
return
0
;
goto
no_block
;
again:
p
=
offsets
;
p
=
offsets
;
if
(
needs_lock
)
lock_ufs
(
sb
);
if
((
flags
&
UFS_TYPE_MASK
)
==
UFS_TYPE_UFS2
)
if
((
flags
&
UFS_TYPE_MASK
)
==
UFS_TYPE_UFS2
)
goto
ufs2
;
goto
ufs2
;
block
=
ufsi
->
i_u1
.
i_data
[
*
p
++
];
if
(
!
grow_chain32
(
ufsi
,
NULL
,
&
ufsi
->
i_u1
.
i_data
[
*
p
++
],
chain
,
q
))
if
(
!
block
)
goto
changed
;
goto
out
;
if
(
!
q
->
key32
)
goto
no_block
;
while
(
--
depth
)
{
while
(
--
depth
)
{
__fs32
*
ptr
;
struct
buffer_head
*
bh
;
struct
buffer_head
*
bh
;
sector_t
n
=
*
p
++
;
unsigned
n
=
*
p
++
;
bh
=
sb_bread
(
sb
,
uspi
->
s_sbbase
+
fs32_to_cpu
(
sb
,
block
)
+
(
n
>>
shift
));
bh
=
sb_bread
(
sb
,
uspi
->
s_sbbase
+
fs32_to_cpu
(
sb
,
q
->
key32
)
+
(
n
>>
shift
));
if
(
!
bh
)
if
(
!
bh
)
goto
out
;
goto
no_block
;
block
=
((
__fs32
*
)
bh
->
b_data
)[
n
&
mask
];
ptr
=
(
__fs32
*
)
bh
->
b_data
+
(
n
&
mask
);
brelse
(
bh
);
if
(
!
grow_chain32
(
ufsi
,
bh
,
ptr
,
chain
,
++
q
))
if
(
!
block
)
goto
changed
;
goto
out
;
if
(
!
q
->
key32
)
}
goto
no_block
;
ret
=
(
u64
)
(
uspi
->
s_sbbase
+
fs32_to_cpu
(
sb
,
block
)
+
(
frag
&
uspi
->
s_fpbmask
));
}
goto
out
;
res
=
fs32_to_cpu
(
sb
,
q
->
key32
);
ufs2:
goto
found
;
u2_block
=
ufsi
->
i_u1
.
u2_i_data
[
*
p
++
];
if
(
!
u2_block
)
goto
out
;
ufs2:
if
(
!
grow_chain64
(
ufsi
,
NULL
,
&
ufsi
->
i_u1
.
u2_i_data
[
*
p
++
],
chain
,
q
))
goto
changed
;
if
(
!
q
->
key64
)
goto
no_block
;
while
(
--
depth
)
{
while
(
--
depth
)
{
__fs64
*
ptr
;
struct
buffer_head
*
bh
;
struct
buffer_head
*
bh
;
sector_t
n
=
*
p
++
;
unsigned
n
=
*
p
++
;
temp
=
(
u64
)(
uspi
->
s_sbbase
)
+
fs64_to_cpu
(
sb
,
u2_block
);
bh
=
sb_bread
(
sb
,
uspi
->
s_sbbase
+
bh
=
sb_bread
(
sb
,
temp
+
(
u64
)
(
n
>>
shift
));
fs64_to_cpu
(
sb
,
q
->
key64
)
+
(
n
>>
shift
));
if
(
!
bh
)
if
(
!
bh
)
goto
out
;
goto
no_block
;
u2_block
=
((
__fs64
*
)
bh
->
b_data
)[
n
&
mask
];
ptr
=
(
__fs64
*
)
bh
->
b_data
+
(
n
&
mask
);
brelse
(
bh
);
if
(
!
grow_chain64
(
ufsi
,
bh
,
ptr
,
chain
,
++
q
))
if
(
!
u2_block
)
goto
changed
;
goto
out
;
if
(
!
q
->
key64
)
goto
no_block
;
}
}
temp
=
(
u64
)
uspi
->
s_sbbase
+
fs64_to_cpu
(
sb
,
u2_block
);
res
=
fs64_to_cpu
(
sb
,
q
->
key64
);
ret
=
temp
+
(
u64
)
(
frag
&
uspi
->
s_fpbmask
);
found:
res
+=
uspi
->
s_sbbase
;
no_block:
while
(
q
>
chain
)
{
brelse
(
q
->
bh
);
q
--
;
}
return
res
;
out:
changed:
if
(
needs_lock
)
while
(
q
>
chain
)
{
unlock_ufs
(
sb
);
brelse
(
q
->
bh
);
return
ret
;
q
--
;
}
goto
again
;
}
/*
* Unpacking tails: we have a file with partial final block and
* we had been asked to extend it. If the fragment being written
* is within the same block, we need to extend the tail just to cover
* that fragment. Otherwise the tail is extended to full block.
*
* Note that we might need to create a _new_ tail, but that will
* be handled elsewhere; this is strictly for resizing old
* ones.
*/
static
bool
ufs_extend_tail
(
struct
inode
*
inode
,
u64
writes_to
,
int
*
err
,
struct
page
*
locked_page
)
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
unsigned
lastfrag
=
ufsi
->
i_lastfrag
;
/* it's a short file, so unsigned is enough */
unsigned
block
=
ufs_fragstoblks
(
lastfrag
);
unsigned
new_size
;
void
*
p
;
u64
tmp
;
if
(
writes_to
<
(
lastfrag
|
uspi
->
s_fpbmask
))
new_size
=
(
writes_to
&
uspi
->
s_fpbmask
)
+
1
;
else
new_size
=
uspi
->
s_fpb
;
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
block
);
tmp
=
ufs_new_fragments
(
inode
,
p
,
lastfrag
,
ufs_data_ptr_to_cpu
(
sb
,
p
),
new_size
,
err
,
locked_page
);
return
tmp
!=
0
;
}
}
/**
/**
* ufs_inode_getfrag() - allocate new fragment(s)
* ufs_inode_getfrag() - allocate new fragment(s)
* @inode: pointer to inode
* @inode: pointer to inode
* @fragment: number of `fragment' which hold pointer
* @index: number of block pointer within the inode's array.
* to new allocated fragment(s)
* @new_fragment: number of new allocated fragment(s)
* @new_fragment: number of new allocated fragment(s)
* @required: how many fragment(s) we require
* @err: we set it if something wrong
* @err: we set it if something wrong
* @phys: pointer to where we save physical number of new allocated fragments,
* NULL if we allocate not data(indirect blocks for example).
* @new: we set it if we allocate new block
* @new: we set it if we allocate new block
* @locked_page: for ufs_new_fragments()
* @locked_page: for ufs_new_fragments()
*/
*/
static
struct
buffer_head
*
static
u64
ufs_inode_getfrag
(
struct
inode
*
inode
,
u
64
fragment
,
ufs_inode_getfrag
(
struct
inode
*
inode
,
u
nsigned
index
,
sector_t
new_fragment
,
unsigned
int
required
,
int
*
err
,
sector_t
new_fragment
,
int
*
err
,
long
*
phys
,
int
*
new
,
struct
page
*
locked_page
)
int
*
new
,
struct
page
*
locked_page
)
{
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
buffer_head
*
result
;
u64
tmp
,
goal
,
lastfrag
;
unsigned
blockoff
,
lastblockoff
;
unsigned
nfrags
=
uspi
->
s_fpb
;
u64
tmp
,
goal
,
lastfrag
,
block
,
lastblock
;
void
*
p
;
void
*
p
,
*
p2
;
UFSD
(
"ENTER, ino %lu, fragment %llu, new_fragment %llu, required %u, "
"metadata %d
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
fragment
,
(
unsigned
long
long
)
new_fragment
,
required
,
!
phys
);
/* TODO : to be done for write support
/* TODO : to be done for write support
if ( (flags & UFS_TYPE_MASK) == UFS_TYPE_UFS2)
if ( (flags & UFS_TYPE_MASK) == UFS_TYPE_UFS2)
goto ufs2;
goto ufs2;
*/
*/
block
=
ufs_fragstoblks
(
fragment
);
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
index
);
blockoff
=
ufs_fragnum
(
fragment
);
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
block
);
goal
=
0
;
repeat:
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
tmp
)
goto
out
;
lastfrag
=
ufsi
->
i_lastfrag
;
lastfrag
=
ufsi
->
i_lastfrag
;
if
(
tmp
&&
fragment
<
lastfrag
)
{
if
(
!
phys
)
{
result
=
sb_getblk
(
sb
,
uspi
->
s_sbbase
+
tmp
+
blockoff
);
if
(
tmp
==
ufs_data_ptr_to_cpu
(
sb
,
p
))
{
UFSD
(
"EXIT, result %llu
\n
"
,
(
unsigned
long
long
)
tmp
+
blockoff
);
return
result
;
}
brelse
(
result
);
goto
repeat
;
}
else
{
*
phys
=
uspi
->
s_sbbase
+
tmp
+
blockoff
;
return
NULL
;
}
}
lastblock
=
ufs_fragstoblks
(
lastfrag
);
/* will that be a new tail? */
lastblockoff
=
ufs_fragnum
(
lastfrag
);
if
(
new_fragment
<
UFS_NDIR_FRAGMENT
&&
new_fragment
>=
lastfrag
)
/*
nfrags
=
(
new_fragment
&
uspi
->
s_fpbmask
)
+
1
;
* We will extend file into new block beyond last allocated block
*/
goal
=
0
;
if
(
lastblock
<
block
)
{
if
(
index
)
{
/*
goal
=
ufs_data_ptr_to_cpu
(
sb
,
* We must reallocate last allocated block
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
index
-
1
));
*/
if
(
goal
)
if
(
lastblockoff
)
{
goal
+=
uspi
->
s_fpb
;
p2
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
lastblock
);
tmp
=
ufs_new_fragments
(
inode
,
p2
,
lastfrag
,
ufs_data_ptr_to_cpu
(
sb
,
p2
),
uspi
->
s_fpb
-
lastblockoff
,
err
,
locked_page
);
if
(
!
tmp
)
{
if
(
lastfrag
!=
ufsi
->
i_lastfrag
)
goto
repeat
;
else
return
NULL
;
}
lastfrag
=
ufsi
->
i_lastfrag
;
}
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
lastblock
));
if
(
tmp
)
goal
=
tmp
+
uspi
->
s_fpb
;
tmp
=
ufs_new_fragments
(
inode
,
p
,
fragment
-
blockoff
,
goal
,
required
+
blockoff
,
err
,
phys
!=
NULL
?
locked_page
:
NULL
);
}
else
if
(
lastblock
==
block
)
{
/*
* We will extend last allocated block
*/
tmp
=
ufs_new_fragments
(
inode
,
p
,
fragment
-
(
blockoff
-
lastblockoff
),
ufs_data_ptr_to_cpu
(
sb
,
p
),
required
+
(
blockoff
-
lastblockoff
),
err
,
phys
!=
NULL
?
locked_page
:
NULL
);
}
else
/* (lastblock > block) */
{
/*
* We will allocate new block before last allocated block
*/
if
(
block
)
{
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
block
-
1
));
if
(
tmp
)
goal
=
tmp
+
uspi
->
s_fpb
;
}
tmp
=
ufs_new_fragments
(
inode
,
p
,
fragment
-
blockoff
,
goal
,
uspi
->
s_fpb
,
err
,
phys
!=
NULL
?
locked_page
:
NULL
);
}
}
tmp
=
ufs_new_fragments
(
inode
,
p
,
ufs_blknum
(
new_fragment
),
goal
,
uspi
->
s_fpb
,
err
,
locked_page
);
if
(
!
tmp
)
{
if
(
!
tmp
)
{
if
((
!
blockoff
&&
ufs_data_ptr_to_cpu
(
sb
,
p
))
||
(
blockoff
&&
lastfrag
!=
ufsi
->
i_lastfrag
))
goto
repeat
;
*
err
=
-
ENOSPC
;
*
err
=
-
ENOSPC
;
return
NULL
;
return
0
;
}
}
if
(
!
phys
)
{
if
(
new
)
result
=
sb_getblk
(
sb
,
uspi
->
s_sbbase
+
tmp
+
blockoff
);
}
else
{
*
phys
=
uspi
->
s_sbbase
+
tmp
+
blockoff
;
result
=
NULL
;
*
err
=
0
;
*
new
=
1
;
*
new
=
1
;
}
inode
->
i_ctime
=
CURRENT_TIME_SEC
;
inode
->
i_ctime
=
CURRENT_TIME_SEC
;
if
(
IS_SYNC
(
inode
))
if
(
IS_SYNC
(
inode
))
ufs_sync_inode
(
inode
);
ufs_sync_inode
(
inode
);
mark_inode_dirty
(
inode
);
mark_inode_dirty
(
inode
);
UFSD
(
"EXIT, result %llu
\n
"
,
(
unsigned
long
long
)
tmp
+
blockoff
);
out:
return
result
;
return
tmp
+
uspi
->
s_sbbase
;
/* This part : To be implemented ....
/* This part : To be implemented ....
Required only for writing, not required for READ-ONLY.
Required only for writing, not required for READ-ONLY.
...
@@ -316,95 +319,70 @@ ufs_inode_getfrag(struct inode *inode, u64 fragment,
...
@@ -316,95 +319,70 @@ ufs_inode_getfrag(struct inode *inode, u64 fragment,
/**
/**
* ufs_inode_getblock() - allocate new block
* ufs_inode_getblock() - allocate new block
* @inode: pointer to inode
* @inode: pointer to inode
* @bh: pointer to block which hold "pointer" to new allocated block
* @ind_block: block number of the indirect block
* @fragment: number of `fragment' which hold pointer
* @index: number of pointer within the indirect block
* to new allocated block
* @new_fragment: number of new allocated fragment
* @new_fragment: number of new allocated fragment
* (block will hold this fragment and also uspi->s_fpb-1)
* (block will hold this fragment and also uspi->s_fpb-1)
* @err: see ufs_inode_getfrag()
* @err: see ufs_inode_getfrag()
* @phys: see ufs_inode_getfrag()
* @new: see ufs_inode_getfrag()
* @new: see ufs_inode_getfrag()
* @locked_page: see ufs_inode_getfrag()
* @locked_page: see ufs_inode_getfrag()
*/
*/
static
struct
buffer_head
*
static
u64
ufs_inode_getblock
(
struct
inode
*
inode
,
struct
buffer_head
*
bh
,
ufs_inode_getblock
(
struct
inode
*
inode
,
u64
ind_block
,
u
64
fragment
,
sector_t
new_fragment
,
int
*
err
,
u
nsigned
index
,
sector_t
new_fragment
,
int
*
err
,
long
*
phys
,
int
*
new
,
struct
page
*
locked_page
)
int
*
new
,
struct
page
*
locked_page
)
{
{
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
buffer_head
*
resul
t
;
int
shift
=
uspi
->
s_apbshift
-
uspi
->
s_fpbshif
t
;
u
nsigned
blockoff
;
u
64
tmp
=
0
,
goal
;
u64
tmp
,
goal
,
block
;
struct
buffer_head
*
bh
;
void
*
p
;
void
*
p
;
block
=
ufs_fragstoblks
(
fragment
);
if
(
!
ind_block
)
blockoff
=
ufs_fragnum
(
fragment
);
return
0
;
UFSD
(
"ENTER, ino %lu, fragment %llu, new_fragment %llu, metadata %d
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
fragment
,
(
unsigned
long
long
)
new_fragment
,
!
phys
);
result
=
NULL
;
bh
=
sb_bread
(
sb
,
ind_block
+
(
index
>>
shift
));
if
(
!
bh
)
if
(
unlikely
(
!
bh
))
{
goto
out
;
*
err
=
-
EIO
;
if
(
!
buffer_uptodate
(
bh
))
{
return
0
;
ll_rw_block
(
READ
,
1
,
&
bh
);
wait_on_buffer
(
bh
);
if
(
!
buffer_uptodate
(
bh
))
goto
out
;
}
}
index
&=
uspi
->
s_apbmask
>>
uspi
->
s_fpbshift
;
if
(
uspi
->
fs_magic
==
UFS2_MAGIC
)
if
(
uspi
->
fs_magic
==
UFS2_MAGIC
)
p
=
(
__fs64
*
)
bh
->
b_data
+
block
;
p
=
(
__fs64
*
)
bh
->
b_data
+
index
;
else
else
p
=
(
__fs32
*
)
bh
->
b_data
+
block
;
p
=
(
__fs32
*
)
bh
->
b_data
+
index
;
repeat:
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
tmp
)
{
if
(
tmp
)
if
(
!
phys
)
{
goto
out
;
result
=
sb_getblk
(
sb
,
uspi
->
s_sbbase
+
tmp
+
blockoff
);
if
(
tmp
==
ufs_data_ptr_to_cpu
(
sb
,
p
))
goto
out
;
brelse
(
result
);
goto
repeat
;
}
else
{
*
phys
=
uspi
->
s_sbbase
+
tmp
+
blockoff
;
goto
out
;
}
}
if
(
block
&&
(
uspi
->
fs_magic
==
UFS2_MAGIC
?
if
(
index
&&
(
uspi
->
fs_magic
==
UFS2_MAGIC
?
(
tmp
=
fs64_to_cpu
(
sb
,
((
__fs64
*
)
bh
->
b_data
)[
block
-
1
]))
:
(
tmp
=
fs64_to_cpu
(
sb
,
((
__fs64
*
)
bh
->
b_data
)[
index
-
1
]))
:
(
tmp
=
fs32_to_cpu
(
sb
,
((
__fs32
*
)
bh
->
b_data
)[
block
-
1
]))))
(
tmp
=
fs32_to_cpu
(
sb
,
((
__fs32
*
)
bh
->
b_data
)[
index
-
1
]))))
goal
=
tmp
+
uspi
->
s_fpb
;
goal
=
tmp
+
uspi
->
s_fpb
;
else
else
goal
=
bh
->
b_blocknr
+
uspi
->
s_fpb
;
goal
=
bh
->
b_blocknr
+
uspi
->
s_fpb
;
tmp
=
ufs_new_fragments
(
inode
,
p
,
ufs_blknum
(
new_fragment
),
goal
,
tmp
=
ufs_new_fragments
(
inode
,
p
,
ufs_blknum
(
new_fragment
),
goal
,
uspi
->
s_fpb
,
err
,
locked_page
);
uspi
->
s_fpb
,
err
,
locked_page
);
if
(
!
tmp
)
{
if
(
!
tmp
)
if
(
ufs_data_ptr_to_cpu
(
sb
,
p
))
goto
repeat
;
goto
out
;
goto
out
;
}
if
(
new
)
if
(
!
phys
)
{
result
=
sb_getblk
(
sb
,
uspi
->
s_sbbase
+
tmp
+
blockoff
);
}
else
{
*
phys
=
uspi
->
s_sbbase
+
tmp
+
blockoff
;
*
new
=
1
;
*
new
=
1
;
}
mark_buffer_dirty
(
bh
);
mark_buffer_dirty
(
bh
);
if
(
IS_SYNC
(
inode
))
if
(
IS_SYNC
(
inode
))
sync_dirty_buffer
(
bh
);
sync_dirty_buffer
(
bh
);
inode
->
i_ctime
=
CURRENT_TIME_SEC
;
inode
->
i_ctime
=
CURRENT_TIME_SEC
;
mark_inode_dirty
(
inode
);
mark_inode_dirty
(
inode
);
UFSD
(
"result %llu
\n
"
,
(
unsigned
long
long
)
tmp
+
blockoff
);
out:
out:
brelse
(
bh
);
brelse
(
bh
);
UFSD
(
"EXIT
\n
"
);
UFSD
(
"EXIT
\n
"
);
return
result
;
if
(
tmp
)
tmp
+=
uspi
->
s_sbbase
;
return
tmp
;
}
}
/**
/**
...
@@ -412,103 +390,64 @@ ufs_inode_getblock(struct inode *inode, struct buffer_head *bh,
...
@@ -412,103 +390,64 @@ ufs_inode_getblock(struct inode *inode, struct buffer_head *bh,
* readpage, writepage and so on
* readpage, writepage and so on
*/
*/
int
ufs_getfrag_block
(
struct
inode
*
inode
,
sector_t
fragment
,
struct
buffer_head
*
bh_result
,
int
create
)
static
int
ufs_getfrag_block
(
struct
inode
*
inode
,
sector_t
fragment
,
struct
buffer_head
*
bh_result
,
int
create
)
{
{
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_info
*
sbi
=
UFS_SB
(
sb
);
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_sb_private_info
*
uspi
=
sbi
->
s_uspi
;
int
err
=
0
,
new
=
0
;
struct
buffer_head
*
bh
;
unsigned
offsets
[
4
];
int
ret
,
err
,
new
;
int
depth
=
ufs_block_to_path
(
inode
,
fragment
>>
uspi
->
s_fpbshift
,
offsets
);
unsigned
long
ptr
,
phys
;
u64
phys64
=
0
;
u64
phys64
=
0
;
bool
needs_lock
=
(
sbi
->
mutex_owner
!=
current
)
;
unsigned
frag
=
fragment
&
uspi
->
s_fpbmask
;
if
(
!
create
)
{
if
(
!
create
)
{
phys64
=
ufs_frag_map
(
inode
,
fragment
,
needs_lock
);
phys64
=
ufs_frag_map
(
inode
,
offsets
,
depth
);
UFSD
(
"phys64 = %llu
\n
"
,
(
unsigned
long
long
)
phys64
);
goto
out
;
if
(
phys64
)
map_bh
(
bh_result
,
sb
,
phys64
);
return
0
;
}
}
/* This code entered only while writing ....? */
/* This code entered only while writing ....? */
err
=
-
EIO
;
mutex_lock
(
&
UFS_I
(
inode
)
->
truncate_mutex
);
new
=
0
;
ret
=
0
;
bh
=
NULL
;
if
(
needs_lock
)
lock_ufs
(
sb
);
UFSD
(
"ENTER, ino %lu, fragment %llu
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
fragment
);
UFSD
(
"ENTER, ino %lu, fragment %llu
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
fragment
);
if
(
fragment
>
if
(
unlikely
(
!
depth
))
{
((
UFS_NDADDR
+
uspi
->
s_apb
+
uspi
->
s_2apb
+
uspi
->
s_3apb
)
ufs_warning
(
sb
,
"ufs_get_block"
,
"block > big"
);
<<
uspi
->
s_fpbshift
))
err
=
-
EIO
;
goto
abort_too_big
;
err
=
0
;
ptr
=
fragment
;
/*
* ok, these macros clean the logic up a bit and make
* it much more readable:
*/
#define GET_INODE_DATABLOCK(x) \
ufs_inode_getfrag(inode, x, fragment, 1, &err, &phys, &new,\
bh_result->b_page)
#define GET_INODE_PTR(x) \
ufs_inode_getfrag(inode, x, fragment, uspi->s_fpb, &err, NULL, NULL,\
bh_result->b_page)
#define GET_INDIRECT_DATABLOCK(x) \
ufs_inode_getblock(inode, bh, x, fragment, \
&err, &phys, &new, bh_result->b_page)
#define GET_INDIRECT_PTR(x) \
ufs_inode_getblock(inode, bh, x, fragment, \
&err, NULL, NULL, NULL)
if
(
ptr
<
UFS_NDIR_FRAGMENT
)
{
bh
=
GET_INODE_DATABLOCK
(
ptr
);
goto
out
;
goto
out
;
}
}
ptr
-=
UFS_NDIR_FRAGMENT
;
if
(
ptr
<
(
1
<<
(
uspi
->
s_apbshift
+
uspi
->
s_fpbshift
)))
{
bh
=
GET_INODE_PTR
(
UFS_IND_FRAGMENT
+
(
ptr
>>
uspi
->
s_apbshift
));
goto
get_indirect
;
}
ptr
-=
1
<<
(
uspi
->
s_apbshift
+
uspi
->
s_fpbshift
);
if
(
ptr
<
(
1
<<
(
uspi
->
s_2apbshift
+
uspi
->
s_fpbshift
)))
{
bh
=
GET_INODE_PTR
(
UFS_DIND_FRAGMENT
+
(
ptr
>>
uspi
->
s_2apbshift
));
goto
get_double
;
}
ptr
-=
1
<<
(
uspi
->
s_2apbshift
+
uspi
->
s_fpbshift
);
bh
=
GET_INODE_PTR
(
UFS_TIND_FRAGMENT
+
(
ptr
>>
uspi
->
s_3apbshift
));
bh
=
GET_INDIRECT_PTR
((
ptr
>>
uspi
->
s_2apbshift
)
&
uspi
->
s_apbmask
);
get_double:
bh
=
GET_INDIRECT_PTR
((
ptr
>>
uspi
->
s_apbshift
)
&
uspi
->
s_apbmask
);
get_indirect:
bh
=
GET_INDIRECT_DATABLOCK
(
ptr
&
uspi
->
s_apbmask
);
#undef GET_INODE_DATABLOCK
#undef GET_INODE_PTR
#undef GET_INDIRECT_DATABLOCK
#undef GET_INDIRECT_PTR
out:
if
(
UFS_I
(
inode
)
->
i_lastfrag
<
UFS_NDIR_FRAGMENT
)
{
if
(
err
)
unsigned
lastfrag
=
UFS_I
(
inode
)
->
i_lastfrag
;
goto
abort
;
unsigned
tailfrags
=
lastfrag
&
uspi
->
s_fpbmask
;
if
(
new
)
if
(
tailfrags
&&
fragment
>=
lastfrag
)
{
set_buffer_new
(
bh_result
);
if
(
!
ufs_extend_tail
(
inode
,
fragment
,
map_bh
(
bh_result
,
sb
,
phys
);
&
err
,
bh_result
->
b_page
))
abort:
goto
out
;
if
(
needs_lock
)
}
unlock_ufs
(
sb
);
}
if
(
depth
==
1
)
{
phys64
=
ufs_inode_getfrag
(
inode
,
offsets
[
0
],
fragment
,
&
err
,
&
new
,
bh_result
->
b_page
);
}
else
{
int
i
;
phys64
=
ufs_inode_getfrag
(
inode
,
offsets
[
0
],
fragment
,
&
err
,
NULL
,
NULL
);
for
(
i
=
1
;
i
<
depth
-
1
;
i
++
)
phys64
=
ufs_inode_getblock
(
inode
,
phys64
,
offsets
[
i
],
fragment
,
&
err
,
NULL
,
NULL
);
phys64
=
ufs_inode_getblock
(
inode
,
phys64
,
offsets
[
depth
-
1
],
fragment
,
&
err
,
&
new
,
bh_result
->
b_page
);
}
out:
if
(
phys64
)
{
phys64
+=
frag
;
map_bh
(
bh_result
,
sb
,
phys64
);
if
(
new
)
set_buffer_new
(
bh_result
);
}
mutex_unlock
(
&
UFS_I
(
inode
)
->
truncate_mutex
);
return
err
;
return
err
;
abort_too_big:
ufs_warning
(
sb
,
"ufs_get_block"
,
"block > big"
);
goto
abort
;
}
}
static
int
ufs_writepage
(
struct
page
*
page
,
struct
writeback_control
*
wbc
)
static
int
ufs_writepage
(
struct
page
*
page
,
struct
writeback_control
*
wbc
)
...
@@ -526,12 +465,16 @@ int ufs_prepare_chunk(struct page *page, loff_t pos, unsigned len)
...
@@ -526,12 +465,16 @@ int ufs_prepare_chunk(struct page *page, loff_t pos, unsigned len)
return
__block_write_begin
(
page
,
pos
,
len
,
ufs_getfrag_block
);
return
__block_write_begin
(
page
,
pos
,
len
,
ufs_getfrag_block
);
}
}
static
void
ufs_truncate_blocks
(
struct
inode
*
);
static
void
ufs_write_failed
(
struct
address_space
*
mapping
,
loff_t
to
)
static
void
ufs_write_failed
(
struct
address_space
*
mapping
,
loff_t
to
)
{
{
struct
inode
*
inode
=
mapping
->
host
;
struct
inode
*
inode
=
mapping
->
host
;
if
(
to
>
inode
->
i_size
)
if
(
to
>
inode
->
i_size
)
{
truncate_pagecache
(
inode
,
inode
->
i_size
);
truncate_pagecache
(
inode
,
inode
->
i_size
);
ufs_truncate_blocks
(
inode
);
}
}
}
static
int
ufs_write_begin
(
struct
file
*
file
,
struct
address_space
*
mapping
,
static
int
ufs_write_begin
(
struct
file
*
file
,
struct
address_space
*
mapping
,
...
@@ -548,6 +491,18 @@ static int ufs_write_begin(struct file *file, struct address_space *mapping,
...
@@ -548,6 +491,18 @@ static int ufs_write_begin(struct file *file, struct address_space *mapping,
return
ret
;
return
ret
;
}
}
static
int
ufs_write_end
(
struct
file
*
file
,
struct
address_space
*
mapping
,
loff_t
pos
,
unsigned
len
,
unsigned
copied
,
struct
page
*
page
,
void
*
fsdata
)
{
int
ret
;
ret
=
generic_write_end
(
file
,
mapping
,
pos
,
len
,
copied
,
page
,
fsdata
);
if
(
ret
<
len
)
ufs_write_failed
(
mapping
,
pos
+
len
);
return
ret
;
}
static
sector_t
ufs_bmap
(
struct
address_space
*
mapping
,
sector_t
block
)
static
sector_t
ufs_bmap
(
struct
address_space
*
mapping
,
sector_t
block
)
{
{
return
generic_block_bmap
(
mapping
,
block
,
ufs_getfrag_block
);
return
generic_block_bmap
(
mapping
,
block
,
ufs_getfrag_block
);
...
@@ -557,7 +512,7 @@ const struct address_space_operations ufs_aops = {
...
@@ -557,7 +512,7 @@ const struct address_space_operations ufs_aops = {
.
readpage
=
ufs_readpage
,
.
readpage
=
ufs_readpage
,
.
writepage
=
ufs_writepage
,
.
writepage
=
ufs_writepage
,
.
write_begin
=
ufs_write_begin
,
.
write_begin
=
ufs_write_begin
,
.
write_end
=
generic
_write_end
,
.
write_end
=
ufs
_write_end
,
.
bmap
=
ufs_bmap
.
bmap
=
ufs_bmap
};
};
...
@@ -599,7 +554,7 @@ static int ufs1_read_inode(struct inode *inode, struct ufs_inode *ufs_inode)
...
@@ -599,7 +554,7 @@ static int ufs1_read_inode(struct inode *inode, struct ufs_inode *ufs_inode)
ufs_error
(
sb
,
"ufs_read_inode"
,
"inode %lu has zero nlink
\n
"
,
inode
->
i_ino
);
ufs_error
(
sb
,
"ufs_read_inode"
,
"inode %lu has zero nlink
\n
"
,
inode
->
i_ino
);
return
-
1
;
return
-
1
;
}
}
/*
/*
* Linux now has 32-bit uid and gid, so we can support EFT.
* Linux now has 32-bit uid and gid, so we can support EFT.
*/
*/
...
@@ -619,7 +574,7 @@ static int ufs1_read_inode(struct inode *inode, struct ufs_inode *ufs_inode)
...
@@ -619,7 +574,7 @@ static int ufs1_read_inode(struct inode *inode, struct ufs_inode *ufs_inode)
ufsi
->
i_shadow
=
fs32_to_cpu
(
sb
,
ufs_inode
->
ui_u3
.
ui_sun
.
ui_shadow
);
ufsi
->
i_shadow
=
fs32_to_cpu
(
sb
,
ufs_inode
->
ui_u3
.
ui_sun
.
ui_shadow
);
ufsi
->
i_oeftflag
=
fs32_to_cpu
(
sb
,
ufs_inode
->
ui_u3
.
ui_sun
.
ui_oeftflag
);
ufsi
->
i_oeftflag
=
fs32_to_cpu
(
sb
,
ufs_inode
->
ui_u3
.
ui_sun
.
ui_oeftflag
);
if
(
S_ISCHR
(
mode
)
||
S_ISBLK
(
mode
)
||
inode
->
i_blocks
)
{
if
(
S_ISCHR
(
mode
)
||
S_ISBLK
(
mode
)
||
inode
->
i_blocks
)
{
memcpy
(
ufsi
->
i_u1
.
i_data
,
&
ufs_inode
->
ui_u2
.
ui_addr
,
memcpy
(
ufsi
->
i_u1
.
i_data
,
&
ufs_inode
->
ui_u2
.
ui_addr
,
sizeof
(
ufs_inode
->
ui_u2
.
ui_addr
));
sizeof
(
ufs_inode
->
ui_u2
.
ui_addr
));
...
@@ -753,7 +708,7 @@ static void ufs1_update_inode(struct inode *inode, struct ufs_inode *ufs_inode)
...
@@ -753,7 +708,7 @@ static void ufs1_update_inode(struct inode *inode, struct ufs_inode *ufs_inode)
ufs_set_inode_uid
(
sb
,
ufs_inode
,
i_uid_read
(
inode
));
ufs_set_inode_uid
(
sb
,
ufs_inode
,
i_uid_read
(
inode
));
ufs_set_inode_gid
(
sb
,
ufs_inode
,
i_gid_read
(
inode
));
ufs_set_inode_gid
(
sb
,
ufs_inode
,
i_gid_read
(
inode
));
ufs_inode
->
ui_size
=
cpu_to_fs64
(
sb
,
inode
->
i_size
);
ufs_inode
->
ui_size
=
cpu_to_fs64
(
sb
,
inode
->
i_size
);
ufs_inode
->
ui_atime
.
tv_sec
=
cpu_to_fs32
(
sb
,
inode
->
i_atime
.
tv_sec
);
ufs_inode
->
ui_atime
.
tv_sec
=
cpu_to_fs32
(
sb
,
inode
->
i_atime
.
tv_sec
);
ufs_inode
->
ui_atime
.
tv_usec
=
0
;
ufs_inode
->
ui_atime
.
tv_usec
=
0
;
...
@@ -855,23 +810,19 @@ static int ufs_update_inode(struct inode * inode, int do_sync)
...
@@ -855,23 +810,19 @@ static int ufs_update_inode(struct inode * inode, int do_sync)
ufs1_update_inode
(
inode
,
ufs_inode
+
ufs_inotofsbo
(
inode
->
i_ino
));
ufs1_update_inode
(
inode
,
ufs_inode
+
ufs_inotofsbo
(
inode
->
i_ino
));
}
}
mark_buffer_dirty
(
bh
);
mark_buffer_dirty
(
bh
);
if
(
do_sync
)
if
(
do_sync
)
sync_dirty_buffer
(
bh
);
sync_dirty_buffer
(
bh
);
brelse
(
bh
);
brelse
(
bh
);
UFSD
(
"EXIT
\n
"
);
UFSD
(
"EXIT
\n
"
);
return
0
;
return
0
;
}
}
int
ufs_write_inode
(
struct
inode
*
inode
,
struct
writeback_control
*
wbc
)
int
ufs_write_inode
(
struct
inode
*
inode
,
struct
writeback_control
*
wbc
)
{
{
int
ret
;
return
ufs_update_inode
(
inode
,
wbc
->
sync_mode
==
WB_SYNC_ALL
);
lock_ufs
(
inode
->
i_sb
);
ret
=
ufs_update_inode
(
inode
,
wbc
->
sync_mode
==
WB_SYNC_ALL
);
unlock_ufs
(
inode
->
i_sb
);
return
ret
;
}
}
int
ufs_sync_inode
(
struct
inode
*
inode
)
int
ufs_sync_inode
(
struct
inode
*
inode
)
...
@@ -888,24 +839,389 @@ void ufs_evict_inode(struct inode * inode)
...
@@ -888,24 +839,389 @@ void ufs_evict_inode(struct inode * inode)
truncate_inode_pages_final
(
&
inode
->
i_data
);
truncate_inode_pages_final
(
&
inode
->
i_data
);
if
(
want_delete
)
{
if
(
want_delete
)
{
loff_t
old_i_size
;
/*UFS_I(inode)->i_dtime = CURRENT_TIME;*/
lock_ufs
(
inode
->
i_sb
);
mark_inode_dirty
(
inode
);
ufs_update_inode
(
inode
,
IS_SYNC
(
inode
));
old_i_size
=
inode
->
i_size
;
inode
->
i_size
=
0
;
inode
->
i_size
=
0
;
if
(
inode
->
i_blocks
&&
ufs_truncate
(
inode
,
old_i_size
))
if
(
inode
->
i_blocks
)
ufs_warning
(
inode
->
i_sb
,
__func__
,
"ufs_truncate failed
\n
"
);
ufs_truncate_blocks
(
inode
);
unlock_ufs
(
inode
->
i_sb
);
}
}
invalidate_inode_buffers
(
inode
);
invalidate_inode_buffers
(
inode
);
clear_inode
(
inode
);
clear_inode
(
inode
);
if
(
want_delete
)
{
if
(
want_delete
)
lock_ufs
(
inode
->
i_sb
);
ufs_free_inode
(
inode
);
ufs_free_inode
(
inode
);
unlock_ufs
(
inode
->
i_sb
);
}
struct
to_free
{
struct
inode
*
inode
;
u64
to
;
unsigned
count
;
};
static
inline
void
free_data
(
struct
to_free
*
ctx
,
u64
from
,
unsigned
count
)
{
if
(
ctx
->
count
&&
ctx
->
to
!=
from
)
{
ufs_free_blocks
(
ctx
->
inode
,
ctx
->
to
-
ctx
->
count
,
ctx
->
count
);
ctx
->
count
=
0
;
}
ctx
->
count
+=
count
;
ctx
->
to
=
from
+
count
;
}
#define DIRECT_BLOCK ((inode->i_size + uspi->s_bsize - 1) >> uspi->s_bshift)
#define DIRECT_FRAGMENT ((inode->i_size + uspi->s_fsize - 1) >> uspi->s_fshift)
static
void
ufs_trunc_direct
(
struct
inode
*
inode
)
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
;
struct
ufs_sb_private_info
*
uspi
;
void
*
p
;
u64
frag1
,
frag2
,
frag3
,
frag4
,
block1
,
block2
;
struct
to_free
ctx
=
{.
inode
=
inode
};
unsigned
i
,
tmp
;
UFSD
(
"ENTER: ino %lu
\n
"
,
inode
->
i_ino
);
sb
=
inode
->
i_sb
;
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
frag1
=
DIRECT_FRAGMENT
;
frag4
=
min_t
(
u64
,
UFS_NDIR_FRAGMENT
,
ufsi
->
i_lastfrag
);
frag2
=
((
frag1
&
uspi
->
s_fpbmask
)
?
((
frag1
|
uspi
->
s_fpbmask
)
+
1
)
:
frag1
);
frag3
=
frag4
&
~
uspi
->
s_fpbmask
;
block1
=
block2
=
0
;
if
(
frag2
>
frag3
)
{
frag2
=
frag4
;
frag3
=
frag4
=
0
;
}
else
if
(
frag2
<
frag3
)
{
block1
=
ufs_fragstoblks
(
frag2
);
block2
=
ufs_fragstoblks
(
frag3
);
}
UFSD
(
"ino %lu, frag1 %llu, frag2 %llu, block1 %llu, block2 %llu,"
" frag3 %llu, frag4 %llu
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
frag1
,
(
unsigned
long
long
)
frag2
,
(
unsigned
long
long
)
block1
,
(
unsigned
long
long
)
block2
,
(
unsigned
long
long
)
frag3
,
(
unsigned
long
long
)
frag4
);
if
(
frag1
>=
frag2
)
goto
next1
;
/*
* Free first free fragments
*/
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
ufs_fragstoblks
(
frag1
));
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
ufs_panic
(
sb
,
"ufs_trunc_direct"
,
"internal error"
);
frag2
-=
frag1
;
frag1
=
ufs_fragnum
(
frag1
);
ufs_free_fragments
(
inode
,
tmp
+
frag1
,
frag2
);
next1:
/*
* Free whole blocks
*/
for
(
i
=
block1
;
i
<
block2
;
i
++
)
{
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
i
);
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
continue
;
write_seqlock
(
&
ufsi
->
meta_lock
);
ufs_data_ptr_clear
(
uspi
,
p
);
write_sequnlock
(
&
ufsi
->
meta_lock
);
free_data
(
&
ctx
,
tmp
,
uspi
->
s_fpb
);
}
free_data
(
&
ctx
,
0
,
0
);
if
(
frag3
>=
frag4
)
goto
next3
;
/*
* Free last free fragments
*/
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
ufs_fragstoblks
(
frag3
));
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
ufs_panic
(
sb
,
"ufs_truncate_direct"
,
"internal error"
);
frag4
=
ufs_fragnum
(
frag4
);
write_seqlock
(
&
ufsi
->
meta_lock
);
ufs_data_ptr_clear
(
uspi
,
p
);
write_sequnlock
(
&
ufsi
->
meta_lock
);
ufs_free_fragments
(
inode
,
tmp
,
frag4
);
next3:
UFSD
(
"EXIT: ino %lu
\n
"
,
inode
->
i_ino
);
}
static
void
free_full_branch
(
struct
inode
*
inode
,
u64
ind_block
,
int
depth
)
{
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_buffer_head
*
ubh
=
ubh_bread
(
sb
,
ind_block
,
uspi
->
s_bsize
);
unsigned
i
;
if
(
!
ubh
)
return
;
if
(
--
depth
)
{
for
(
i
=
0
;
i
<
uspi
->
s_apb
;
i
++
)
{
void
*
p
=
ubh_get_data_ptr
(
uspi
,
ubh
,
i
);
u64
block
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
block
)
free_full_branch
(
inode
,
block
,
depth
);
}
}
else
{
struct
to_free
ctx
=
{.
inode
=
inode
};
for
(
i
=
0
;
i
<
uspi
->
s_apb
;
i
++
)
{
void
*
p
=
ubh_get_data_ptr
(
uspi
,
ubh
,
i
);
u64
block
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
block
)
free_data
(
&
ctx
,
block
,
uspi
->
s_fpb
);
}
free_data
(
&
ctx
,
0
,
0
);
}
ubh_bforget
(
ubh
);
ufs_free_blocks
(
inode
,
ind_block
,
uspi
->
s_fpb
);
}
static
void
free_branch_tail
(
struct
inode
*
inode
,
unsigned
from
,
struct
ufs_buffer_head
*
ubh
,
int
depth
)
{
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
unsigned
i
;
if
(
--
depth
)
{
for
(
i
=
from
;
i
<
uspi
->
s_apb
;
i
++
)
{
void
*
p
=
ubh_get_data_ptr
(
uspi
,
ubh
,
i
);
u64
block
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
block
)
{
write_seqlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
ufs_data_ptr_clear
(
uspi
,
p
);
write_sequnlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
ubh_mark_buffer_dirty
(
ubh
);
free_full_branch
(
inode
,
block
,
depth
);
}
}
}
else
{
struct
to_free
ctx
=
{.
inode
=
inode
};
for
(
i
=
from
;
i
<
uspi
->
s_apb
;
i
++
)
{
void
*
p
=
ubh_get_data_ptr
(
uspi
,
ubh
,
i
);
u64
block
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
block
)
{
write_seqlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
ufs_data_ptr_clear
(
uspi
,
p
);
write_sequnlock
(
&
UFS_I
(
inode
)
->
meta_lock
);
ubh_mark_buffer_dirty
(
ubh
);
free_data
(
&
ctx
,
block
,
uspi
->
s_fpb
);
}
}
free_data
(
&
ctx
,
0
,
0
);
}
if
(
IS_SYNC
(
inode
)
&&
ubh_buffer_dirty
(
ubh
))
ubh_sync_block
(
ubh
);
ubh_brelse
(
ubh
);
}
static
int
ufs_alloc_lastblock
(
struct
inode
*
inode
,
loff_t
size
)
{
int
err
=
0
;
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
address_space
*
mapping
=
inode
->
i_mapping
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
unsigned
i
,
end
;
sector_t
lastfrag
;
struct
page
*
lastpage
;
struct
buffer_head
*
bh
;
u64
phys64
;
lastfrag
=
(
size
+
uspi
->
s_fsize
-
1
)
>>
uspi
->
s_fshift
;
if
(
!
lastfrag
)
goto
out
;
lastfrag
--
;
lastpage
=
ufs_get_locked_page
(
mapping
,
lastfrag
>>
(
PAGE_CACHE_SHIFT
-
inode
->
i_blkbits
));
if
(
IS_ERR
(
lastpage
))
{
err
=
-
EIO
;
goto
out
;
}
end
=
lastfrag
&
((
1
<<
(
PAGE_CACHE_SHIFT
-
inode
->
i_blkbits
))
-
1
);
bh
=
page_buffers
(
lastpage
);
for
(
i
=
0
;
i
<
end
;
++
i
)
bh
=
bh
->
b_this_page
;
err
=
ufs_getfrag_block
(
inode
,
lastfrag
,
bh
,
1
);
if
(
unlikely
(
err
))
goto
out_unlock
;
if
(
buffer_new
(
bh
))
{
clear_buffer_new
(
bh
);
unmap_underlying_metadata
(
bh
->
b_bdev
,
bh
->
b_blocknr
);
/*
* we do not zeroize fragment, because of
* if it maped to hole, it already contains zeroes
*/
set_buffer_uptodate
(
bh
);
mark_buffer_dirty
(
bh
);
set_page_dirty
(
lastpage
);
}
if
(
lastfrag
>=
UFS_IND_FRAGMENT
)
{
end
=
uspi
->
s_fpb
-
ufs_fragnum
(
lastfrag
)
-
1
;
phys64
=
bh
->
b_blocknr
+
1
;
for
(
i
=
0
;
i
<
end
;
++
i
)
{
bh
=
sb_getblk
(
sb
,
i
+
phys64
);
lock_buffer
(
bh
);
memset
(
bh
->
b_data
,
0
,
sb
->
s_blocksize
);
set_buffer_uptodate
(
bh
);
mark_buffer_dirty
(
bh
);
unlock_buffer
(
bh
);
sync_dirty_buffer
(
bh
);
brelse
(
bh
);
}
}
out_unlock:
ufs_put_locked_page
(
lastpage
);
out:
return
err
;
}
static
void
__ufs_truncate_blocks
(
struct
inode
*
inode
)
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
unsigned
offsets
[
4
];
int
depth
=
ufs_block_to_path
(
inode
,
DIRECT_BLOCK
,
offsets
);
int
depth2
;
unsigned
i
;
struct
ufs_buffer_head
*
ubh
[
3
];
void
*
p
;
u64
block
;
if
(
!
depth
)
return
;
/* find the last non-zero in offsets[] */
for
(
depth2
=
depth
-
1
;
depth2
;
depth2
--
)
if
(
offsets
[
depth2
])
break
;
mutex_lock
(
&
ufsi
->
truncate_mutex
);
if
(
depth
==
1
)
{
ufs_trunc_direct
(
inode
);
offsets
[
0
]
=
UFS_IND_BLOCK
;
}
else
{
/* get the blocks that should be partially emptied */
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
offsets
[
0
]);
for
(
i
=
0
;
i
<
depth2
;
i
++
)
{
offsets
[
i
]
++
;
/* next branch is fully freed */
block
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
block
)
break
;
ubh
[
i
]
=
ubh_bread
(
sb
,
block
,
uspi
->
s_bsize
);
if
(
!
ubh
[
i
])
{
write_seqlock
(
&
ufsi
->
meta_lock
);
ufs_data_ptr_clear
(
uspi
,
p
);
write_sequnlock
(
&
ufsi
->
meta_lock
);
break
;
}
p
=
ubh_get_data_ptr
(
uspi
,
ubh
[
i
],
offsets
[
i
+
1
]);
}
while
(
i
--
)
free_branch_tail
(
inode
,
offsets
[
i
+
1
],
ubh
[
i
],
depth
-
i
-
1
);
}
for
(
i
=
offsets
[
0
];
i
<=
UFS_TIND_BLOCK
;
i
++
)
{
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
i
);
block
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
block
)
{
write_seqlock
(
&
ufsi
->
meta_lock
);
ufs_data_ptr_clear
(
uspi
,
p
);
write_sequnlock
(
&
ufsi
->
meta_lock
);
free_full_branch
(
inode
,
block
,
i
-
UFS_IND_BLOCK
+
1
);
}
}
}
ufsi
->
i_lastfrag
=
DIRECT_FRAGMENT
;
mark_inode_dirty
(
inode
);
mutex_unlock
(
&
ufsi
->
truncate_mutex
);
}
static
int
ufs_truncate
(
struct
inode
*
inode
,
loff_t
size
)
{
int
err
=
0
;
UFSD
(
"ENTER: ino %lu, i_size: %llu, old_i_size: %llu
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
size
,
(
unsigned
long
long
)
i_size_read
(
inode
));
if
(
!
(
S_ISREG
(
inode
->
i_mode
)
||
S_ISDIR
(
inode
->
i_mode
)
||
S_ISLNK
(
inode
->
i_mode
)))
return
-
EINVAL
;
if
(
IS_APPEND
(
inode
)
||
IS_IMMUTABLE
(
inode
))
return
-
EPERM
;
err
=
ufs_alloc_lastblock
(
inode
,
size
);
if
(
err
)
goto
out
;
block_truncate_page
(
inode
->
i_mapping
,
size
,
ufs_getfrag_block
);
truncate_setsize
(
inode
,
size
);
__ufs_truncate_blocks
(
inode
);
inode
->
i_mtime
=
inode
->
i_ctime
=
CURRENT_TIME_SEC
;
mark_inode_dirty
(
inode
);
out:
UFSD
(
"EXIT: err %d
\n
"
,
err
);
return
err
;
}
void
ufs_truncate_blocks
(
struct
inode
*
inode
)
{
if
(
!
(
S_ISREG
(
inode
->
i_mode
)
||
S_ISDIR
(
inode
->
i_mode
)
||
S_ISLNK
(
inode
->
i_mode
)))
return
;
if
(
IS_APPEND
(
inode
)
||
IS_IMMUTABLE
(
inode
))
return
;
__ufs_truncate_blocks
(
inode
);
}
int
ufs_setattr
(
struct
dentry
*
dentry
,
struct
iattr
*
attr
)
{
struct
inode
*
inode
=
d_inode
(
dentry
);
unsigned
int
ia_valid
=
attr
->
ia_valid
;
int
error
;
error
=
inode_change_ok
(
inode
,
attr
);
if
(
error
)
return
error
;
if
(
ia_valid
&
ATTR_SIZE
&&
attr
->
ia_size
!=
inode
->
i_size
)
{
error
=
ufs_truncate
(
inode
,
attr
->
ia_size
);
if
(
error
)
return
error
;
}
setattr_copy
(
inode
,
attr
);
mark_inode_dirty
(
inode
);
return
0
;
}
}
const
struct
inode_operations
ufs_file_inode_operations
=
{
.
setattr
=
ufs_setattr
,
};
fs/ufs/super.c
View file @
b5f5914c
...
@@ -94,22 +94,6 @@
...
@@ -94,22 +94,6 @@
#include "swab.h"
#include "swab.h"
#include "util.h"
#include "util.h"
void
lock_ufs
(
struct
super_block
*
sb
)
{
struct
ufs_sb_info
*
sbi
=
UFS_SB
(
sb
);
mutex_lock
(
&
sbi
->
mutex
);
sbi
->
mutex_owner
=
current
;
}
void
unlock_ufs
(
struct
super_block
*
sb
)
{
struct
ufs_sb_info
*
sbi
=
UFS_SB
(
sb
);
sbi
->
mutex_owner
=
NULL
;
mutex_unlock
(
&
sbi
->
mutex
);
}
static
struct
inode
*
ufs_nfs_get_inode
(
struct
super_block
*
sb
,
u64
ino
,
u32
generation
)
static
struct
inode
*
ufs_nfs_get_inode
(
struct
super_block
*
sb
,
u64
ino
,
u32
generation
)
{
{
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
...
@@ -694,7 +678,6 @@ static int ufs_sync_fs(struct super_block *sb, int wait)
...
@@ -694,7 +678,6 @@ static int ufs_sync_fs(struct super_block *sb, int wait)
struct
ufs_super_block_third
*
usb3
;
struct
ufs_super_block_third
*
usb3
;
unsigned
flags
;
unsigned
flags
;
lock_ufs
(
sb
);
mutex_lock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_lock
(
&
UFS_SB
(
sb
)
->
s_lock
);
UFSD
(
"ENTER
\n
"
);
UFSD
(
"ENTER
\n
"
);
...
@@ -714,7 +697,6 @@ static int ufs_sync_fs(struct super_block *sb, int wait)
...
@@ -714,7 +697,6 @@ static int ufs_sync_fs(struct super_block *sb, int wait)
UFSD
(
"EXIT
\n
"
);
UFSD
(
"EXIT
\n
"
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
0
;
return
0
;
}
}
...
@@ -758,7 +740,6 @@ static void ufs_put_super(struct super_block *sb)
...
@@ -758,7 +740,6 @@ static void ufs_put_super(struct super_block *sb)
ubh_brelse_uspi
(
sbi
->
s_uspi
);
ubh_brelse_uspi
(
sbi
->
s_uspi
);
kfree
(
sbi
->
s_uspi
);
kfree
(
sbi
->
s_uspi
);
mutex_destroy
(
&
sbi
->
mutex
);
kfree
(
sbi
);
kfree
(
sbi
);
sb
->
s_fs_info
=
NULL
;
sb
->
s_fs_info
=
NULL
;
UFSD
(
"EXIT
\n
"
);
UFSD
(
"EXIT
\n
"
);
...
@@ -801,7 +782,6 @@ static int ufs_fill_super(struct super_block *sb, void *data, int silent)
...
@@ -801,7 +782,6 @@ static int ufs_fill_super(struct super_block *sb, void *data, int silent)
UFSD
(
"flag %u
\n
"
,
(
int
)(
sb
->
s_flags
&
MS_RDONLY
));
UFSD
(
"flag %u
\n
"
,
(
int
)(
sb
->
s_flags
&
MS_RDONLY
));
mutex_init
(
&
sbi
->
mutex
);
mutex_init
(
&
sbi
->
s_lock
);
mutex_init
(
&
sbi
->
s_lock
);
spin_lock_init
(
&
sbi
->
work_lock
);
spin_lock_init
(
&
sbi
->
work_lock
);
INIT_DELAYED_WORK
(
&
sbi
->
sync_work
,
delayed_sync_fs
);
INIT_DELAYED_WORK
(
&
sbi
->
sync_work
,
delayed_sync_fs
);
...
@@ -1257,7 +1237,6 @@ static int ufs_fill_super(struct super_block *sb, void *data, int silent)
...
@@ -1257,7 +1237,6 @@ static int ufs_fill_super(struct super_block *sb, void *data, int silent)
return
0
;
return
0
;
failed:
failed:
mutex_destroy
(
&
sbi
->
mutex
);
if
(
ubh
)
if
(
ubh
)
ubh_brelse_uspi
(
uspi
);
ubh_brelse_uspi
(
uspi
);
kfree
(
uspi
);
kfree
(
uspi
);
...
@@ -1280,7 +1259,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
...
@@ -1280,7 +1259,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
unsigned
flags
;
unsigned
flags
;
sync_filesystem
(
sb
);
sync_filesystem
(
sb
);
lock_ufs
(
sb
);
mutex_lock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_lock
(
&
UFS_SB
(
sb
)
->
s_lock
);
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
flags
=
UFS_SB
(
sb
)
->
s_flags
;
flags
=
UFS_SB
(
sb
)
->
s_flags
;
...
@@ -1296,7 +1274,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
...
@@ -1296,7 +1274,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
ufs_set_opt
(
new_mount_opt
,
ONERROR_LOCK
);
ufs_set_opt
(
new_mount_opt
,
ONERROR_LOCK
);
if
(
!
ufs_parse_options
(
data
,
&
new_mount_opt
))
{
if
(
!
ufs_parse_options
(
data
,
&
new_mount_opt
))
{
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
-
EINVAL
;
return
-
EINVAL
;
}
}
if
(
!
(
new_mount_opt
&
UFS_MOUNT_UFSTYPE
))
{
if
(
!
(
new_mount_opt
&
UFS_MOUNT_UFSTYPE
))
{
...
@@ -1304,14 +1281,12 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
...
@@ -1304,14 +1281,12 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
}
else
if
((
new_mount_opt
&
UFS_MOUNT_UFSTYPE
)
!=
ufstype
)
{
}
else
if
((
new_mount_opt
&
UFS_MOUNT_UFSTYPE
)
!=
ufstype
)
{
pr_err
(
"ufstype can't be changed during remount
\n
"
);
pr_err
(
"ufstype can't be changed during remount
\n
"
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
-
EINVAL
;
return
-
EINVAL
;
}
}
if
((
*
mount_flags
&
MS_RDONLY
)
==
(
sb
->
s_flags
&
MS_RDONLY
))
{
if
((
*
mount_flags
&
MS_RDONLY
)
==
(
sb
->
s_flags
&
MS_RDONLY
))
{
UFS_SB
(
sb
)
->
s_mount_opt
=
new_mount_opt
;
UFS_SB
(
sb
)
->
s_mount_opt
=
new_mount_opt
;
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
0
;
return
0
;
}
}
...
@@ -1335,7 +1310,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
...
@@ -1335,7 +1310,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
#ifndef CONFIG_UFS_FS_WRITE
#ifndef CONFIG_UFS_FS_WRITE
pr_err
(
"ufs was compiled with read-only support, can't be mounted as read-write
\n
"
);
pr_err
(
"ufs was compiled with read-only support, can't be mounted as read-write
\n
"
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
-
EINVAL
;
return
-
EINVAL
;
#else
#else
if
(
ufstype
!=
UFS_MOUNT_UFSTYPE_SUN
&&
if
(
ufstype
!=
UFS_MOUNT_UFSTYPE_SUN
&&
...
@@ -1345,13 +1319,11 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
...
@@ -1345,13 +1319,11 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
ufstype
!=
UFS_MOUNT_UFSTYPE_UFS2
)
{
ufstype
!=
UFS_MOUNT_UFSTYPE_UFS2
)
{
pr_err
(
"this ufstype is read-only supported
\n
"
);
pr_err
(
"this ufstype is read-only supported
\n
"
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
-
EINVAL
;
return
-
EINVAL
;
}
}
if
(
!
ufs_read_cylinder_structures
(
sb
))
{
if
(
!
ufs_read_cylinder_structures
(
sb
))
{
pr_err
(
"failed during remounting
\n
"
);
pr_err
(
"failed during remounting
\n
"
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
-
EPERM
;
return
-
EPERM
;
}
}
sb
->
s_flags
&=
~
MS_RDONLY
;
sb
->
s_flags
&=
~
MS_RDONLY
;
...
@@ -1359,7 +1331,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
...
@@ -1359,7 +1331,6 @@ static int ufs_remount (struct super_block *sb, int *mount_flags, char *data)
}
}
UFS_SB
(
sb
)
->
s_mount_opt
=
new_mount_opt
;
UFS_SB
(
sb
)
->
s_mount_opt
=
new_mount_opt
;
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
unlock_ufs
(
sb
);
return
0
;
return
0
;
}
}
...
@@ -1391,8 +1362,7 @@ static int ufs_statfs(struct dentry *dentry, struct kstatfs *buf)
...
@@ -1391,8 +1362,7 @@ static int ufs_statfs(struct dentry *dentry, struct kstatfs *buf)
struct
ufs_super_block_third
*
usb3
;
struct
ufs_super_block_third
*
usb3
;
u64
id
=
huge_encode_dev
(
sb
->
s_bdev
->
bd_dev
);
u64
id
=
huge_encode_dev
(
sb
->
s_bdev
->
bd_dev
);
lock_ufs
(
sb
);
mutex_lock
(
&
UFS_SB
(
sb
)
->
s_lock
);
usb3
=
ubh_get_usb_third
(
uspi
);
usb3
=
ubh_get_usb_third
(
uspi
);
if
((
flags
&
UFS_TYPE_MASK
)
==
UFS_TYPE_UFS2
)
{
if
((
flags
&
UFS_TYPE_MASK
)
==
UFS_TYPE_UFS2
)
{
...
@@ -1413,7 +1383,7 @@ static int ufs_statfs(struct dentry *dentry, struct kstatfs *buf)
...
@@ -1413,7 +1383,7 @@ static int ufs_statfs(struct dentry *dentry, struct kstatfs *buf)
buf
->
f_fsid
.
val
[
0
]
=
(
u32
)
id
;
buf
->
f_fsid
.
val
[
0
]
=
(
u32
)
id
;
buf
->
f_fsid
.
val
[
1
]
=
(
u32
)(
id
>>
32
);
buf
->
f_fsid
.
val
[
1
]
=
(
u32
)(
id
>>
32
);
unlock_ufs
(
sb
);
mutex_unlock
(
&
UFS_SB
(
sb
)
->
s_lock
);
return
0
;
return
0
;
}
}
...
@@ -1429,6 +1399,8 @@ static struct inode *ufs_alloc_inode(struct super_block *sb)
...
@@ -1429,6 +1399,8 @@ static struct inode *ufs_alloc_inode(struct super_block *sb)
return
NULL
;
return
NULL
;
ei
->
vfs_inode
.
i_version
=
1
;
ei
->
vfs_inode
.
i_version
=
1
;
seqlock_init
(
&
ei
->
meta_lock
);
mutex_init
(
&
ei
->
truncate_mutex
);
return
&
ei
->
vfs_inode
;
return
&
ei
->
vfs_inode
;
}
}
...
...
fs/ufs/truncate.c
deleted
100644 → 0
View file @
15cf3b7a
/*
* linux/fs/ufs/truncate.c
*
* Copyright (C) 1998
* Daniel Pirkl <daniel.pirkl@email.cz>
* Charles University, Faculty of Mathematics and Physics
*
* from
*
* linux/fs/ext2/truncate.c
*
* Copyright (C) 1992, 1993, 1994, 1995
* Remy Card (card@masi.ibp.fr)
* Laboratoire MASI - Institut Blaise Pascal
* Universite Pierre et Marie Curie (Paris VI)
*
* from
*
* linux/fs/minix/truncate.c
*
* Copyright (C) 1991, 1992 Linus Torvalds
*
* Big-endian to little-endian byte-swapping/bitmaps by
* David S. Miller (davem@caip.rutgers.edu), 1995
*/
/*
* Real random numbers for secure rm added 94/02/18
* Idea from Pierre del Perugia <delperug@gla.ecoledoc.ibp.fr>
*/
/*
* Adoptation to use page cache and UFS2 write support by
* Evgeniy Dushistov <dushistov@mail.ru>, 2006-2007
*/
#include <linux/errno.h>
#include <linux/fs.h>
#include <linux/fcntl.h>
#include <linux/time.h>
#include <linux/stat.h>
#include <linux/string.h>
#include <linux/buffer_head.h>
#include <linux/blkdev.h>
#include <linux/sched.h>
#include "ufs_fs.h"
#include "ufs.h"
#include "swab.h"
#include "util.h"
/*
* Secure deletion currently doesn't work. It interacts very badly
* with buffers shared with memory mappings, and for that reason
* can't be done in the truncate() routines. It should instead be
* done separately in "release()" before calling the truncate routines
* that will release the actual file blocks.
*
* Linus
*/
#define DIRECT_BLOCK ((inode->i_size + uspi->s_bsize - 1) >> uspi->s_bshift)
#define DIRECT_FRAGMENT ((inode->i_size + uspi->s_fsize - 1) >> uspi->s_fshift)
static
int
ufs_trunc_direct
(
struct
inode
*
inode
)
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
;
struct
ufs_sb_private_info
*
uspi
;
void
*
p
;
u64
frag1
,
frag2
,
frag3
,
frag4
,
block1
,
block2
;
unsigned
frag_to_free
,
free_count
;
unsigned
i
,
tmp
;
int
retry
;
UFSD
(
"ENTER: ino %lu
\n
"
,
inode
->
i_ino
);
sb
=
inode
->
i_sb
;
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
frag_to_free
=
0
;
free_count
=
0
;
retry
=
0
;
frag1
=
DIRECT_FRAGMENT
;
frag4
=
min_t
(
u64
,
UFS_NDIR_FRAGMENT
,
ufsi
->
i_lastfrag
);
frag2
=
((
frag1
&
uspi
->
s_fpbmask
)
?
((
frag1
|
uspi
->
s_fpbmask
)
+
1
)
:
frag1
);
frag3
=
frag4
&
~
uspi
->
s_fpbmask
;
block1
=
block2
=
0
;
if
(
frag2
>
frag3
)
{
frag2
=
frag4
;
frag3
=
frag4
=
0
;
}
else
if
(
frag2
<
frag3
)
{
block1
=
ufs_fragstoblks
(
frag2
);
block2
=
ufs_fragstoblks
(
frag3
);
}
UFSD
(
"ino %lu, frag1 %llu, frag2 %llu, block1 %llu, block2 %llu,"
" frag3 %llu, frag4 %llu
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
frag1
,
(
unsigned
long
long
)
frag2
,
(
unsigned
long
long
)
block1
,
(
unsigned
long
long
)
block2
,
(
unsigned
long
long
)
frag3
,
(
unsigned
long
long
)
frag4
);
if
(
frag1
>=
frag2
)
goto
next1
;
/*
* Free first free fragments
*/
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
ufs_fragstoblks
(
frag1
));
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
ufs_panic
(
sb
,
"ufs_trunc_direct"
,
"internal error"
);
frag2
-=
frag1
;
frag1
=
ufs_fragnum
(
frag1
);
ufs_free_fragments
(
inode
,
tmp
+
frag1
,
frag2
);
mark_inode_dirty
(
inode
);
frag_to_free
=
tmp
+
frag1
;
next1:
/*
* Free whole blocks
*/
for
(
i
=
block1
;
i
<
block2
;
i
++
)
{
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
i
);
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
continue
;
ufs_data_ptr_clear
(
uspi
,
p
);
if
(
free_count
==
0
)
{
frag_to_free
=
tmp
;
free_count
=
uspi
->
s_fpb
;
}
else
if
(
free_count
>
0
&&
frag_to_free
==
tmp
-
free_count
)
free_count
+=
uspi
->
s_fpb
;
else
{
ufs_free_blocks
(
inode
,
frag_to_free
,
free_count
);
frag_to_free
=
tmp
;
free_count
=
uspi
->
s_fpb
;
}
mark_inode_dirty
(
inode
);
}
if
(
free_count
>
0
)
ufs_free_blocks
(
inode
,
frag_to_free
,
free_count
);
if
(
frag3
>=
frag4
)
goto
next3
;
/*
* Free last free fragments
*/
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
ufs_fragstoblks
(
frag3
));
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
ufs_panic
(
sb
,
"ufs_truncate_direct"
,
"internal error"
);
frag4
=
ufs_fragnum
(
frag4
);
ufs_data_ptr_clear
(
uspi
,
p
);
ufs_free_fragments
(
inode
,
tmp
,
frag4
);
mark_inode_dirty
(
inode
);
next3:
UFSD
(
"EXIT: ino %lu
\n
"
,
inode
->
i_ino
);
return
retry
;
}
static
int
ufs_trunc_indirect
(
struct
inode
*
inode
,
u64
offset
,
void
*
p
)
{
struct
super_block
*
sb
;
struct
ufs_sb_private_info
*
uspi
;
struct
ufs_buffer_head
*
ind_ubh
;
void
*
ind
;
u64
tmp
,
indirect_block
,
i
,
frag_to_free
;
unsigned
free_count
;
int
retry
;
UFSD
(
"ENTER: ino %lu, offset %llu, p: %p
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
offset
,
p
);
BUG_ON
(
!
p
);
sb
=
inode
->
i_sb
;
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
frag_to_free
=
0
;
free_count
=
0
;
retry
=
0
;
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
return
0
;
ind_ubh
=
ubh_bread
(
sb
,
tmp
,
uspi
->
s_bsize
);
if
(
tmp
!=
ufs_data_ptr_to_cpu
(
sb
,
p
))
{
ubh_brelse
(
ind_ubh
);
return
1
;
}
if
(
!
ind_ubh
)
{
ufs_data_ptr_clear
(
uspi
,
p
);
return
0
;
}
indirect_block
=
(
DIRECT_BLOCK
>
offset
)
?
(
DIRECT_BLOCK
-
offset
)
:
0
;
for
(
i
=
indirect_block
;
i
<
uspi
->
s_apb
;
i
++
)
{
ind
=
ubh_get_data_ptr
(
uspi
,
ind_ubh
,
i
);
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
ind
);
if
(
!
tmp
)
continue
;
ufs_data_ptr_clear
(
uspi
,
ind
);
ubh_mark_buffer_dirty
(
ind_ubh
);
if
(
free_count
==
0
)
{
frag_to_free
=
tmp
;
free_count
=
uspi
->
s_fpb
;
}
else
if
(
free_count
>
0
&&
frag_to_free
==
tmp
-
free_count
)
free_count
+=
uspi
->
s_fpb
;
else
{
ufs_free_blocks
(
inode
,
frag_to_free
,
free_count
);
frag_to_free
=
tmp
;
free_count
=
uspi
->
s_fpb
;
}
mark_inode_dirty
(
inode
);
}
if
(
free_count
>
0
)
{
ufs_free_blocks
(
inode
,
frag_to_free
,
free_count
);
}
for
(
i
=
0
;
i
<
uspi
->
s_apb
;
i
++
)
if
(
!
ufs_is_data_ptr_zero
(
uspi
,
ubh_get_data_ptr
(
uspi
,
ind_ubh
,
i
)))
break
;
if
(
i
>=
uspi
->
s_apb
)
{
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
ufs_data_ptr_clear
(
uspi
,
p
);
ufs_free_blocks
(
inode
,
tmp
,
uspi
->
s_fpb
);
mark_inode_dirty
(
inode
);
ubh_bforget
(
ind_ubh
);
ind_ubh
=
NULL
;
}
if
(
IS_SYNC
(
inode
)
&&
ind_ubh
&&
ubh_buffer_dirty
(
ind_ubh
))
ubh_sync_block
(
ind_ubh
);
ubh_brelse
(
ind_ubh
);
UFSD
(
"EXIT: ino %lu
\n
"
,
inode
->
i_ino
);
return
retry
;
}
static
int
ufs_trunc_dindirect
(
struct
inode
*
inode
,
u64
offset
,
void
*
p
)
{
struct
super_block
*
sb
;
struct
ufs_sb_private_info
*
uspi
;
struct
ufs_buffer_head
*
dind_bh
;
u64
i
,
tmp
,
dindirect_block
;
void
*
dind
;
int
retry
=
0
;
UFSD
(
"ENTER: ino %lu
\n
"
,
inode
->
i_ino
);
sb
=
inode
->
i_sb
;
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
dindirect_block
=
(
DIRECT_BLOCK
>
offset
)
?
((
DIRECT_BLOCK
-
offset
)
>>
uspi
->
s_apbshift
)
:
0
;
retry
=
0
;
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
if
(
!
tmp
)
return
0
;
dind_bh
=
ubh_bread
(
sb
,
tmp
,
uspi
->
s_bsize
);
if
(
tmp
!=
ufs_data_ptr_to_cpu
(
sb
,
p
))
{
ubh_brelse
(
dind_bh
);
return
1
;
}
if
(
!
dind_bh
)
{
ufs_data_ptr_clear
(
uspi
,
p
);
return
0
;
}
for
(
i
=
dindirect_block
;
i
<
uspi
->
s_apb
;
i
++
)
{
dind
=
ubh_get_data_ptr
(
uspi
,
dind_bh
,
i
);
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
dind
);
if
(
!
tmp
)
continue
;
retry
|=
ufs_trunc_indirect
(
inode
,
offset
+
(
i
<<
uspi
->
s_apbshift
),
dind
);
ubh_mark_buffer_dirty
(
dind_bh
);
}
for
(
i
=
0
;
i
<
uspi
->
s_apb
;
i
++
)
if
(
!
ufs_is_data_ptr_zero
(
uspi
,
ubh_get_data_ptr
(
uspi
,
dind_bh
,
i
)))
break
;
if
(
i
>=
uspi
->
s_apb
)
{
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
ufs_data_ptr_clear
(
uspi
,
p
);
ufs_free_blocks
(
inode
,
tmp
,
uspi
->
s_fpb
);
mark_inode_dirty
(
inode
);
ubh_bforget
(
dind_bh
);
dind_bh
=
NULL
;
}
if
(
IS_SYNC
(
inode
)
&&
dind_bh
&&
ubh_buffer_dirty
(
dind_bh
))
ubh_sync_block
(
dind_bh
);
ubh_brelse
(
dind_bh
);
UFSD
(
"EXIT: ino %lu
\n
"
,
inode
->
i_ino
);
return
retry
;
}
static
int
ufs_trunc_tindirect
(
struct
inode
*
inode
)
{
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
ufs_buffer_head
*
tind_bh
;
u64
tindirect_block
,
tmp
,
i
;
void
*
tind
,
*
p
;
int
retry
;
UFSD
(
"ENTER: ino %lu
\n
"
,
inode
->
i_ino
);
retry
=
0
;
tindirect_block
=
(
DIRECT_BLOCK
>
(
UFS_NDADDR
+
uspi
->
s_apb
+
uspi
->
s_2apb
))
?
((
DIRECT_BLOCK
-
UFS_NDADDR
-
uspi
->
s_apb
-
uspi
->
s_2apb
)
>>
uspi
->
s_2apbshift
)
:
0
;
p
=
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
UFS_TIND_BLOCK
);
if
(
!
(
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
)))
return
0
;
tind_bh
=
ubh_bread
(
sb
,
tmp
,
uspi
->
s_bsize
);
if
(
tmp
!=
ufs_data_ptr_to_cpu
(
sb
,
p
))
{
ubh_brelse
(
tind_bh
);
return
1
;
}
if
(
!
tind_bh
)
{
ufs_data_ptr_clear
(
uspi
,
p
);
return
0
;
}
for
(
i
=
tindirect_block
;
i
<
uspi
->
s_apb
;
i
++
)
{
tind
=
ubh_get_data_ptr
(
uspi
,
tind_bh
,
i
);
retry
|=
ufs_trunc_dindirect
(
inode
,
UFS_NDADDR
+
uspi
->
s_apb
+
((
i
+
1
)
<<
uspi
->
s_2apbshift
),
tind
);
ubh_mark_buffer_dirty
(
tind_bh
);
}
for
(
i
=
0
;
i
<
uspi
->
s_apb
;
i
++
)
if
(
!
ufs_is_data_ptr_zero
(
uspi
,
ubh_get_data_ptr
(
uspi
,
tind_bh
,
i
)))
break
;
if
(
i
>=
uspi
->
s_apb
)
{
tmp
=
ufs_data_ptr_to_cpu
(
sb
,
p
);
ufs_data_ptr_clear
(
uspi
,
p
);
ufs_free_blocks
(
inode
,
tmp
,
uspi
->
s_fpb
);
mark_inode_dirty
(
inode
);
ubh_bforget
(
tind_bh
);
tind_bh
=
NULL
;
}
if
(
IS_SYNC
(
inode
)
&&
tind_bh
&&
ubh_buffer_dirty
(
tind_bh
))
ubh_sync_block
(
tind_bh
);
ubh_brelse
(
tind_bh
);
UFSD
(
"EXIT: ino %lu
\n
"
,
inode
->
i_ino
);
return
retry
;
}
static
int
ufs_alloc_lastblock
(
struct
inode
*
inode
)
{
int
err
=
0
;
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
address_space
*
mapping
=
inode
->
i_mapping
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
unsigned
i
,
end
;
sector_t
lastfrag
;
struct
page
*
lastpage
;
struct
buffer_head
*
bh
;
u64
phys64
;
lastfrag
=
(
i_size_read
(
inode
)
+
uspi
->
s_fsize
-
1
)
>>
uspi
->
s_fshift
;
if
(
!
lastfrag
)
goto
out
;
lastfrag
--
;
lastpage
=
ufs_get_locked_page
(
mapping
,
lastfrag
>>
(
PAGE_CACHE_SHIFT
-
inode
->
i_blkbits
));
if
(
IS_ERR
(
lastpage
))
{
err
=
-
EIO
;
goto
out
;
}
end
=
lastfrag
&
((
1
<<
(
PAGE_CACHE_SHIFT
-
inode
->
i_blkbits
))
-
1
);
bh
=
page_buffers
(
lastpage
);
for
(
i
=
0
;
i
<
end
;
++
i
)
bh
=
bh
->
b_this_page
;
err
=
ufs_getfrag_block
(
inode
,
lastfrag
,
bh
,
1
);
if
(
unlikely
(
err
))
goto
out_unlock
;
if
(
buffer_new
(
bh
))
{
clear_buffer_new
(
bh
);
unmap_underlying_metadata
(
bh
->
b_bdev
,
bh
->
b_blocknr
);
/*
* we do not zeroize fragment, because of
* if it maped to hole, it already contains zeroes
*/
set_buffer_uptodate
(
bh
);
mark_buffer_dirty
(
bh
);
set_page_dirty
(
lastpage
);
}
if
(
lastfrag
>=
UFS_IND_FRAGMENT
)
{
end
=
uspi
->
s_fpb
-
ufs_fragnum
(
lastfrag
)
-
1
;
phys64
=
bh
->
b_blocknr
+
1
;
for
(
i
=
0
;
i
<
end
;
++
i
)
{
bh
=
sb_getblk
(
sb
,
i
+
phys64
);
lock_buffer
(
bh
);
memset
(
bh
->
b_data
,
0
,
sb
->
s_blocksize
);
set_buffer_uptodate
(
bh
);
mark_buffer_dirty
(
bh
);
unlock_buffer
(
bh
);
sync_dirty_buffer
(
bh
);
brelse
(
bh
);
}
}
out_unlock:
ufs_put_locked_page
(
lastpage
);
out:
return
err
;
}
int
ufs_truncate
(
struct
inode
*
inode
,
loff_t
old_i_size
)
{
struct
ufs_inode_info
*
ufsi
=
UFS_I
(
inode
);
struct
super_block
*
sb
=
inode
->
i_sb
;
struct
ufs_sb_private_info
*
uspi
=
UFS_SB
(
sb
)
->
s_uspi
;
int
retry
,
err
=
0
;
UFSD
(
"ENTER: ino %lu, i_size: %llu, old_i_size: %llu
\n
"
,
inode
->
i_ino
,
(
unsigned
long
long
)
i_size_read
(
inode
),
(
unsigned
long
long
)
old_i_size
);
if
(
!
(
S_ISREG
(
inode
->
i_mode
)
||
S_ISDIR
(
inode
->
i_mode
)
||
S_ISLNK
(
inode
->
i_mode
)))
return
-
EINVAL
;
if
(
IS_APPEND
(
inode
)
||
IS_IMMUTABLE
(
inode
))
return
-
EPERM
;
err
=
ufs_alloc_lastblock
(
inode
);
if
(
err
)
{
i_size_write
(
inode
,
old_i_size
);
goto
out
;
}
block_truncate_page
(
inode
->
i_mapping
,
inode
->
i_size
,
ufs_getfrag_block
);
while
(
1
)
{
retry
=
ufs_trunc_direct
(
inode
);
retry
|=
ufs_trunc_indirect
(
inode
,
UFS_IND_BLOCK
,
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
UFS_IND_BLOCK
));
retry
|=
ufs_trunc_dindirect
(
inode
,
UFS_IND_BLOCK
+
uspi
->
s_apb
,
ufs_get_direct_data_ptr
(
uspi
,
ufsi
,
UFS_DIND_BLOCK
));
retry
|=
ufs_trunc_tindirect
(
inode
);
if
(
!
retry
)
break
;
if
(
IS_SYNC
(
inode
)
&&
(
inode
->
i_state
&
I_DIRTY
))
ufs_sync_inode
(
inode
);
yield
();
}
inode
->
i_mtime
=
inode
->
i_ctime
=
CURRENT_TIME_SEC
;
ufsi
->
i_lastfrag
=
DIRECT_FRAGMENT
;
mark_inode_dirty
(
inode
);
out:
UFSD
(
"EXIT: err %d
\n
"
,
err
);
return
err
;
}
int
ufs_setattr
(
struct
dentry
*
dentry
,
struct
iattr
*
attr
)
{
struct
inode
*
inode
=
d_inode
(
dentry
);
unsigned
int
ia_valid
=
attr
->
ia_valid
;
int
error
;
error
=
inode_change_ok
(
inode
,
attr
);
if
(
error
)
return
error
;
if
(
ia_valid
&
ATTR_SIZE
&&
attr
->
ia_size
!=
inode
->
i_size
)
{
loff_t
old_i_size
=
inode
->
i_size
;
/* XXX(truncate): truncate_setsize should be called last */
truncate_setsize
(
inode
,
attr
->
ia_size
);
lock_ufs
(
inode
->
i_sb
);
error
=
ufs_truncate
(
inode
,
old_i_size
);
unlock_ufs
(
inode
->
i_sb
);
if
(
error
)
return
error
;
}
setattr_copy
(
inode
,
attr
);
mark_inode_dirty
(
inode
);
return
0
;
}
const
struct
inode_operations
ufs_file_inode_operations
=
{
.
setattr
=
ufs_setattr
,
};
fs/ufs/ufs.h
View file @
b5f5914c
...
@@ -24,8 +24,6 @@ struct ufs_sb_info {
...
@@ -24,8 +24,6 @@ struct ufs_sb_info {
unsigned
s_cgno
[
UFS_MAX_GROUP_LOADED
];
unsigned
s_cgno
[
UFS_MAX_GROUP_LOADED
];
unsigned
short
s_cg_loaded
;
unsigned
short
s_cg_loaded
;
unsigned
s_mount_opt
;
unsigned
s_mount_opt
;
struct
mutex
mutex
;
struct
task_struct
*
mutex_owner
;
struct
super_block
*
sb
;
struct
super_block
*
sb
;
int
work_queued
;
/* non-zero if the delayed work is queued */
int
work_queued
;
/* non-zero if the delayed work is queued */
struct
delayed_work
sync_work
;
/* FS sync delayed work */
struct
delayed_work
sync_work
;
/* FS sync delayed work */
...
@@ -46,6 +44,8 @@ struct ufs_inode_info {
...
@@ -46,6 +44,8 @@ struct ufs_inode_info {
__u32
i_oeftflag
;
__u32
i_oeftflag
;
__u16
i_osync
;
__u16
i_osync
;
__u64
i_lastfrag
;
__u64
i_lastfrag
;
seqlock_t
meta_lock
;
struct
mutex
truncate_mutex
;
__u32
i_dir_start_lookup
;
__u32
i_dir_start_lookup
;
struct
inode
vfs_inode
;
struct
inode
vfs_inode
;
};
};
...
@@ -122,7 +122,7 @@ extern struct inode *ufs_iget(struct super_block *, unsigned long);
...
@@ -122,7 +122,7 @@ extern struct inode *ufs_iget(struct super_block *, unsigned long);
extern
int
ufs_write_inode
(
struct
inode
*
,
struct
writeback_control
*
);
extern
int
ufs_write_inode
(
struct
inode
*
,
struct
writeback_control
*
);
extern
int
ufs_sync_inode
(
struct
inode
*
);
extern
int
ufs_sync_inode
(
struct
inode
*
);
extern
void
ufs_evict_inode
(
struct
inode
*
);
extern
void
ufs_evict_inode
(
struct
inode
*
);
extern
int
ufs_
getfrag_block
(
struct
inode
*
inode
,
sector_t
fragment
,
struct
buffer_head
*
bh_result
,
int
create
);
extern
int
ufs_
setattr
(
struct
dentry
*
dentry
,
struct
iattr
*
attr
);
/* namei.c */
/* namei.c */
extern
const
struct
file_operations
ufs_dir_operations
;
extern
const
struct
file_operations
ufs_dir_operations
;
...
@@ -140,10 +140,6 @@ void ufs_mark_sb_dirty(struct super_block *sb);
...
@@ -140,10 +140,6 @@ void ufs_mark_sb_dirty(struct super_block *sb);
extern
const
struct
inode_operations
ufs_fast_symlink_inode_operations
;
extern
const
struct
inode_operations
ufs_fast_symlink_inode_operations
;
extern
const
struct
inode_operations
ufs_symlink_inode_operations
;
extern
const
struct
inode_operations
ufs_symlink_inode_operations
;
/* truncate.c */
extern
int
ufs_truncate
(
struct
inode
*
,
loff_t
);
extern
int
ufs_setattr
(
struct
dentry
*
dentry
,
struct
iattr
*
attr
);
static
inline
struct
ufs_sb_info
*
UFS_SB
(
struct
super_block
*
sb
)
static
inline
struct
ufs_sb_info
*
UFS_SB
(
struct
super_block
*
sb
)
{
{
return
sb
->
s_fs_info
;
return
sb
->
s_fs_info
;
...
@@ -170,7 +166,4 @@ static inline u32 ufs_dtogd(struct ufs_sb_private_info * uspi, u64 b)
...
@@ -170,7 +166,4 @@ static inline u32 ufs_dtogd(struct ufs_sb_private_info * uspi, u64 b)
return
do_div
(
b
,
uspi
->
s_fpg
);
return
do_div
(
b
,
uspi
->
s_fpg
);
}
}
extern
void
lock_ufs
(
struct
super_block
*
sb
);
extern
void
unlock_ufs
(
struct
super_block
*
sb
);
#endif
/* _UFS_UFS_H */
#endif
/* _UFS_UFS_H */
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment