blob: d10a5dfe7e288be574abfe63133c6573b1ea32df [file] [log] [blame]
/*
cluster.c (03.09.09)
exFAT file system implementation library.
Copyright (C) 2010-2013 Andrew Nayenko
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
#include "exfat.h"
#include <errno.h>
#include <string.h>
/*
* Sector to absolute offset.
*/
static off64_t s2o(const struct exfat* ef, off64_t sector)
{
return sector << ef->sb->sector_bits;
}
/*
* Cluster to sector.
*/
static off64_t c2s(const struct exfat* ef, cluster_t cluster)
{
if (cluster < EXFAT_FIRST_DATA_CLUSTER)
exfat_bug("invalid cluster number %u", cluster);
return le32_to_cpu(ef->sb->cluster_sector_start) +
((off64_t) (cluster - EXFAT_FIRST_DATA_CLUSTER) << ef->sb->spc_bits);
}
/*
* Cluster to absolute offset.
*/
off64_t exfat_c2o(const struct exfat* ef, cluster_t cluster)
{
return s2o(ef, c2s(ef, cluster));
}
/*
* Sector to cluster.
*/
static cluster_t s2c(const struct exfat* ef, off64_t sector)
{
return ((sector - le32_to_cpu(ef->sb->cluster_sector_start)) >>
ef->sb->spc_bits) + EXFAT_FIRST_DATA_CLUSTER;
}
/*
* Size in bytes to size in clusters (rounded upwards).
*/
static uint32_t bytes2clusters(const struct exfat* ef, uint64_t bytes)
{
uint64_t cluster_size = CLUSTER_SIZE(*ef->sb);
return (bytes + cluster_size - 1) / cluster_size;
}
cluster_t exfat_next_cluster(const struct exfat* ef,
const struct exfat_node* node, cluster_t cluster)
{
le32_t next;
off64_t fat_offset;
if (cluster < EXFAT_FIRST_DATA_CLUSTER)
exfat_bug("bad cluster 0x%x", cluster);
if (IS_CONTIGUOUS(*node))
return cluster + 1;
fat_offset = s2o(ef, le32_to_cpu(ef->sb->fat_sector_start))
+ cluster * sizeof(cluster_t);
exfat_pread(ef->dev, &next, sizeof(next), fat_offset);
return le32_to_cpu(next);
}
cluster_t exfat_advance_cluster(const struct exfat* ef,
struct exfat_node* node, uint32_t count)
{
uint32_t i;
if (node->fptr_index > count)
{
node->fptr_index = 0;
node->fptr_cluster = node->start_cluster;
}
for (i = node->fptr_index; i < count; i++)
{
node->fptr_cluster = exfat_next_cluster(ef, node, node->fptr_cluster);
if (CLUSTER_INVALID(node->fptr_cluster))
break; /* the caller should handle this and print appropriate
error message */
}
node->fptr_index = count;
return node->fptr_cluster;
}
static cluster_t find_bit_and_set(uint8_t* bitmap, size_t start, size_t end)
{
const size_t start_index = start / 8;
const size_t end_index = DIV_ROUND_UP(end, 8);
size_t i;
size_t c;
for (i = start_index; i < end_index; i++)
{
if (bitmap[i] == 0xff)
continue;
for (c = MAX(i * 8, start); c < MIN((i + 1) * 8, end); c++)
if (BMAP_GET(bitmap, c) == 0)
{
BMAP_SET(bitmap, c);
return c + EXFAT_FIRST_DATA_CLUSTER;
}
}
return EXFAT_CLUSTER_END;
}
void exfat_flush_cmap(struct exfat* ef)
{
exfat_pwrite(ef->dev, ef->cmap.chunk, (ef->cmap.chunk_size + 7) / 8,
exfat_c2o(ef, ef->cmap.start_cluster));
ef->cmap.dirty = false;
}
static void set_next_cluster(const struct exfat* ef, bool contiguous,
cluster_t current, cluster_t next)
{
off64_t fat_offset;
le32_t next_le32;
if (contiguous)
return;
fat_offset = s2o(ef, le32_to_cpu(ef->sb->fat_sector_start))
+ current * sizeof(cluster_t);
next_le32 = cpu_to_le32(next);
exfat_pwrite(ef->dev, &next_le32, sizeof(next_le32), fat_offset);
}
static cluster_t allocate_cluster(struct exfat* ef, cluster_t hint)
{
cluster_t cluster;
hint -= EXFAT_FIRST_DATA_CLUSTER;
if (hint >= ef->cmap.chunk_size)
hint = 0;
cluster = find_bit_and_set(ef->cmap.chunk, hint, ef->cmap.chunk_size);
if (cluster == EXFAT_CLUSTER_END)
cluster = find_bit_and_set(ef->cmap.chunk, 0, hint);
if (cluster == EXFAT_CLUSTER_END)
{
exfat_error("no free space left");
return EXFAT_CLUSTER_END;
}
ef->cmap.dirty = true;
return cluster;
}
static void free_cluster(struct exfat* ef, cluster_t cluster)
{
if (CLUSTER_INVALID(cluster))
exfat_bug("freeing invalid cluster 0x%x", cluster);
if (cluster - EXFAT_FIRST_DATA_CLUSTER >= ef->cmap.size)
exfat_bug("freeing non-existing cluster 0x%x (0x%x)", cluster,
ef->cmap.size);
BMAP_CLR(ef->cmap.chunk, cluster - EXFAT_FIRST_DATA_CLUSTER);
ef->cmap.dirty = true;
}
static void make_noncontiguous(const struct exfat* ef, cluster_t first,
cluster_t last)
{
cluster_t c;
for (c = first; c < last; c++)
set_next_cluster(ef, false, c, c + 1);
}
static int shrink_file(struct exfat* ef, struct exfat_node* node,
uint32_t current, uint32_t difference);
static int grow_file(struct exfat* ef, struct exfat_node* node,
uint32_t current, uint32_t difference)
{
cluster_t previous;
cluster_t next;
uint32_t allocated = 0;
if (difference == 0)
exfat_bug("zero clusters count passed");
if (node->start_cluster != EXFAT_CLUSTER_FREE)
{
/* get the last cluster of the file */
previous = exfat_advance_cluster(ef, node, current - 1);
if (CLUSTER_INVALID(previous))
{
exfat_error("invalid cluster 0x%x while growing", previous);
return -EIO;
}
}
else
{
if (node->fptr_index != 0)
exfat_bug("non-zero pointer index (%u)", node->fptr_index);
/* file does not have clusters (i.e. is empty), allocate
the first one for it */
previous = allocate_cluster(ef, 0);
if (CLUSTER_INVALID(previous))
return -ENOSPC;
node->fptr_cluster = node->start_cluster = previous;
allocated = 1;
/* file consists of only one cluster, so it's contiguous */
node->flags |= EXFAT_ATTRIB_CONTIGUOUS;
}
while (allocated < difference)
{
next = allocate_cluster(ef, previous + 1);
if (CLUSTER_INVALID(next))
{
if (allocated != 0)
shrink_file(ef, node, current + allocated, allocated);
return -ENOSPC;
}
if (next != previous - 1 && IS_CONTIGUOUS(*node))
{
/* it's a pity, but we are not able to keep the file contiguous
anymore */
make_noncontiguous(ef, node->start_cluster, previous);
node->flags &= ~EXFAT_ATTRIB_CONTIGUOUS;
node->flags |= EXFAT_ATTRIB_DIRTY;
}
set_next_cluster(ef, IS_CONTIGUOUS(*node), previous, next);
previous = next;
allocated++;
}
set_next_cluster(ef, IS_CONTIGUOUS(*node), previous, EXFAT_CLUSTER_END);
return 0;
}
static int shrink_file(struct exfat* ef, struct exfat_node* node,
uint32_t current, uint32_t difference)
{
cluster_t previous;
cluster_t next;
if (difference == 0)
exfat_bug("zero difference passed");
if (node->start_cluster == EXFAT_CLUSTER_FREE)
exfat_bug("unable to shrink empty file (%u clusters)", current);
if (current < difference)
exfat_bug("file underflow (%u < %u)", current, difference);
/* crop the file */
if (current > difference)
{
cluster_t last = exfat_advance_cluster(ef, node,
current - difference - 1);
if (CLUSTER_INVALID(last))
{
exfat_error("invalid cluster 0x%x while shrinking", last);
return -EIO;
}
previous = exfat_next_cluster(ef, node, last);
set_next_cluster(ef, IS_CONTIGUOUS(*node), last, EXFAT_CLUSTER_END);
}
else
{
previous = node->start_cluster;
node->start_cluster = EXFAT_CLUSTER_FREE;
}
node->fptr_index = 0;
node->fptr_cluster = node->start_cluster;
/* free remaining clusters */
while (difference--)
{
if (CLUSTER_INVALID(previous))
{
exfat_error("invalid cluster 0x%x while freeing after shrink",
previous);
return -EIO;
}
next = exfat_next_cluster(ef, node, previous);
set_next_cluster(ef, IS_CONTIGUOUS(*node), previous,
EXFAT_CLUSTER_FREE);
free_cluster(ef, previous);
previous = next;
}
return 0;
}
static void erase_raw(struct exfat* ef, size_t size, off64_t offset)
{
exfat_pwrite(ef->dev, ef->zero_cluster, size, offset);
}
static int erase_range(struct exfat* ef, struct exfat_node* node,
uint64_t begin, uint64_t end)
{
uint64_t cluster_boundary;
cluster_t cluster;
if (begin >= end)
return 0;
cluster_boundary = (begin | (CLUSTER_SIZE(*ef->sb) - 1)) + 1;
cluster = exfat_advance_cluster(ef, node,
begin / CLUSTER_SIZE(*ef->sb));
if (CLUSTER_INVALID(cluster))
{
exfat_error("invalid cluster 0x%x while erasing", cluster);
return -EIO;
}
/* erase from the beginning to the closest cluster boundary */
erase_raw(ef, MIN(cluster_boundary, end) - begin,
exfat_c2o(ef, cluster) + begin % CLUSTER_SIZE(*ef->sb));
/* erase whole clusters */
while (cluster_boundary < end)
{
cluster = exfat_next_cluster(ef, node, cluster);
/* the cluster cannot be invalid because we have just allocated it */
if (CLUSTER_INVALID(cluster))
exfat_bug("invalid cluster 0x%x after allocation", cluster);
erase_raw(ef, CLUSTER_SIZE(*ef->sb), exfat_c2o(ef, cluster));
cluster_boundary += CLUSTER_SIZE(*ef->sb);
}
return 0;
}
int exfat_truncate(struct exfat* ef, struct exfat_node* node, uint64_t size,
bool erase)
{
uint32_t c1 = bytes2clusters(ef, node->size);
uint32_t c2 = bytes2clusters(ef, size);
int rc = 0;
if (node->references == 0 && node->parent)
exfat_bug("no references, node changes can be lost");
if (node->size == size)
return 0;
if (c1 < c2)
rc = grow_file(ef, node, c1, c2 - c1);
else if (c1 > c2)
rc = shrink_file(ef, node, c1, c1 - c2);
if (rc != 0)
return rc;
if (erase)
{
rc = erase_range(ef, node, node->size, size);
if (rc != 0)
return rc;
}
exfat_update_mtime(node);
node->size = size;
node->flags |= EXFAT_ATTRIB_DIRTY;
return 0;
}
uint32_t exfat_count_free_clusters(const struct exfat* ef)
{
uint32_t free_clusters = 0;
uint32_t i;
for (i = 0; i < ef->cmap.size; i++)
if (BMAP_GET(ef->cmap.chunk, i) == 0)
free_clusters++;
return free_clusters;
}
static int find_used_clusters(const struct exfat* ef,
cluster_t* a, cluster_t* b)
{
const cluster_t end = le32_to_cpu(ef->sb->cluster_count);
/* find first used cluster */
for (*a = *b + 1; *a < end; (*a)++)
if (BMAP_GET(ef->cmap.chunk, *a - EXFAT_FIRST_DATA_CLUSTER))
break;
if (*a >= end)
return 1;
/* find last contiguous used cluster */
for (*b = *a; *b < end; (*b)++)
if (BMAP_GET(ef->cmap.chunk, *b - EXFAT_FIRST_DATA_CLUSTER) == 0)
{
(*b)--;
break;
}
return 0;
}
int exfat_find_used_sectors(const struct exfat* ef, off64_t* a, off64_t* b)
{
cluster_t ca, cb;
if (*a == 0 && *b == 0)
ca = cb = EXFAT_FIRST_DATA_CLUSTER - 1;
else
{
ca = s2c(ef, *a);
cb = s2c(ef, *b);
}
if (find_used_clusters(ef, &ca, &cb) != 0)
return 1;
if (*a != 0 || *b != 0)
*a = c2s(ef, ca);
*b = c2s(ef, cb) + (CLUSTER_SIZE(*ef->sb) - 1) / SECTOR_SIZE(*ef->sb);
return 0;
}