|
@@ -47,43 +47,10 @@ static inline size_t buffer_start(struct persistent_ram_zone *prz)
|
|
|
return atomic_read(&prz->buffer->start);
|
|
|
}
|
|
|
|
|
|
-/* increase and wrap the start pointer, returning the old value */
|
|
|
-static size_t buffer_start_add_atomic(struct persistent_ram_zone *prz, size_t a)
|
|
|
-{
|
|
|
- int old;
|
|
|
- int new;
|
|
|
-
|
|
|
- do {
|
|
|
- old = atomic_read(&prz->buffer->start);
|
|
|
- new = old + a;
|
|
|
- while (unlikely(new >= prz->buffer_size))
|
|
|
- new -= prz->buffer_size;
|
|
|
- } while (atomic_cmpxchg(&prz->buffer->start, old, new) != old);
|
|
|
-
|
|
|
- return old;
|
|
|
-}
|
|
|
-
|
|
|
-/* increase the size counter until it hits the max size */
|
|
|
-static void buffer_size_add_atomic(struct persistent_ram_zone *prz, size_t a)
|
|
|
-{
|
|
|
- size_t old;
|
|
|
- size_t new;
|
|
|
-
|
|
|
- if (atomic_read(&prz->buffer->size) == prz->buffer_size)
|
|
|
- return;
|
|
|
-
|
|
|
- do {
|
|
|
- old = atomic_read(&prz->buffer->size);
|
|
|
- new = old + a;
|
|
|
- if (new > prz->buffer_size)
|
|
|
- new = prz->buffer_size;
|
|
|
- } while (atomic_cmpxchg(&prz->buffer->size, old, new) != old);
|
|
|
-}
|
|
|
-
|
|
|
static DEFINE_RAW_SPINLOCK(buffer_lock);
|
|
|
|
|
|
/* increase and wrap the start pointer, returning the old value */
|
|
|
-static size_t buffer_start_add_locked(struct persistent_ram_zone *prz, size_t a)
|
|
|
+static size_t buffer_start_add(struct persistent_ram_zone *prz, size_t a)
|
|
|
{
|
|
|
int old;
|
|
|
int new;
|
|
@@ -103,7 +70,7 @@ static size_t buffer_start_add_locked(struct persistent_ram_zone *prz, size_t a)
|
|
|
}
|
|
|
|
|
|
/* increase the size counter until it hits the max size */
|
|
|
-static void buffer_size_add_locked(struct persistent_ram_zone *prz, size_t a)
|
|
|
+static void buffer_size_add(struct persistent_ram_zone *prz, size_t a)
|
|
|
{
|
|
|
size_t old;
|
|
|
size_t new;
|
|
@@ -124,9 +91,6 @@ exit:
|
|
|
raw_spin_unlock_irqrestore(&buffer_lock, flags);
|
|
|
}
|
|
|
|
|
|
-static size_t (*buffer_start_add)(struct persistent_ram_zone *, size_t) = buffer_start_add_atomic;
|
|
|
-static void (*buffer_size_add)(struct persistent_ram_zone *, size_t) = buffer_size_add_atomic;
|
|
|
-
|
|
|
static void notrace persistent_ram_encode_rs8(struct persistent_ram_zone *prz,
|
|
|
uint8_t *data, size_t len, uint8_t *ecc)
|
|
|
{
|
|
@@ -426,9 +390,6 @@ static void *persistent_ram_iomap(phys_addr_t start, size_t size,
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
|
- buffer_start_add = buffer_start_add_locked;
|
|
|
- buffer_size_add = buffer_size_add_locked;
|
|
|
-
|
|
|
if (memtype)
|
|
|
va = ioremap(start, size);
|
|
|
else
|