gecko-dev/dom/quota/DirectoryLockImpl.cpp
Jan Varga 554a32e5a3 Bug 1968869 - QM: Optimize DirectoryLockImpl::LocksMustWaitForInternal; r=dom-storage-reviewers,hsingh
LocksMustWaitForInternal is called at the start of every directory lock
acquisition and currently traverses all existing locks to determine conflicts.
However, shared locks never block other shared locks.

This patch optimizes performance by traversing only exclusive locks when
acquiring a shared lock. This reduces the cost of acquisition checks in common
scenarios while maintaining correctness.

Differential Revision: https://phabricator.services.mozilla.com/D251688
2025-06-07 02:11:26 +00:00

461 lines
14 KiB
C++

/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
/* vim: set ts=8 sts=2 et sw=2 tw=80: */
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this file,
* You can obtain one at http://mozilla.org/MPL/2.0/. */
#include "DirectoryLockImpl.h"
#include "nsError.h"
#include "nsString.h"
#include "nsThreadUtils.h"
#include "mozilla/ReverseIterator.h"
#include "mozilla/dom/quota/QuotaCommon.h"
#include "mozilla/dom/quota/QuotaManager.h"
namespace mozilla::dom::quota {
namespace {
/**
* Automatically log information about a directory lock if acquiring of the
* directory lock takes this long. We've chosen a value that is long enough
* that it is unlikely for the problem to be falsely triggered by slow system
* I/O. We've also chosen a value long enough so that testers can notice the
* timeout; we want to know about the timeouts, not hide them. On the other
* hand this value is less than 45 seconds which is used by quota manager to
* crash a hung quota manager shutdown.
*/
const uint32_t kAcquireTimeoutMs = 30000;
} // namespace
DirectoryLockImpl::DirectoryLockImpl(
MovingNotNull<RefPtr<QuotaManager>> aQuotaManager,
const PersistenceScope& aPersistenceScope, const OriginScope& aOriginScope,
const ClientStorageScope& aClientStorageScope, const bool aExclusive,
const bool aInternal,
const ShouldUpdateLockIdTableFlag aShouldUpdateLockIdTableFlag,
const DirectoryLockCategory aCategory)
: mQuotaManager(std::move(aQuotaManager)),
mPersistenceScope(aPersistenceScope),
mOriginScope(aOriginScope),
mClientStorageScope(aClientStorageScope),
mId(mQuotaManager->GenerateDirectoryLockId()),
mExclusive(aExclusive),
mInternal(aInternal),
mShouldUpdateLockIdTable(aShouldUpdateLockIdTableFlag ==
ShouldUpdateLockIdTableFlag::Yes),
mCategory(aCategory),
mRegistered(false) {
AssertIsOnOwningThread();
MOZ_ASSERT_IF(aOriginScope.IsOrigin(), !aOriginScope.GetOrigin().IsEmpty());
MOZ_ASSERT_IF(!aInternal, aPersistenceScope.IsValue());
MOZ_ASSERT_IF(!aInternal,
aPersistenceScope.GetValue() != PERSISTENCE_TYPE_INVALID);
MOZ_ASSERT_IF(!aInternal, aOriginScope.IsOrigin());
MOZ_ASSERT_IF(!aInternal, aClientStorageScope.IsClient());
MOZ_ASSERT_IF(!aInternal,
aClientStorageScope.GetClientType() < Client::TypeMax());
}
DirectoryLockImpl::~DirectoryLockImpl() {
AssertIsOnOwningThread();
MOZ_DIAGNOSTIC_ASSERT(!mRegistered);
}
bool DirectoryLockImpl::MustWait() const {
AssertIsOnOwningThread();
MOZ_ASSERT(!mRegistered);
// Shared locks never block other shared locks, so when acquiring a shared
// lock, we only need to consider existing exclusive locks. This reduces the
// cost of traversal when many locks are active. Exclusive locks must still
// consider all existing locks (both shared and exclusive). See also
// DirectoryLockImpl::MustWaitFor.
const auto& existingLocks = mExclusive
? mQuotaManager->mDirectoryLocks
: mQuotaManager->mExclusiveDirectoryLocks;
for (const DirectoryLockImpl* const existingLock : existingLocks) {
if (MustWaitFor(*existingLock)) {
return true;
}
}
return false;
}
nsTArray<RefPtr<DirectoryLockImpl>> DirectoryLockImpl::LocksMustWaitFor()
const {
AssertIsOnOwningThread();
return LocksMustWaitForInternal<RefPtr<DirectoryLockImpl>>();
}
DirectoryLockImpl::PrepareInfo DirectoryLockImpl::Prepare() const {
return PrepareInfo{*this};
}
RefPtr<BoolPromise> DirectoryLockImpl::Acquire() {
auto prepareInfo = Prepare();
return Acquire(std::move(prepareInfo));
}
RefPtr<BoolPromise> DirectoryLockImpl::Acquire(PrepareInfo&& aPrepareInfo) {
AssertIsOnOwningThread();
RefPtr<BoolPromise> result = mAcquirePromiseHolder.Ensure(__func__);
AcquireInternal(std::move(aPrepareInfo));
return result;
}
void DirectoryLockImpl::AcquireImmediately() {
AssertIsOnOwningThread();
MOZ_ASSERT(!MustWait());
mQuotaManager->RegisterDirectoryLock(*this);
mAcquired.Flip();
}
#ifdef DEBUG
void DirectoryLockImpl::AssertIsAcquiredExclusively() {
AssertIsOnOwningThread();
MOZ_ASSERT(mBlockedOn.IsEmpty());
MOZ_ASSERT(mExclusive);
MOZ_ASSERT(mInternal);
MOZ_ASSERT(mRegistered);
MOZ_ASSERT(!mInvalidated);
MOZ_ASSERT(mAcquired);
bool found = false;
for (const DirectoryLockImpl* const existingLock :
mQuotaManager->mDirectoryLocks) {
if (existingLock == this) {
MOZ_ASSERT(!found);
found = true;
} else if (existingLock->mAcquired) {
MOZ_ASSERT(false);
}
}
MOZ_ASSERT(found);
}
#endif
RefPtr<BoolPromise> DirectoryLockImpl::Drop() {
AssertIsOnOwningThread();
MOZ_ASSERT_IF(!mRegistered, mBlocking.IsEmpty());
mDropped.Flip();
return InvokeAsync(GetCurrentSerialEventTarget(), __func__,
[self = RefPtr(this)]() {
if (self->mRegistered) {
self->Unregister();
}
return BoolPromise::CreateAndResolve(true, __func__);
});
}
void DirectoryLockImpl::OnInvalidate(std::function<void()>&& aCallback) {
mInvalidateCallback = std::move(aCallback);
}
void DirectoryLockImpl::Log() const {
AssertIsOnOwningThread();
if (!QM_LOG_TEST()) {
return;
}
QM_LOG(("DirectoryLockImpl [%p]", this));
nsCString persistenceScope;
if (mPersistenceScope.IsNull()) {
persistenceScope.AssignLiteral("null");
} else if (mPersistenceScope.IsValue()) {
persistenceScope.Assign(
PersistenceTypeToString(mPersistenceScope.GetValue()));
} else {
MOZ_ASSERT(mPersistenceScope.IsSet());
for (auto persistenceType : mPersistenceScope.GetSet()) {
persistenceScope.Append(PersistenceTypeToString(persistenceType) +
" "_ns);
}
}
QM_LOG((" mPersistenceScope: %s", persistenceScope.get()));
nsCString originScope;
if (mOriginScope.IsOrigin()) {
originScope.AssignLiteral("origin:");
originScope.Append(mOriginScope.GetOrigin());
} else if (mOriginScope.IsPrefix()) {
originScope.AssignLiteral("prefix:");
originScope.Append(mOriginScope.GetOriginNoSuffix());
} else if (mOriginScope.IsPattern()) {
originScope.AssignLiteral("pattern:");
// Can't call GetJSONPattern since it only works on the main thread.
} else {
MOZ_ASSERT(mOriginScope.IsNull());
originScope.AssignLiteral("null");
}
QM_LOG((" mOriginScope: %s", originScope.get()));
nsCString clientStorageScope;
if (mClientStorageScope.IsNull()) {
clientStorageScope.AssignLiteral("null");
} else if (mClientStorageScope.IsClient()) {
clientStorageScope.Assign(
Client::TypeToText(mClientStorageScope.GetClientType()));
} else {
MOZ_ASSERT(mClientStorageScope.IsMetadata());
clientStorageScope.AssignLiteral("metadata");
}
QM_LOG((" mClientStorageScope: %s", clientStorageScope.get()));
nsCString blockedOnString;
for (auto blockedOn : mBlockedOn) {
blockedOnString.Append(
nsPrintfCString(" [%p]", static_cast<void*>(blockedOn)));
}
QM_LOG((" mBlockedOn:%s", blockedOnString.get()));
QM_LOG((" mExclusive: %d", mExclusive));
QM_LOG((" mInternal: %d", mInternal));
QM_LOG((" mInvalidated: %d", static_cast<bool>(mInvalidated)));
for (auto blockedOn : mBlockedOn) {
blockedOn->Log();
}
}
#ifdef DEBUG
void DirectoryLockImpl::AssertIsOnOwningThread() const {
mQuotaManager->AssertIsOnOwningThread();
}
#endif // DEBUG
bool DirectoryLockImpl::Overlaps(const DirectoryLockImpl& aLock) const {
AssertIsOnOwningThread();
// If the persistence types don't overlap, the op can proceed.
bool match = aLock.mPersistenceScope.Matches(mPersistenceScope);
if (!match) {
return false;
}
// If the origin scopes don't overlap, the op can proceed.
match = aLock.mOriginScope.Matches(mOriginScope);
if (!match) {
return false;
}
// If the client storage scopes don't overlap, the op can proceed.
match = aLock.mClientStorageScope.Matches(mClientStorageScope);
if (!match) {
return false;
}
// Otherwise, when all attributes overlap (persistence type, origin scope and
// client type) the op must wait.
return true;
}
bool DirectoryLockImpl::MustWaitFor(const DirectoryLockImpl& aLock) const {
AssertIsOnOwningThread();
// Waiting is never required if the ops in comparison represent shared locks.
// Note that this condition is also used to optimize traversal in MustWait
// and LocksMustWaitForInternal. If this logic changes, that optimization
// must be revisited to ensure correctness.
if (!aLock.mExclusive && !mExclusive) {
return false;
}
// Wait if the ops overlap.
return Overlaps(aLock);
}
void DirectoryLockImpl::NotifyOpenListener() {
AssertIsOnOwningThread();
if (mAcquireTimer) {
mAcquireTimer->Cancel();
mAcquireTimer = nullptr;
}
if (mInvalidated) {
mAcquirePromiseHolder.Reject(NS_ERROR_FAILURE, __func__);
} else {
mAcquired.Flip();
mAcquirePromiseHolder.Resolve(true, __func__);
}
MOZ_ASSERT(mAcquirePromiseHolder.IsEmpty());
mQuotaManager->RemovePendingDirectoryLock(*this);
mPending.Flip();
if (mInvalidated) {
mDropped.Flip();
Unregister();
}
}
template <typename T>
nsTArray<T> DirectoryLockImpl::LocksMustWaitForInternal() const {
AssertIsOnOwningThread();
MOZ_ASSERT(!mRegistered);
nsTArray<T> locks;
// Shared locks never block other shared locks, so when acquiring a shared
// lock, we only need to consider existing exclusive locks. This reduces the
// cost of traversal when many locks are active. Exclusive locks must still
// consider all existing locks (both shared and exclusive). See also
// DirectoryLockImpl::MustWaitFor.
const auto& existingLocks = mExclusive
? mQuotaManager->mDirectoryLocks
: mQuotaManager->mExclusiveDirectoryLocks;
// XXX It is probably unnecessary to iterate this in reverse order.
for (DirectoryLockImpl* const existingLock : Reversed(existingLocks)) {
if (MustWaitFor(*existingLock)) {
if constexpr (std::is_same_v<T, NotNull<DirectoryLockImpl*>>) {
locks.AppendElement(WrapNotNull(existingLock));
} else {
locks.AppendElement(existingLock);
}
}
}
return locks;
}
void DirectoryLockImpl::AcquireInternal(PrepareInfo&& aPrepareInfo) {
AssertIsOnOwningThread();
mQuotaManager->AddPendingDirectoryLock(*this);
// See if this lock needs to wait. This has to be done before the lock is
// registered, we would be comparing the lock against itself otherwise.
mBlockedOn = std::move(aPrepareInfo.mBlockedOn);
// After the traversal of existing locks is done, this lock can be
// registered and will become an existing lock as well.
mQuotaManager->RegisterDirectoryLock(*this);
// If this lock is not blocked by some other existing lock, notify the open
// listener immediately and return.
if (mBlockedOn.IsEmpty()) {
NotifyOpenListener();
return;
}
// Add this lock as a blocking lock to all locks which block it, so the
// locks can update this lock when they are unregistered and eventually
// unblock this lock.
for (auto& blockedOnLock : mBlockedOn) {
blockedOnLock->AddBlockingLock(*this);
}
mAcquireTimer = NS_NewTimer();
MOZ_ALWAYS_SUCCEEDS(mAcquireTimer->InitWithNamedFuncCallback(
[](nsITimer* aTimer, void* aClosure) {
if (!QM_LOG_TEST()) {
return;
}
auto* const lock = static_cast<DirectoryLockImpl*>(aClosure);
QM_LOG(("Directory lock [%p] is taking too long to be acquired", lock));
lock->Log();
},
this, kAcquireTimeoutMs, nsITimer::TYPE_ONE_SHOT,
"quota::DirectoryLockImpl::AcquireInternal"));
if (!mExclusive || !mInternal) {
return;
}
// All the locks that block this new exclusive internal lock need to be
// invalidated. We also need to notify clients to abort operations for them.
QuotaManager::DirectoryLockIdTableArray lockIds;
lockIds.SetLength(Client::TypeMax());
const auto& blockedOnLocks = GetBlockedOnLocks();
MOZ_ASSERT(!blockedOnLocks.IsEmpty());
for (DirectoryLockImpl* blockedOnLock : blockedOnLocks) {
if (!blockedOnLock->IsInternal()) {
blockedOnLock->Invalidate();
// Clients don't have to handle pending locks. Invalidation is sufficient
// in that case (once a lock is ready and the listener needs to be
// notified, we will call DirectoryLockFailed instead of
// DirectoryLockAcquired which should release any remaining references to
// the lock).
if (!blockedOnLock->IsPending()) {
lockIds[blockedOnLock->ClientType()].Put(blockedOnLock->Id());
}
}
}
mQuotaManager->AbortOperationsForLocks(lockIds);
}
void DirectoryLockImpl::Invalidate() {
AssertIsOnOwningThread();
mInvalidated.EnsureFlipped();
if (mInvalidateCallback) {
MOZ_ALWAYS_SUCCEEDS(GetCurrentSerialEventTarget()->Dispatch(
NS_NewRunnableFunction("DirectoryLockImpl::Invalidate",
[invalidateCallback = mInvalidateCallback]() {
invalidateCallback();
}),
NS_DISPATCH_NORMAL));
}
}
void DirectoryLockImpl::Unregister() {
AssertIsOnOwningThread();
MOZ_ASSERT(mRegistered);
// We must call UnregisterDirectoryLock before unblocking other locks because
// UnregisterDirectoryLock also updates the origin last access time and the
// access flag (if the last lock for given origin is unregistered). One of the
// blocked locks could be requested by the clear/reset operation which stores
// cached information about origins in storage.sqlite. So if the access flag
// is not updated before unblocking the lock for reset/clear, we might store
// invalid information which can lead to omitting origin initialization during
// next temporary storage initialization.
mQuotaManager->UnregisterDirectoryLock(*this);
MOZ_ASSERT(!mRegistered);
for (NotNull<RefPtr<DirectoryLockImpl>> blockingLock : mBlocking) {
blockingLock->MaybeUnblock(*this);
}
mBlocking.Clear();
}
} // namespace mozilla::dom::quota