tutanota/src/common/api/worker/rest/DefaultEntityRestCache.ts

996 lines
42 KiB
TypeScript
Raw Normal View History

import {
CacheMode,
EntityRestClient,
EntityRestClientEraseOptions,
EntityRestClientLoadOptions,
EntityRestClientSetupOptions,
EntityRestInterface,
OwnerEncSessionKeyProvider,
} from "./EntityRestClient"
2022-12-27 15:37:40 +01:00
import { resolveTypeReference } from "../../common/EntityFunctions"
import { OperationType } from "../../common/TutanotaConstants"
import { assertNotNull, difference, getFirstOrThrow, getTypeId, groupBy, isEmpty, isSameTypeRef, lastThrow, TypeRef } from "@tutao/tutanota-utils"
import {
AuditLogEntryTypeRef,
BucketPermissionTypeRef,
EntityEventBatchTypeRef,
EntityUpdate,
GroupKeyTypeRef,
KeyRotationTypeRef,
PermissionTypeRef,
RecoverCodeTypeRef,
RejectedSenderTypeRef,
SecondFactorTypeRef,
SessionTypeRef,
User,
UserGroupKeyDistributionTypeRef,
UserGroupRootTypeRef,
2022-12-27 15:37:40 +01:00
UserTypeRef,
} from "../../entities/sys/TypeRefs.js"
2024-07-05 11:32:43 +02:00
import { ValueType } from "../../common/EntityConstants.js"
2022-12-27 15:37:40 +01:00
import { NotAuthorizedError, NotFoundError } from "../../common/error/RestError"
import {
CalendarEventUidIndexTypeRef,
Mail,
MailDetailsBlobTypeRef,
MailFolderTypeRef,
MailSetEntryTypeRef,
MailTypeRef,
} from "../../entities/tutanota/TypeRefs.js"
import { CUSTOM_MAX_ID, CUSTOM_MIN_ID, firstBiggerThanSecond, GENERATED_MAX_ID, GENERATED_MIN_ID, getElementId, isSameId } from "../../common/utils/EntityUtils"
2022-12-27 15:37:40 +01:00
import { ProgrammingError } from "../../common/error/ProgrammingError"
import { assertWorkerOrNode } from "../../common/Env"
Fix outdated mails from notification Notification process on mobile does insert Mail instances into offline db in order to show the mail preview as quickly as possible when the user clicks on notification. In most cases it happens before login, and it happens independently of entity update processing. EntityRestCache does only download new list element instances when they are within the cached range. In most cases new mails are within the cached range as mail indexer caches mail bag ranges during initial indexing. However, if the mail bag size has been reached there will be a new mail bag with mail list that is not cached by the client (there is no range stored for this new list). EventQueue optimizes event updates. If CREATE event is followed by an UPDATE event they will be folded into a single CREATE event. In a situation where the email has been updated by another client after it has been inserted into offline db by notification process but before the login, and that email belongs to uncached mail bag, CREATE event (and optimized away UPDATE events) for that mail would be skipped and the user would see an outdated Mail instance. We changed the behavior for Mail instances to be always downloaded on CREATE event when we use Offline cache so that we do not miss those update operations. This was the previously expected behavior because of mail indexing and is also a common behavior now, until new mail bag is created so it does not lead to additional requests in most cases. Alternatively we could check if the (new) instance is already cached as individual instance. We decided against that as it adds the delay to event processing and also does not fit the logic of processing (why should cache check if new instance is already cache?). In the future we should try to optimize loading of new instances so that the sync time does not increase from downloading single emails. Close #8041 Co-authored-by: ivk <ivk@tutao.de>
2024-12-17 15:46:05 +01:00
import type { ElementEntity, ListElementEntity, SomeEntity, TypeModel } from "../../common/EntityTypes"
2023-02-10 12:01:42 +01:00
import { QueuedBatch } from "../EventQueue.js"
2022-12-27 15:37:40 +01:00
import { ENTITY_EVENT_BATCH_EXPIRE_MS } from "../EventBusClient"
import { CustomCacheHandlerMap } from "./CustomCacheHandler.js"
import { containsEventOfType, EntityUpdateData, getEventOfType } from "../../common/utils/EntityUpdateUtils.js"
import { isCustomIdType } from "../offline/OfflineStorage.js"
2017-08-15 13:54:22 +02:00
assertWorkerOrNode()
/**
*
* The minimum size of a range request when extending an existing range
* Because we extend by making (potentially) many range requests until we reach the startId
* We want to avoid that the requests are too small
*/
export const EXTEND_RANGE_MIN_CHUNK_SIZE = 40
const IGNORED_TYPES = [
EntityEventBatchTypeRef,
PermissionTypeRef,
BucketPermissionTypeRef,
SessionTypeRef,
SecondFactorTypeRef,
RecoverCodeTypeRef,
RejectedSenderTypeRef,
// when doing automatic calendar updates, we will miss uid index entity updates if we're using the cache.
// this is mainly caused by some calendaring apps sending the same update multiple times in the same mail.
// the earliest place where we could deduplicate would be in entityEventsReceived on the calendarModel.
CalendarEventUidIndexTypeRef,
KeyRotationTypeRef,
UserGroupRootTypeRef,
UserGroupKeyDistributionTypeRef,
AuditLogEntryTypeRef, // Should not be part of cached data because there are errors inside entity event processing after rotating the admin group key
] as const
/**
* List of types containing a customId that we want to explicitly enable caching for.
* CustomId types are not cached by default because their id is using base64UrlEncoding while GeneratedUId types are using base64Ext encoding.
* base64Url encoding results in a different sort order of elements that we have on the server, this is problematic for caching LET and their ranges.
* When enabling caching for customId types we convert the id that we store in cache from base64Url to base64Ext so we have the same sort order. (see function
* OfflineStorage.ensureBase64Ext). In theory, we can try to enable caching for all types but as of now we enable it for a limited amount of types because there
* are other ways to cache customId types (see implementation of CustomCacheHandler)
*/
const CACHABLE_CUSTOMID_TYPES = [MailSetEntryTypeRef, GroupKeyTypeRef] as const
export interface EntityRestCache extends EntityRestInterface {
/**
* Clear out the contents of the cache.
*/
2022-12-27 15:37:40 +01:00
purgeStorage(): Promise<void>
/**
* Get the batch id of the most recently processed batch for the given group.
*/
2022-12-27 15:37:40 +01:00
getLastEntityEventBatchForGroup(groupId: Id): Promise<Id | null>
/**
* Saved tha batch id of the most recently processed batch manually.
*
* Is needed when the cache is new but we want to make sure that the next time we will download from this moment, even if we don't receive any events.
*/
2022-12-27 15:37:40 +01:00
setLastEntityEventBatchForGroup(groupId: Id, batchId: Id): Promise<void>
/**
* Persist the last time client downloaded event batches. This is not the last *processed* item, merely when things were *downloaded*. We use it to
* detect out-of-sync.
*/
2022-12-27 15:37:40 +01:00
recordSyncTime(): Promise<void>
/**
* Fetch the time since last time we downloaded event batches.
*/
2022-12-27 15:37:40 +01:00
timeSinceLastSyncMs(): Promise<number | null>
/**
* Detect if out of sync based on stored "lastUpdateTime" and the current server time
*/
2022-12-27 15:37:40 +01:00
isOutOfSync(): Promise<boolean>
}
2022-12-27 15:37:40 +01:00
export type Range = { lower: Id; upper: Id }
2022-12-27 15:37:40 +01:00
export type LastUpdateTime = { type: "recorded"; time: number } | { type: "never" } | { type: "uninitialized" }
/**
* Part of the cache storage only with subset of CacheStorage functionality
*
* Separate from the rest of the cache as a narrow interface to not expose the whole storage for cases where we want to only get the cached part of the list to
* display it even if we can't load the full page from the server or need some metadata.
*
* also exposes functions to repair an outdated cache in case we can't access the server without getting a new version of a cached entity
* (mainly password changes)
*/
export interface ExposedCacheStorage {
get<T extends SomeEntity>(typeRef: TypeRef<T>, listId: Id | null, id: Id): Promise<T | null>
/**
* Load range of entities. Does not include {@param start}.
* If {@param reverse} is false then returns entities newer than {@param start} in ascending order sorted by
* elementId.
* If {@param reverse} is true then returns entities older than {@param start} in descending order sorted by
* elementId.
*/
2022-12-27 15:37:40 +01:00
provideFromRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, start: Id, count: number, reverse: boolean): Promise<T[]>
/**
* Load a set of list element entities by id. Missing elements are not returned, no error is thrown.
*/
provideMultiple<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, elementIds: Id[]): Promise<Array<T>>
/**
* retrieve all list elements that are in the cache
* @param typeRef
* @param listId
*/
getWholeList<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id): Promise<Array<T>>
getLastUpdateTime(): Promise<LastUpdateTime>
clearExcludedData(): Promise<void>
/**
* remove an ElementEntity from the cache by typeRef and Id.
* the exposed interface is intentionally more narrow than the internal cacheStorage because
* we must maintain the integrity of our list ranges.
* */
deleteIfExists<T extends ElementEntity>(typeRef: TypeRef<T>, listId: null, id: Id): Promise<void>
/** delete all instances of the given type that share {@param listId}. also deletes the range of that list. */
deleteWholeList<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id): Promise<void>
}
export interface CacheStorage extends ExposedCacheStorage {
/**
* Get a given entity from the cache, expects that you have already checked for existence
*/
2022-12-27 15:37:40 +01:00
get<T extends SomeEntity>(typeRef: TypeRef<T>, listId: Id | null, id: Id): Promise<T | null>
/**
* get a map with cache handlers for the customId types this storage implementation supports
* customId types that don't have a custom handler don't get served from the cache
*/
2022-12-27 15:37:40 +01:00
getCustomCacheHandlerMap(entityRestClient: EntityRestClient): CustomCacheHandlerMap
2022-12-27 15:37:40 +01:00
isElementIdInCacheRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, id: Id): Promise<boolean>
2022-12-27 15:37:40 +01:00
put(originalEntity: SomeEntity): Promise<void>
2022-12-27 15:37:40 +01:00
getRangeForList<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id): Promise<Range | null>
2022-12-27 15:37:40 +01:00
setUpperRangeForList<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, id: Id): Promise<void>
2022-12-27 15:37:40 +01:00
setLowerRangeForList<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, id: Id): Promise<void>
/**
* Creates a new list cache if there is none. Resets everything but elements.
* @param typeRef
* @param listId
* @param lower
* @param upper
*/
2022-12-27 15:37:40 +01:00
setNewRangeForList<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, lower: Id, upper: Id): Promise<void>
2022-12-27 15:37:40 +01:00
getIdsInRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id): Promise<Array<Id>>
/**
* Persist the last processed batch for a given group id.
*/
2022-12-27 15:37:40 +01:00
putLastBatchIdForGroup(groupId: Id, batchId: Id): Promise<void>
/**
* Retrieve the least processed batch id for a given group.
*/
2022-12-27 15:37:40 +01:00
getLastBatchIdForGroup(groupId: Id): Promise<Id | null>
deleteIfExists<T extends SomeEntity>(typeRef: TypeRef<T>, listId: Id | null, id: Id): Promise<void>
purgeStorage(): Promise<void>
putLastUpdateTime(value: number): Promise<void>
getUserId(): Id
deleteAllOwnedBy(owner: Id): Promise<void>
/**
* We want to lock the access to the "ranges" db when updating / reading the
* offline available mail list ranges for each mail list (referenced using the listId)
* @param listId the mail list that we want to lock
*/
lockRangesDbAccess(listId: Id): Promise<void>
/**
* This is the counterpart to the function "lockRangesDbAccess(listId)"
* @param listId the mail list that we want to unlock
*/
unlockRangesDbAccess(listId: Id): Promise<void>
}
2017-08-15 13:54:22 +02:00
/**
* This implementation provides a caching mechanism to the rest chain.
* It forwards requests to the entity rest client.
* The cache works as follows:
* If a read from the target fails, the request fails.
* If a read from the target is successful, the cache is written and the element returned.
* For LETs the cache stores one range per list id. if a range is requested starting in the stored range or at the range ends the missing elements are loaded from the server.
* Only ranges with elements with generated ids are stored in the cache. Custom id elements are only stored as single element currently. If needed this has to be extended for ranges.
* Range requests starting outside the stored range are only allowed if the direction is away from the stored range. In this case we load from the range end to avoid gaps in the stored range.
* Requests for creating or updating elements are always forwarded and not directly stored in the cache.
* On EventBusClient notifications updated elements are stored in the cache if the element already exists in the cache.
* On EventBusClient notifications new elements are only stored in the cache if they are LETs and in the stored range.
* On EventBusClient notifications deleted elements are removed from the cache.
*
* Range handling:
* | <|> c d e f g h i j k <|> |
* MIN_ID lowerRangeId ids in range upperRangeId MAX_ID
2017-08-15 13:54:22 +02:00
* lowerRangeId may be anything from MIN_ID to c, upperRangeId may be anything from k to MAX_ID
*/
export class DefaultEntityRestCache implements EntityRestCache {
constructor(private readonly entityRestClient: EntityRestClient, private readonly storage: CacheStorage) {}
async load<T extends SomeEntity>(typeRef: TypeRef<T>, id: PropertyType<T, "_id">, opts: EntityRestClientLoadOptions = {}): Promise<T> {
const { queryParams, cacheMode = CacheMode.Cache } = opts
2022-12-27 15:37:40 +01:00
const { listId, elementId } = expandId(id)
// if a specific version is requested we have to load again and do not want to store it in the cache
if (queryParams?.version != null) {
return await this.entityRestClient.load(typeRef, id, opts)
}
let cachedEntity: T | null
if (cacheMode === CacheMode.Cache) {
cachedEntity = await this.storage.get(typeRef, listId, elementId)
} else {
cachedEntity = null
}
if (cachedEntity == null) {
const entity = await this.entityRestClient.load(typeRef, id, opts)
if (!isIgnoredType(typeRef)) {
await this.storage.put(entity)
}
return entity
}
return cachedEntity
}
loadMultiple<T extends SomeEntity>(
typeRef: TypeRef<T>,
listId: Id | null,
elementIds: Array<Id>,
2023-10-19 17:59:51 +02:00
ownerEncSessionKeyProvider?: OwnerEncSessionKeyProvider,
): Promise<Array<T>> {
if (isIgnoredType(typeRef)) {
2023-10-19 17:59:51 +02:00
return this.entityRestClient.loadMultiple(typeRef, listId, elementIds, ownerEncSessionKeyProvider)
}
2023-10-19 17:59:51 +02:00
return this._loadMultiple(typeRef, listId, elementIds, ownerEncSessionKeyProvider)
}
setup<T extends SomeEntity>(listId: Id | null, instance: T, extraHeaders?: Dict, options?: EntityRestClientSetupOptions): Promise<Id> {
return this.entityRestClient.setup(listId, instance, extraHeaders, options)
}
setupMultiple<T extends SomeEntity>(listId: Id | null, instances: Array<T>): Promise<Array<Id>> {
return this.entityRestClient.setupMultiple(listId, instances)
}
update<T extends SomeEntity>(instance: T): Promise<void> {
return this.entityRestClient.update(instance)
}
erase<T extends SomeEntity>(instance: T, options?: EntityRestClientEraseOptions): Promise<void> {
return this.entityRestClient.erase(instance, options)
}
getLastEntityEventBatchForGroup(groupId: Id): Promise<Id | null> {
return this.storage.getLastBatchIdForGroup(groupId)
}
setLastEntityEventBatchForGroup(groupId: Id, batchId: Id): Promise<void> {
return this.storage.putLastBatchIdForGroup(groupId, batchId)
}
purgeStorage(): Promise<void> {
console.log("Purging the user's offline database")
return this.storage.purgeStorage()
}
async isOutOfSync(): Promise<boolean> {
const timeSinceLastSync = await this.timeSinceLastSyncMs()
return timeSinceLastSync != null && timeSinceLastSync > ENTITY_EVENT_BATCH_EXPIRE_MS
}
async recordSyncTime(): Promise<void> {
const timestamp = this.getServerTimestampMs()
await this.storage.putLastUpdateTime(timestamp)
}
async timeSinceLastSyncMs(): Promise<number | null> {
const lastUpdate = await this.storage.getLastUpdateTime()
let lastUpdateTime: number
switch (lastUpdate.type) {
case "recorded":
lastUpdateTime = lastUpdate.time
break
case "never":
return null
case "uninitialized":
throw new ProgrammingError("Offline storage is not initialized")
}
const now = this.getServerTimestampMs()
return now - lastUpdateTime
}
private getServerTimestampMs(): number {
return this.entityRestClient.getRestClient().getServerTimestampMs()
}
/**
* Delete a cached entity. Sometimes this is necessary to do to ensure you always load the new version
*/
deleteFromCacheIfExists<T extends SomeEntity>(typeRef: TypeRef<T>, listId: Id | null, elementId: Id): Promise<void> {
return this.storage.deleteIfExists(typeRef, listId, elementId)
}
private async _loadMultiple<T extends SomeEntity>(
typeRef: TypeRef<T>,
listId: Id | null,
ids: Array<Id>,
2023-10-19 17:59:51 +02:00
ownerEncSessionKeyProvider?: OwnerEncSessionKeyProvider,
): Promise<Array<T>> {
const entitiesInCache: T[] = []
const idsToLoad: Id[] = []
for (let id of ids) {
const cachedEntity = await this.storage.get(typeRef, listId, id)
if (cachedEntity != null) {
entitiesInCache.push(cachedEntity)
} else {
idsToLoad.push(id)
}
}
const entitiesFromServer: T[] = []
if (idsToLoad.length > 0) {
2023-10-19 17:59:51 +02:00
const entities = await this.entityRestClient.loadMultiple(typeRef, listId, idsToLoad, ownerEncSessionKeyProvider)
for (let entity of entities) {
await this.storage.put(entity)
entitiesFromServer.push(entity)
}
}
return entitiesFromServer.concat(entitiesInCache)
}
async loadRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, start: Id, count: number, reverse: boolean): Promise<T[]> {
Fix outdated mails from notification Notification process on mobile does insert Mail instances into offline db in order to show the mail preview as quickly as possible when the user clicks on notification. In most cases it happens before login, and it happens independently of entity update processing. EntityRestCache does only download new list element instances when they are within the cached range. In most cases new mails are within the cached range as mail indexer caches mail bag ranges during initial indexing. However, if the mail bag size has been reached there will be a new mail bag with mail list that is not cached by the client (there is no range stored for this new list). EventQueue optimizes event updates. If CREATE event is followed by an UPDATE event they will be folded into a single CREATE event. In a situation where the email has been updated by another client after it has been inserted into offline db by notification process but before the login, and that email belongs to uncached mail bag, CREATE event (and optimized away UPDATE events) for that mail would be skipped and the user would see an outdated Mail instance. We changed the behavior for Mail instances to be always downloaded on CREATE event when we use Offline cache so that we do not miss those update operations. This was the previously expected behavior because of mail indexing and is also a common behavior now, until new mail bag is created so it does not lead to additional requests in most cases. Alternatively we could check if the (new) instance is already cached as individual instance. We decided against that as it adds the delay to event processing and also does not fit the logic of processing (why should cache check if new instance is already cache?). In the future we should try to optimize loading of new instances so that the sync time does not increase from downloading single emails. Close #8041 Co-authored-by: ivk <ivk@tutao.de>
2024-12-17 15:46:05 +01:00
const customHandler = this.storage.getCustomCacheHandlerMap(this.entityRestClient).get(typeRef)
if (customHandler && customHandler.loadRange) {
return await customHandler.loadRange(this.storage, listId, start, count, reverse)
}
const typeModel = await resolveTypeReference(typeRef)
if (!isCachedType(typeModel, typeRef)) {
return this.entityRestClient.loadRange(typeRef, listId, start, count, reverse)
}
2021-12-28 13:53:11 +01:00
// We lock access to the "ranges" db here in order to prevent race conditions when accessing the ranges database.
await this.storage.lockRangesDbAccess(listId)
2021-12-28 13:53:11 +01:00
try {
const range = await this.storage.getRangeForList(typeRef, listId)
2022-12-27 15:37:40 +01:00
if (range == null) {
await this.populateNewListWithRange(typeRef, listId, start, count, reverse)
} else if (isStartIdWithinRange(range, start, typeModel)) {
await this.extendFromWithinRange(typeRef, listId, start, count, reverse)
} else if (isRangeRequestAwayFromExistingRange(range, reverse, start, typeModel)) {
await this.extendAwayFromRange(typeRef, listId, start, count, reverse)
} else {
await this.extendTowardsRange(typeRef, listId, start, count, reverse)
}
2021-12-23 14:03:23 +01:00
return this.storage.provideFromRange(typeRef, listId, start, count, reverse)
} finally {
// We unlock access to the "ranges" db here. We lock it in order to prevent race conditions when accessing the "ranges" database.
await this.storage.unlockRangesDbAccess(listId)
}
}
/**
* Creates a new list range, reading everything from the server that it can
* range: (none)
* request: *--------->
* range becomes: |---------|
* @private
*/
2022-12-27 15:37:40 +01:00
private async populateNewListWithRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, start: Id, count: number, reverse: boolean) {
// Create a new range and load everything
const entities = await this.entityRestClient.loadRange(typeRef, listId, start, count, reverse)
// Initialize a new range for this list
await this.storage.setNewRangeForList(typeRef, listId, start, start)
// The range bounds will be updated in here
await this.updateRangeInStorage(typeRef, listId, count, reverse, entities)
}
/**
* Returns part of a request from the cache, and the remainder is loaded from the server
* range: |---------|
* request: *-------------->
* range becomes: |--------------------|
*/
private async extendFromWithinRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, start: Id, count: number, reverse: boolean) {
2022-12-27 15:37:40 +01:00
const { newStart, newCount } = await this.recalculateRangeRequest(typeRef, listId, start, count, reverse)
if (newCount > 0) {
// We will be able to provide some entities from the cache, so we just want to load the remaining entities from the server
const entities = await this.entityRestClient.loadRange(typeRef, listId, newStart, newCount, reverse)
await this.updateRangeInStorage(typeRef, listId, newCount, reverse, entities)
}
}
/**
* Start was outside the range, and we are loading away from the range
* Keeps loading elements from the end of the range in the direction of the startId.
* Returns once all available elements have been loaded or the requested number is in cache
* range: |---------|
* request: *------->
* range becomes: |--------------------|
*/
2022-12-27 15:37:40 +01:00
private async extendAwayFromRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, start: Id, count: number, reverse: boolean) {
// Start is outside the range, and we are loading away from the range, so we grow until we are able to provide enough
// entities starting at startId
while (true) {
const range = assertNotNull(await this.storage.getRangeForList(typeRef, listId))
// Which end of the range to start loading from
2022-12-27 15:37:40 +01:00
const loadStartId = reverse ? range.lower : range.upper
const requestCount = Math.max(count, EXTEND_RANGE_MIN_CHUNK_SIZE)
// Load some entities
const entities = await this.entityRestClient.loadRange(typeRef, listId, loadStartId, requestCount, reverse)
await this.updateRangeInStorage(typeRef, listId, requestCount, reverse, entities)
// If we exhausted the entities from the server
if (entities.length < requestCount) {
break
}
// Try to get enough entities from cache
const entitiesFromCache = await this.storage.provideFromRange(typeRef, listId, start, count, reverse)
// If cache is now capable of providing the whole request
if (entitiesFromCache.length === count) {
break
}
}
}
/**
* Loads all elements from the startId in the direction of the range
* Once complete, returns as many elements as it can from the original request
* range: |---------|
* request: <------*
* range becomes: |--------------------|
* or
* range: |---------|
* request: <-------------------*
* range becomes: |--------------------|
*/
private async extendTowardsRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, start: Id, count: number, reverse: boolean) {
while (true) {
const range = assertNotNull(await this.storage.getRangeForList(typeRef, listId))
2022-12-27 15:37:40 +01:00
const loadStartId = reverse ? range.upper : range.lower
const requestCount = Math.max(count, EXTEND_RANGE_MIN_CHUNK_SIZE)
const entities = await this.entityRestClient.loadRange(typeRef, listId, loadStartId, requestCount, !reverse)
await this.updateRangeInStorage(typeRef, listId, requestCount, !reverse, entities)
// The call to `updateRangeInStorage` will have set the range bounds to GENERATED_MIN_ID/GENERATED_MAX_ID
// in the case that we have exhausted all elements from the server, so if that happens, we will also end up breaking here
if (await this.storage.isElementIdInCacheRange(typeRef, listId, start)) {
break
}
}
await this.extendFromWithinRange(typeRef, listId, start, count, reverse)
}
/**
* Given the parameters and result of a range request,
* Inserts the result into storage, and updates the range bounds
* based on number of entities requested and the actual amount that were received
*/
private async updateRangeInStorage<T extends ListElementEntity>(
typeRef: TypeRef<T>,
listId: Id,
countRequested: number,
wasReverseRequest: boolean,
receivedEntities: T[],
) {
const isCustomId = isCustomIdType(await resolveTypeReference(typeRef))
let elementsToAdd = receivedEntities
if (wasReverseRequest) {
// Ensure that elements are cached in ascending (not reverse) order
elementsToAdd = receivedEntities.reverse()
if (receivedEntities.length < countRequested) {
console.log("finished loading, setting min id")
await this.storage.setLowerRangeForList(typeRef, listId, isCustomId ? CUSTOM_MIN_ID : GENERATED_MIN_ID)
} else {
// After reversing the list the first element in the list is the lower range limit
await this.storage.setLowerRangeForList(typeRef, listId, getElementId(getFirstOrThrow(receivedEntities)))
}
} else {
// Last element in the list is the upper range limit
if (receivedEntities.length < countRequested) {
// all elements have been loaded, so the upper range must be set to MAX_ID
console.log("finished loading, setting max id")
await this.storage.setUpperRangeForList(typeRef, listId, isCustomId ? CUSTOM_MAX_ID : GENERATED_MAX_ID)
} else {
await this.storage.setUpperRangeForList(typeRef, listId, getElementId(lastThrow(receivedEntities)))
}
}
2022-12-27 15:37:40 +01:00
await Promise.all(elementsToAdd.map((element) => this.storage.put(element)))
}
/**
* Calculates the new start value for the getElementRange request and the number of elements to read in
* order to read no duplicate values.
* @return returns the new start and count value.
*/
2022-12-27 15:37:40 +01:00
private async recalculateRangeRequest<T extends ListElementEntity>(
typeRef: TypeRef<T>,
listId: Id,
start: Id,
count: number,
reverse: boolean,
): Promise<{ newStart: string; newCount: number }> {
let allRangeList = await this.storage.getIdsInRange(typeRef, listId)
let elementsToRead = count
let startElementId = start
const range = await this.storage.getRangeForList(typeRef, listId)
if (range == null) {
2022-12-27 15:37:40 +01:00
return { newStart: start, newCount: count }
}
2022-12-27 15:37:40 +01:00
const { lower, upper } = range
let indexOfStart = allRangeList.indexOf(start)
const typeModel = await resolveTypeReference(typeRef)
const isCustomId = isCustomIdType(typeModel)
if (
(!reverse && (isCustomId ? upper === CUSTOM_MAX_ID : upper === GENERATED_MAX_ID)) ||
(reverse && (isCustomId ? lower === CUSTOM_MIN_ID : lower === GENERATED_MIN_ID))
) {
// we have already loaded the complete range in the desired direction, so we do not have to load from server
elementsToRead = 0
2022-12-27 15:37:40 +01:00
} else if (allRangeList.length === 0) {
// Element range is empty, so read all elements
elementsToRead = count
2022-12-27 15:37:40 +01:00
} else if (indexOfStart !== -1) {
// Start element is located in allRange read only elements that are not in allRange.
if (reverse) {
elementsToRead = count - indexOfStart
startElementId = allRangeList[0] // use the lowest id in allRange as start element
} else {
elementsToRead = count - (allRangeList.length - 1 - indexOfStart)
startElementId = allRangeList[allRangeList.length - 1] // use the highest id in allRange as start element
}
} else if (lower === start || (firstBiggerThanSecond(start, lower, typeModel) && firstBiggerThanSecond(allRangeList[0], start, typeModel))) {
2022-12-27 15:37:40 +01:00
// Start element is not in allRange but has been used has start element for a range request, eg. EntityRestInterface.GENERATED_MIN_ID, or start is between lower range id and lowest element in range
if (!reverse) {
// if not reverse read only elements that are not in allRange
startElementId = allRangeList[allRangeList.length - 1] // use the highest id in allRange as start element
elementsToRead = count - allRangeList.length
}
// if reverse read all elements
} else if (
upper === start ||
(firstBiggerThanSecond(start, allRangeList[allRangeList.length - 1], typeModel) && firstBiggerThanSecond(upper, start, typeModel))
) {
2022-12-27 15:37:40 +01:00
// Start element is not in allRange but has been used has start element for a range request, eg. EntityRestInterface.GENERATED_MAX_ID, or start is between upper range id and highest element in range
if (reverse) {
// if not reverse read only elements that are not in allRange
startElementId = allRangeList[0] // use the highest id in allRange as start element
elementsToRead = count - allRangeList.length
}
// if not reverse read all elements
}
2022-12-27 15:37:40 +01:00
return { newStart: startElementId, newCount: elementsToRead }
}
/**
* Resolves when the entity is loaded from the server if necessary
* @pre The last call of this function must be resolved. This is needed to avoid that e.g. while
* loading a created instance from the server we receive an update of that instance and ignore it because the instance is not in the cache yet.
*
* @return Promise, which resolves to the array of valid events (if response is NotFound or NotAuthorized we filter it out)
*/
async entityEventsReceived(batch: QueuedBatch): Promise<Array<EntityUpdate>> {
await this.recordSyncTime()
// we handle post multiple create operations separately to optimize the number of requests with getMultiple
const createUpdatesForLETs: EntityUpdate[] = []
const regularUpdates: EntityUpdate[] = [] // all updates not resulting from post multiple requests
const updatesArray = batch.events
for (const update of updatesArray) {
const typeRef = new TypeRef(update.application, update.type)
// monitor application is ignored
if (update.application === "monitor") continue
// mails are ignored because move operations are handled as a special event (and no post multiple is possible)
if (update.operation === OperationType.CREATE && getUpdateInstanceId(update).instanceListId != null && !isSameTypeRef(typeRef, MailTypeRef)) {
createUpdatesForLETs.push(update)
} else {
regularUpdates.push(update)
await this.checkForMailSetMigration(typeRef, update)
}
}
const createUpdatesForLETsPerList = groupBy(createUpdatesForLETs, (update) => update.instanceListId)
const postMultipleEventUpdates: EntityUpdate[][] = []
// we first handle potential post multiple updates in get multiple requests
for (let [instanceListId, updates] of createUpdatesForLETsPerList) {
const firstUpdate = updates[0]
const typeRef = new TypeRef<ListElementEntity>(firstUpdate.application, firstUpdate.type)
2022-12-27 15:37:40 +01:00
const ids = updates.map((update) => update.instanceId)
// We only want to load the instances that are in cache range
Fix outdated mails from notification Notification process on mobile does insert Mail instances into offline db in order to show the mail preview as quickly as possible when the user clicks on notification. In most cases it happens before login, and it happens independently of entity update processing. EntityRestCache does only download new list element instances when they are within the cached range. In most cases new mails are within the cached range as mail indexer caches mail bag ranges during initial indexing. However, if the mail bag size has been reached there will be a new mail bag with mail list that is not cached by the client (there is no range stored for this new list). EventQueue optimizes event updates. If CREATE event is followed by an UPDATE event they will be folded into a single CREATE event. In a situation where the email has been updated by another client after it has been inserted into offline db by notification process but before the login, and that email belongs to uncached mail bag, CREATE event (and optimized away UPDATE events) for that mail would be skipped and the user would see an outdated Mail instance. We changed the behavior for Mail instances to be always downloaded on CREATE event when we use Offline cache so that we do not miss those update operations. This was the previously expected behavior because of mail indexing and is also a common behavior now, until new mail bag is created so it does not lead to additional requests in most cases. Alternatively we could check if the (new) instance is already cached as individual instance. We decided against that as it adds the delay to event processing and also does not fit the logic of processing (why should cache check if new instance is already cache?). In the future we should try to optimize loading of new instances so that the sync time does not increase from downloading single emails. Close #8041 Co-authored-by: ivk <ivk@tutao.de>
2024-12-17 15:46:05 +01:00
const customHandler = this.storage.getCustomCacheHandlerMap(this.entityRestClient).get(typeRef)
const idsInCacheRange =
customHandler && customHandler.getElementIdsInCacheRange
? await customHandler.getElementIdsInCacheRange(this.storage, instanceListId, ids)
: await this.getElementIdsInCacheRange(typeRef, instanceListId, ids)
if (idsInCacheRange.length === 0) {
postMultipleEventUpdates.push(updates)
} else {
2022-12-27 15:37:40 +01:00
const updatesNotInCacheRange =
idsInCacheRange.length === updates.length ? [] : updates.filter((update) => !idsInCacheRange.includes(update.instanceId))
try {
// loadMultiple is only called to cache the elements and check which ones return errors
const returnedInstances = await this._loadMultiple(typeRef, instanceListId, idsInCacheRange)
//We do not want to pass updates that caused an error
if (returnedInstances.length !== idsInCacheRange.length) {
2022-12-27 15:37:40 +01:00
const returnedIds = returnedInstances.map((instance) => getElementId(instance))
postMultipleEventUpdates.push(updates.filter((update) => returnedIds.includes(update.instanceId)).concat(updatesNotInCacheRange))
} else {
postMultipleEventUpdates.push(updates)
}
} catch (e) {
if (e instanceof NotAuthorizedError) {
// return updates that are not in cache Range if NotAuthorizedError (for those updates that are in cache range)
postMultipleEventUpdates.push(updatesNotInCacheRange)
} else {
throw e
}
}
}
}
const otherEventUpdates: EntityUpdate[] = []
for (let update of regularUpdates) {
2022-12-27 15:37:40 +01:00
const { operation, type, application } = update
const { instanceListId, instanceId } = getUpdateInstanceId(update)
const typeRef = new TypeRef<SomeEntity>(application, type)
switch (operation) {
case OperationType.UPDATE: {
const handledUpdate = await this.processUpdateEvent(typeRef, update)
if (handledUpdate) {
otherEventUpdates.push(handledUpdate)
}
break // do break instead of continue to avoid ide warnings
}
case OperationType.DELETE: {
2022-12-27 15:37:40 +01:00
if (
isSameTypeRef(MailTypeRef, typeRef) &&
containsEventOfType(updatesArray as Readonly<EntityUpdateData[]>, OperationType.CREATE, instanceId)
) {
// move for mail is handled in create event.
} else if (isSameTypeRef(MailTypeRef, typeRef)) {
// delete mailDetails if they are available (as we don't send an event for this type)
const mail = await this.storage.get(MailTypeRef, instanceListId, instanceId)
await this.storage.deleteIfExists(typeRef, instanceListId, instanceId)
if (mail?.mailDetails != null) {
await this.storage.deleteIfExists(MailDetailsBlobTypeRef, mail.mailDetails[0], mail.mailDetails[1])
}
} else {
await this.storage.deleteIfExists(typeRef, instanceListId, instanceId)
}
otherEventUpdates.push(update)
break // do break instead of continue to avoid ide warnings
}
case OperationType.CREATE: {
const handledUpdate = await this.processCreateEvent(typeRef, update, updatesArray)
if (handledUpdate) {
otherEventUpdates.push(handledUpdate)
}
break // do break instead of continue to avoid ide warnings
}
default:
throw new ProgrammingError("Unknown operation type: " + operation)
}
}
// the whole batch has been written successfully
await this.storage.putLastBatchIdForGroup(batch.groupId, batch.batchId)
// merge the results
return otherEventUpdates.concat(postMultipleEventUpdates.flat())
}
/** Returns {null} when the update should be skipped. */
2022-12-27 15:37:40 +01:00
private async processCreateEvent(typeRef: TypeRef<any>, update: EntityUpdate, batch: ReadonlyArray<EntityUpdate>): Promise<EntityUpdate | null> {
// do not return undefined to avoid implicit returns
const { instanceId, instanceListId } = getUpdateInstanceId(update)
// We put new instances into cache only when it's a new instance in the cached range which is only for the list instances.
if (instanceListId != null) {
const deleteEvent = getEventOfType(batch, OperationType.DELETE, instanceId)
const mail = deleteEvent && isSameTypeRef(MailTypeRef, typeRef) ? await this.storage.get(MailTypeRef, deleteEvent.instanceListId, instanceId) : null
// avoid downloading new mail element for non-mailSet user.
// can be removed once all mailbox have been migrated to mailSet (once lastNonOutdatedClientVersion is >= v242)
if (deleteEvent != null && mail != null && isEmpty(mail.sets)) {
// It is a move event for cached mail
await this.storage.deleteIfExists(typeRef, deleteEvent.instanceListId, instanceId)
await this.updateListIdOfMailAndUpdateCache(mail, instanceListId, instanceId)
return update
} else {
Fix outdated mails from notification Notification process on mobile does insert Mail instances into offline db in order to show the mail preview as quickly as possible when the user clicks on notification. In most cases it happens before login, and it happens independently of entity update processing. EntityRestCache does only download new list element instances when they are within the cached range. In most cases new mails are within the cached range as mail indexer caches mail bag ranges during initial indexing. However, if the mail bag size has been reached there will be a new mail bag with mail list that is not cached by the client (there is no range stored for this new list). EventQueue optimizes event updates. If CREATE event is followed by an UPDATE event they will be folded into a single CREATE event. In a situation where the email has been updated by another client after it has been inserted into offline db by notification process but before the login, and that email belongs to uncached mail bag, CREATE event (and optimized away UPDATE events) for that mail would be skipped and the user would see an outdated Mail instance. We changed the behavior for Mail instances to be always downloaded on CREATE event when we use Offline cache so that we do not miss those update operations. This was the previously expected behavior because of mail indexing and is also a common behavior now, until new mail bag is created so it does not lead to additional requests in most cases. Alternatively we could check if the (new) instance is already cached as individual instance. We decided against that as it adds the delay to event processing and also does not fit the logic of processing (why should cache check if new instance is already cache?). In the future we should try to optimize loading of new instances so that the sync time does not increase from downloading single emails. Close #8041 Co-authored-by: ivk <ivk@tutao.de>
2024-12-17 15:46:05 +01:00
// If there is a custom handler we follow its decision.
// Otherwise, we do a range check to see if we need to keep the range up-to-date.
const shouldLoad =
(await this.storage.getCustomCacheHandlerMap(this.entityRestClient).get(typeRef)?.shouldLoadOnCreateEvent?.(update)) ??
(await this.storage.isElementIdInCacheRange(typeRef, instanceListId, instanceId))
if (shouldLoad) {
// No need to try to download something that's not there anymore
// We do not consult custom handlers here because they are only needed for list elements.
console.log("downloading create event for", getTypeId(typeRef), instanceListId, instanceId)
return this.entityRestClient
.load(typeRef, [instanceListId, instanceId])
.then((entity) => this.storage.put(entity))
.then(() => update)
.catch((e) => {
if (isExpectedErrorForSynchronization(e)) {
return null
} else {
throw e
}
})
} else {
return update
}
}
} else {
return update
}
}
/**
* Updates the given mail with the new list id and add it to the cache.
*/
private async updateListIdOfMailAndUpdateCache(mail: Mail, newListId: Id, elementId: Id) {
// In case of a move operation we have to replace the list id always, as the mail is stored in another folder.
mail._id = [newListId, elementId]
if (mail.bucketKey != null) {
// With the simplified permission system (MailDetails) we also have to update the bucketEncSessionKey for the mail,
// which also references the mail list id. We need this for some cases when the move operation was executed
// before the UpdateSessionKeyService has been executed, e.g. when using inbox rules.
// The UpdateSessionKeyService would remove the bucketKey from the mail and there is no need to synchronize it anymore.
const mailSessionKey = mail.bucketKey.bucketEncSessionKeys.find((bucketEncSessionKey) => isSameId(bucketEncSessionKey.instanceId, elementId))
if (mailSessionKey) {
mailSessionKey.instanceList = newListId
}
}
await this.storage.put(mail)
}
/** Returns {null} when the update should be skipped. */
private async processUpdateEvent(typeRef: TypeRef<SomeEntity>, update: EntityUpdate): Promise<EntityUpdate | null> {
2022-12-27 15:37:40 +01:00
const { instanceId, instanceListId } = getUpdateInstanceId(update)
const cached = await this.storage.get(typeRef, instanceListId, instanceId)
// No need to try to download something that's not there anymore
if (cached != null) {
try {
2023-03-22 15:37:54 +01:00
// in case this is an update for the user instance: if the password changed we'll be logged out at this point
// if we don't catch the expected NotAuthenticated Error that results from trying to load anything with
// the old user.
// Letting the NotAuthenticatedError propagate to the main thread instead of trying to handle it ourselves
// or throwing out the update drops us onto the login page and into the session recovery flow if the user
// clicks their saved credentials again, but lets them still use offline login if they try to use the
// outdated credentials while not connected to the internet.
const newEntity = await this.entityRestClient.load(typeRef, collapseId(instanceListId, instanceId))
if (isSameTypeRef(typeRef, UserTypeRef)) {
await this.handleUpdatedUser(cached, newEntity)
}
await this.storage.put(newEntity)
return update
} catch (e) {
// If the entity is not there anymore we should evict it from the cache and not keep the outdated/nonexisting instance around.
// Even for list elements this should be safe as the instance is not there anymore and is definitely not in this version
if (isExpectedErrorForSynchronization(e)) {
console.log(`Instance not found when processing update for ${JSON.stringify(update)}, deleting from the cache.`)
await this.storage.deleteIfExists(typeRef, instanceListId, instanceId)
return null
} else {
throw e
}
}
}
return update
}
private async handleUpdatedUser(cached: SomeEntity, newEntity: SomeEntity) {
// When we are removed from a group we just get an update for our user
// with no membership on it. We need to clean up all the entities that
// belong to that group since we shouldn't be able to access them anymore
// and we won't get any update or another chance to clean them up.
const oldUser = cached as User
if (oldUser._id !== this.storage.getUserId()) {
return
}
const newUser = newEntity as User
const removedShips = difference(oldUser.memberships, newUser.memberships, (l, r) => l._id === r._id)
for (const ship of removedShips) {
console.log("Lost membership on ", ship._id, ship.groupType)
await this.storage.deleteAllOwnedBy(ship.group)
}
}
/**
*
* @returns {Array<Id>} the ids that are in cache range and therefore should be cached
*/
private async getElementIdsInCacheRange<T extends ListElementEntity>(typeRef: TypeRef<T>, listId: Id, ids: Id[]): Promise<Id[]> {
const ret: Id[] = []
for (let i = 0; i < ids.length; i++) {
if (await this.storage.isElementIdInCacheRange(typeRef, listId, ids[i])) {
ret.push(ids[i])
}
}
return ret
}
/**
* to avoid excessive entity updates and inconsistent offline storages, we don't send entity updates for each mail set migrated mail.
* instead we detect the mail set migration for each folder and drop its whole list from offline.
*/
private async checkForMailSetMigration(typeRef: TypeRef<unknown>, update: EntityUpdate): Promise<void> {
if (update.operation !== OperationType.UPDATE || !isSameTypeRef(typeRef, MailFolderTypeRef)) return
// load the old version of the folder now to check if it was migrated to mail set
const oldFolder = await this.storage.get(MailFolderTypeRef, update.instanceListId, update.instanceId)
// if it already is a mail set, we're done.
// we also delete the mails in the case where we don't have the folder itself in the cache.
// because we cache after loading the folder, we won't do it again on the next update event.
if (oldFolder != null && oldFolder.isMailSet) return
const updatedFolder = await this.entityRestClient.load(MailFolderTypeRef, [update.instanceListId, update.instanceId])
if (!updatedFolder.isMailSet) return
await this.storage.deleteWholeList(MailTypeRef, updatedFolder.mails)
await this.storage.put(updatedFolder)
}
}
/**
* Returns whether the error is expected for the cases where our local state might not be up-to-date with the server yet. E.g. we might be processing an update
* for the instance that was already deleted. Normally this would be optimized away but it might still happen due to timing.
*/
function isExpectedErrorForSynchronization(e: Error): boolean {
return e instanceof NotFoundError || e instanceof NotAuthorizedError
}
2022-12-27 15:37:40 +01:00
export function expandId(id: Id | IdTuple): { listId: Id | null; elementId: Id } {
if (typeof id === "string") {
return {
listId: null,
2022-12-27 15:37:40 +01:00
elementId: id,
}
} else {
const [listId, elementId] = id
return {
2022-12-27 15:37:40 +01:00
listId,
elementId,
}
}
}
2021-12-23 14:03:23 +01:00
export function collapseId(listId: Id | null, elementId: Id): Id | IdTuple {
if (listId != null) {
return [listId, elementId]
} else {
return elementId
}
}
2022-12-27 15:37:40 +01:00
export function getUpdateInstanceId(update: EntityUpdate): { instanceListId: Id | null; instanceId: Id } {
let instanceListId
if (update.instanceListId === "") {
instanceListId = null
} else {
instanceListId = update.instanceListId
}
2022-12-27 15:37:40 +01:00
return { instanceListId, instanceId: update.instanceId }
}
/**
* Check if a range request begins inside an existing range
*/
function isStartIdWithinRange(range: Range, startId: Id, typeModel: TypeModel): boolean {
return !firstBiggerThanSecond(startId, range.upper, typeModel) && !firstBiggerThanSecond(range.lower, startId, typeModel)
}
/**
* Check if a range request is going away from an existing range
* Assumes that the range request doesn't start inside the range
*/
function isRangeRequestAwayFromExistingRange(range: Range, reverse: boolean, start: string, typeModel: TypeModel) {
return reverse ? firstBiggerThanSecond(range.lower, start, typeModel) : firstBiggerThanSecond(start, range.upper, typeModel)
}
/**
* some types are completely ignored by the cache and always served from a request.
* Note:
* isCachedType(ref) ---> !isIgnoredType(ref) but
* isIgnoredType(ref) -/-> !isCachedType(ref) because of opted-in CustomId types.
*/
function isIgnoredType(typeRef: TypeRef<unknown>): boolean {
2022-12-27 15:37:40 +01:00
return typeRef.app === "monitor" || IGNORED_TYPES.some((ref) => isSameTypeRef(typeRef, ref))
}
/**
* Checks if for the given type, that contains a customId, caching is enabled.
*/
function isCachableCustomIdType(typeRef: TypeRef<unknown>): boolean {
return CACHABLE_CUSTOMID_TYPES.some((ref) => isSameTypeRef(typeRef, ref))
}
/**
* customId types are normally not cached, but some are opted in.
* Note:
* isCachedType(ref) ---> !isIgnoredType(ref) but
* isIgnoredType(ref) -/-> !isCachedType(ref)
*/
function isCachedType(typeModel: TypeModel, typeRef: TypeRef<unknown>): boolean {
return (!isIgnoredType(typeRef) && isGeneratedIdType(typeModel)) || isCachableCustomIdType(typeRef)
}
function isGeneratedIdType(typeModel: TypeModel): boolean {
return typeModel.values._id.type === ValueType.GeneratedId
2022-12-27 15:37:40 +01:00
}