/** * Firestore Lite * * @remarks Firestore Lite is a small online-only SDK that allows read * and write access to your Firestore database. All operations connect * directly to the backend, and `onSnapshot()` APIs are not supported. * @packageDocumentation */ import { DocumentData as DocumentData_2 } from '@firebase/firestore-types'; import { EmulatorMockTokenOptions } from '@firebase/util'; import { FirebaseApp } from '@firebase/app'; import { FirebaseError } from '@firebase/util'; import { LogLevelString as LogLevel } from '@firebase/logger'; import { SetOptions as SetOptions_2 } from '@firebase/firestore-types'; /* Excluded from this release type: AbstractUserDataWriter */ /** * Describes a map whose keys are active target ids. We do not care about the type of the * values. */ declare type ActiveTargets = SortedMap; /** * Add a new document to specified `CollectionReference` with the given data, * assigning it a document ID automatically. * * The result of this write will only be reflected in document reads that occur * after the returned promise resolves. If the client is offline, the * write fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @param reference - A reference to the collection to add this document to. * @param data - An Object containing the data for the new document. * @throws Error - If the provided input is not a valid Firestore document. * @returns A `Promise` resolved with a `DocumentReference` pointing to the * newly created document after it has been written to the backend. */ export declare function addDoc(reference: CollectionReference, data: WithFieldValue): Promise>; /** * Returns a new map where every key is prefixed with the outer key appended * to a dot. */ export declare type AddPrefixToKeys> = { [K in keyof T & string as `${Prefix}.${K}`]+?: string extends K ? any : T[K]; }; /** * Represents an aggregation that can be performed by Firestore. */ export declare class AggregateField { readonly _internalFieldPath?: FieldPath_2 | undefined; /** A type string to uniquely identify instances of this class. */ readonly type = "AggregateField"; /** Indicates the aggregation operation of this AggregateField. */ readonly aggregateType: AggregateType; /* Excluded from this release type: __constructor */ } /** * Compares two 'AggregateField` instances for equality. * * @param left Compare this AggregateField to the `right`. * @param right Compare this AggregateField to the `left`. */ export declare function aggregateFieldEqual(left: AggregateField, right: AggregateField): boolean; /** * The union of all `AggregateField` types that are supported by Firestore. */ export declare type AggregateFieldType = ReturnType | ReturnType | ReturnType; /** * The results of executing an aggregation query. */ export declare class AggregateQuerySnapshot { private readonly _userDataWriter; private readonly _data; /** A type string to uniquely identify instances of this class. */ readonly type = "AggregateQuerySnapshot"; /** * The underlying query over which the aggregations recorded in this * `AggregateQuerySnapshot` were performed. */ readonly query: Query; /** @hideconstructor */ constructor(query: Query, _userDataWriter: AbstractUserDataWriter, _data: ApiClientObjectMap); /** * Returns the results of the aggregations performed over the underlying * query. * * The keys of the returned object will be the same as those of the * `AggregateSpec` object specified to the aggregation method, and the values * will be the corresponding aggregation result. * * @returns The results of the aggregations performed over the underlying * query. */ data(): AggregateSpecData; } /** * Compares two `AggregateQuerySnapshot` instances for equality. * * Two `AggregateQuerySnapshot` instances are considered "equal" if they have * underlying queries that compare equal, and the same data. * * @param left - The first `AggregateQuerySnapshot` to compare. * @param right - The second `AggregateQuerySnapshot` to compare. * * @returns `true` if the objects are "equal", as defined above, or `false` * otherwise. */ export declare function aggregateQuerySnapshotEqual(left: AggregateQuerySnapshot, right: AggregateQuerySnapshot): boolean; /** * Specifies a set of aggregations and their aliases. */ export declare interface AggregateSpec { [field: string]: AggregateFieldType; } /** * A type whose keys are taken from an `AggregateSpec`, and whose values are the * result of the aggregation performed by the corresponding `AggregateField` * from the input `AggregateSpec`. */ export declare type AggregateSpecData = { [P in keyof T]: T[P] extends AggregateField ? U : never; }; /** * Union type representing the aggregate type to be performed. */ export declare type AggregateType = 'count' | 'avg' | 'sum'; /** * Creates a new {@link QueryCompositeFilterConstraint} that is a conjunction of * the given filter constraints. A conjunction filter includes a document if it * satisfies all of the given filters. * * @param queryConstraints - Optional. The list of * {@link QueryFilterConstraint}s to perform a conjunction for. These must be * created with calls to {@link where}, {@link or}, or {@link and}. * @returns The newly created {@link QueryCompositeFilterConstraint}. */ export declare function and(...queryConstraints: QueryFilterConstraint[]): QueryCompositeFilterConstraint; declare interface ApiClientObjectMap { [k: string]: T; } /** * An `AppliableConstraint` is an abstraction of a constraint that can be applied * to a Firestore query. */ declare abstract class AppliableConstraint { /** * Takes the provided {@link Query} and returns a copy of the {@link Query} with this * {@link AppliableConstraint} applied. */ abstract _apply(query: Query): Query; } /** * Returns a special value that can be used with {@link (setDoc:1)} or {@link * updateDoc:1} that tells the server to remove the given elements from any * array value that already exists on the server. All instances of each element * specified will be removed from the array. If the field being modified is not * already an array it will be overwritten with an empty array. * * @param elements - The elements to remove from the array. * @returns The `FieldValue` sentinel for use in a call to `setDoc()` or * `updateDoc()` */ export declare function arrayRemove(...elements: unknown[]): FieldValue; /** * Returns a special value that can be used with {@link @firebase/firestore/lite#(setDoc:1)} or {@link * @firebase/firestore/lite#(updateDoc:1)} that tells the server to union the given elements with any array * value that already exists on the server. Each specified element that doesn't * already exist in the array will be added to the end. If the field being * modified is not already an array it will be overwritten with an array * containing exactly the specified elements. * * @param elements - The elements to union into the array. * @returns The `FieldValue` sentinel for use in a call to `setDoc()` or * `updateDoc()`. */ export declare function arrayUnion(...elements: unknown[]): FieldValue; declare interface AsyncQueue { readonly isShuttingDown: boolean; /** * Adds a new operation to the queue without waiting for it to complete (i.e. * we ignore the Promise result). */ enqueueAndForget(op: () => Promise): void; /** * Regardless if the queue has initialized shutdown, adds a new operation to the * queue without waiting for it to complete (i.e. we ignore the Promise result). */ enqueueAndForgetEvenWhileRestricted(op: () => Promise): void; /** * Initialize the shutdown of this queue. Once this method is called, the * only possible way to request running an operation is through * `enqueueEvenWhileRestricted()`. * * @param purgeExistingTasks Whether already enqueued tasked should be * rejected (unless enqueued wih `enqueueEvenWhileRestricted()`). Defaults * to false. */ enterRestrictedMode(purgeExistingTasks?: boolean): void; /** * Adds a new operation to the queue. Returns a promise that will be resolved * when the promise returned by the new operation is (with its value). */ enqueue(op: () => Promise): Promise; /** * Enqueue a retryable operation. * * A retryable operation is rescheduled with backoff if it fails with a * IndexedDbTransactionError (the error type used by SimpleDb). All * retryable operations are executed in order and only run if all prior * operations were retried successfully. */ enqueueRetryable(op: () => Promise): void; /** * Schedules an operation to be queued on the AsyncQueue once the specified * `delayMs` has elapsed. The returned DelayedOperation can be used to cancel * or fast-forward the operation prior to its running. */ enqueueAfterDelay(timerId: TimerId, delayMs: number, op: () => Promise): DelayedOperation; /** * Verifies there's an operation currently in-progress on the AsyncQueue. * Unfortunately we can't verify that the running code is in the promise chain * of that operation, so this isn't a foolproof check, but it should be enough * to catch some bugs. */ verifyOperationInProgress(): void; } /* Excluded from this release type: AuthTokenFactory */ /** * Create an AggregateField object that can be used to compute the average of * a specified field over a range of documents in the result set of a query. * @param field Specifies the field to average across the result set. */ export declare function average(field: string | FieldPath): AggregateField; /** * Path represents an ordered sequence of string segments. */ declare abstract class BasePath> { private segments; private offset; private len; constructor(segments: string[], offset?: number, length?: number); /** * Abstract constructor method to construct an instance of B with the given * parameters. */ protected abstract construct(segments: string[], offset?: number, length?: number): B; /** * Returns a String representation. * * Implementing classes are required to provide deterministic implementations as * the String representation is used to obtain canonical Query IDs. */ abstract toString(): string; get length(): number; isEqual(other: B): boolean; child(nameOrPath: string | B): B; /** The index of one past the last segment of the path. */ private limit; popFirst(size?: number): B; popLast(): B; firstSegment(): string; lastSegment(): string; get(index: number): string; isEmpty(): boolean; isPrefixOf(other: this): boolean; isImmediateParentOf(potentialChild: this): boolean; forEach(fn: (segment: string) => void): void; toArray(): string[]; static comparator>(p1: BasePath, p2: BasePath): number; } /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * BatchID is a locally assigned ID for a batch of mutations that have been * applied. */ declare type BatchId = number; /** * Represents a bound of a query. * * The bound is specified with the given components representing a position and * whether it's just before or just after the position (relative to whatever the * query order is). * * The position represents a logical index position for a query. It's a prefix * of values for the (potentially implicit) order by clauses of a query. * * Bound provides a function to determine whether a document comes before or * after a bound. This is influenced by whether the position is just before or * just after the provided values. */ declare class Bound { readonly position: Value[]; readonly inclusive: boolean; constructor(position: Value[], inclusive: boolean); } /** * Provides interfaces to save and read Firestore bundles. */ declare interface BundleCache { /** * Gets the saved `BundleMetadata` for a given `bundleId`, returns undefined * if no bundle metadata is found under the given id. */ getBundleMetadata(transaction: PersistenceTransaction, bundleId: string): PersistencePromise; /** * Saves a `BundleMetadata` from a bundle into local storage, using its id as * the persistent key. */ saveBundleMetadata(transaction: PersistenceTransaction, metadata: BundleMetadata_2): PersistencePromise; /** * Gets a saved `NamedQuery` for the given query name. Returns undefined if * no queries are found under the given name. */ getNamedQuery(transaction: PersistenceTransaction, queryName: string): PersistencePromise; /** * Saves a `NamedQuery` from a bundle, using its name as the persistent key. */ saveNamedQuery(transaction: PersistenceTransaction, query: NamedQuery_2): PersistencePromise; } /** Properties of a BundledQuery. */ declare interface BundledQuery { /** BundledQuery parent */ parent?: string | null; /** BundledQuery structuredQuery */ structuredQuery?: StructuredQuery | null; /** BundledQuery limitType */ limitType?: LimitType_2 | null; } /** * Represents a Firestore bundle saved by the SDK in its local storage. */ declare interface BundleMetadata { /** * Id of the bundle. It is used together with `createTime` to determine if a * bundle has been loaded by the SDK. */ readonly id: string; /** Schema version of the bundle. */ readonly version: number; /** * Set to the snapshot version of the bundle if created by the Server SDKs. * Otherwise set to SnapshotVersion.MIN. */ readonly createTime: SnapshotVersion; } /** Properties of a BundleMetadata. */ declare interface BundleMetadata_2 { /** BundleMetadata id */ id?: string | null; /** BundleMetadata createTime */ createTime?: Timestamp_2 | null; /** BundleMetadata version */ version?: number | null; /** BundleMetadata totalDocuments */ totalDocuments?: number | null; /** BundleMetadata totalBytes */ totalBytes?: number | null; } /** * An immutable object representing an array of bytes. */ export declare class Bytes { _byteString: ByteString; /** @hideconstructor */ constructor(byteString: ByteString); /** * Creates a new `Bytes` object from the given Base64 string, converting it to * bytes. * * @param base64 - The Base64 string used to create the `Bytes` object. */ static fromBase64String(base64: string): Bytes; /** * Creates a new `Bytes` object from the given Uint8Array. * * @param array - The Uint8Array used to create the `Bytes` object. */ static fromUint8Array(array: Uint8Array): Bytes; /** * Returns the underlying bytes as a Base64-encoded string. * * @returns The Base64-encoded string created from the `Bytes` object. */ toBase64(): string; /** * Returns the underlying bytes in a new `Uint8Array`. * * @returns The Uint8Array created from the `Bytes` object. */ toUint8Array(): Uint8Array; /** * Returns a string representation of the `Bytes` object. * * @returns A string representation of the `Bytes` object. */ toString(): string; /** * Returns true if this `Bytes` object is equal to the provided one. * * @param other - The `Bytes` object to compare against. * @returns true if this `Bytes` object is equal to the provided one. */ isEqual(other: Bytes): boolean; } /* Excluded from this release type: ByteString */ declare const enum ChangeType { Added = 0, Removed = 1, Modified = 2, Metadata = 3 } /** * Helper for calculating the nested fields for a given type T1. This is needed * to distribute union types such as `undefined | {...}` (happens for optional * props) or `{a: A} | {b: B}`. * * In this use case, `V` is used to distribute the union types of `T[K]` on * `Record`, since `T[K]` is evaluated as an expression and not distributed. * * See https://www.typescriptlang.org/docs/handbook/advanced-types.html#distributive-conditional-types */ export declare type ChildUpdateFields = V extends Record ? AddPrefixToKeys> : never; /** * A randomly-generated key assigned to each Firestore instance at startup. */ declare type ClientId = string; /** * Gets a `CollectionReference` instance that refers to the collection at * the specified absolute path. * * @param firestore - A reference to the root `Firestore` instance. * @param path - A slash-separated path to a collection. * @param pathSegments - Additional path segments to apply relative to the first * argument. * @throws If the final path has an even number of segments and does not point * to a collection. * @returns The `CollectionReference` instance. */ export declare function collection(firestore: Firestore, path: string, ...pathSegments: string[]): CollectionReference; /** * Gets a `CollectionReference` instance that refers to a subcollection of * `reference` at the the specified relative path. * * @param reference - A reference to a collection. * @param path - A slash-separated path to a collection. * @param pathSegments - Additional path segments to apply relative to the first * argument. * @throws If the final path has an even number of segments and does not point * to a collection. * @returns The `CollectionReference` instance. */ export declare function collection(reference: CollectionReference, path: string, ...pathSegments: string[]): CollectionReference; /** * Gets a `CollectionReference` instance that refers to a subcollection of * `reference` at the the specified relative path. * * @param reference - A reference to a Firestore document. * @param path - A slash-separated path to a collection. * @param pathSegments - Additional path segments that will be applied relative * to the first argument. * @throws If the final path has an even number of segments and does not point * to a collection. * @returns The `CollectionReference` instance. */ export declare function collection(reference: DocumentReference, path: string, ...pathSegments: string[]): CollectionReference; /** * Creates and returns a new `Query` instance that includes all documents in the * database that are contained in a collection or subcollection with the * given `collectionId`. * * @param firestore - A reference to the root `Firestore` instance. * @param collectionId - Identifies the collections to query over. Every * collection or subcollection with this ID as the last segment of its path * will be included. Cannot contain a slash. * @returns The created `Query`. */ export declare function collectionGroup(firestore: Firestore, collectionId: string): Query; /** * A `CollectionReference` object can be used for adding documents, getting * document references, and querying for documents (using {@link (query:1)}). */ export declare class CollectionReference extends Query { readonly _path: ResourcePath; /** The type of this Firestore reference. */ readonly type = "collection"; /** @hideconstructor */ constructor(firestore: Firestore, converter: FirestoreDataConverter | null, _path: ResourcePath); /** The collection's identifier. */ get id(): string; /** * A string representing the path of the referenced collection (relative * to the root of the database). */ get path(): string; /** * A reference to the containing `DocumentReference` if this is a * subcollection. If this isn't a subcollection, the reference is null. */ get parent(): DocumentReference | null; /** * Applies a custom data converter to this `CollectionReference`, allowing you * to use your own custom model objects with Firestore. When you call {@link * addDoc} with the returned `CollectionReference` instance, the provided * converter will convert between Firestore data of type `NewDbModelType` and * your custom type `NewAppModelType`. * * @param converter - Converts objects to and from Firestore. * @returns A `CollectionReference` that uses the provided converter. */ withConverter(converter: FirestoreDataConverter): CollectionReference; /** * Removes the current converter. * * @param converter - `null` removes the current converter. * @returns A `CollectionReference` that does not * use a converter. */ withConverter(converter: null): CollectionReference; } /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ declare type Comparator = (key1: K, key2: K) => number; declare interface ComponentConfiguration { asyncQueue: AsyncQueue; databaseInfo: DatabaseInfo; authCredentials: CredentialsProvider; appCheckCredentials: CredentialsProvider; clientId: ClientId; initialUser: User; maxConcurrentLimboResolutions: number; } declare type CompositeFilterOp = 'OPERATOR_UNSPECIFIED' | 'AND' | 'OR'; declare const enum CompositeOperator { OR = "or", AND = "and" } /** * Modify this instance to communicate with the Cloud Firestore emulator. * * Note: This must be called before this instance has been used to do any * operations. * * @param firestore - The `Firestore` instance to configure to connect to the * emulator. * @param host - the emulator host (ex: localhost). * @param port - the emulator port (ex: 9000). * @param options.mockUserToken - the mock auth token to use for unit testing * Security Rules. */ export declare function connectFirestoreEmulator(firestore: Firestore, host: string, port: number, options?: { mockUserToken?: EmulatorMockTokenOptions | string; }): void; /** * Create an AggregateField object that can be used to compute the count of * documents in the result set of a query. */ export declare function count(): AggregateField; /** * A Listener for credential change events. The listener should fetch a new * token and may need to invalidate other state if the current user has also * changed. */ declare type CredentialChangeListener = (credential: T) => Promise; /** * Provides methods for getting the uid and token for the current user and * listening for changes. */ declare interface CredentialsProvider { /** * Starts the credentials provider and specifies a listener to be notified of * credential changes (sign-in / sign-out, token changes). It is immediately * called once with the initial user. * * The change listener is invoked on the provided AsyncQueue. */ start(asyncQueue: AsyncQueue, changeListener: CredentialChangeListener): void; /** Requests a token for the current user. */ getToken(): Promise; /** * Marks the last retrieved token as invalid, making the next GetToken request * force-refresh the token. */ invalidateToken(): void; shutdown(): void; } /** Settings for private credentials */ declare type CredentialsSettings = FirstPartyCredentialsSettings | ProviderCredentialsSettings; /* Excluded from this release type: DatabaseId */ /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ declare class DatabaseInfo { readonly databaseId: DatabaseId; readonly appId: string; readonly persistenceKey: string; readonly host: string; readonly ssl: boolean; readonly forceLongPolling: boolean; readonly autoDetectLongPolling: boolean; readonly longPollingOptions: ExperimentalLongPollingOptions; readonly useFetchStreams: boolean; /** * Constructs a DatabaseInfo using the provided host, databaseId and * persistenceKey. * * @param databaseId - The database to use. * @param appId - The Firebase App Id. * @param persistenceKey - A unique identifier for this Firestore's local * storage (used in conjunction with the databaseId). * @param host - The Firestore backend host to connect to. * @param ssl - Whether to use SSL when connecting. * @param forceLongPolling - Whether to use the forceLongPolling option * when using WebChannel as the network transport. * @param autoDetectLongPolling - Whether to use the detectBufferingProxy * option when using WebChannel as the network transport. * @param longPollingOptions Options that configure long-polling. * @param useFetchStreams Whether to use the Fetch API instead of * XMLHTTPRequest */ constructor(databaseId: DatabaseId, appId: string, persistenceKey: string, host: string, ssl: boolean, forceLongPolling: boolean, autoDetectLongPolling: boolean, longPollingOptions: ExperimentalLongPollingOptions, useFetchStreams: boolean); } /** * Datastore and its related methods are a wrapper around the external Google * Cloud Datastore grpc API, which provides an interface that is more convenient * for the rest of the client SDK architecture to consume. */ declare abstract class Datastore { abstract terminate(): void; abstract serializer: JsonProtoSerializer; } /** * Represents an operation scheduled to be run in the future on an AsyncQueue. * * It is created via DelayedOperation.createAndSchedule(). * * Supports cancellation (via cancel()) and early execution (via skipDelay()). * * Note: We implement `PromiseLike` instead of `Promise`, as the `Promise` type * in newer versions of TypeScript defines `finally`, which is not available in * IE. */ declare class DelayedOperation implements PromiseLike { private readonly asyncQueue; readonly timerId: TimerId; readonly targetTimeMs: number; private readonly op; private readonly removalCallback; private timerHandle; private readonly deferred; private constructor(); get promise(): Promise; /** * Creates and returns a DelayedOperation that has been scheduled to be * executed on the provided asyncQueue after the provided delayMs. * * @param asyncQueue - The queue to schedule the operation on. * @param id - A Timer ID identifying the type of operation this is. * @param delayMs - The delay (ms) before the operation should be scheduled. * @param op - The operation to run. * @param removalCallback - A callback to be called synchronously once the * operation is executed or canceled, notifying the AsyncQueue to remove it * from its delayedOperations list. * PORTING NOTE: This exists to prevent making removeDelayedOperation() and * the DelayedOperation class public. */ static createAndSchedule(asyncQueue: AsyncQueue, timerId: TimerId, delayMs: number, op: () => Promise, removalCallback: (op: DelayedOperation) => void): DelayedOperation; /** * Starts the timer. This is called immediately after construction by * createAndSchedule(). */ private start; /** * Queues the operation to run immediately (if it hasn't already been run or * canceled). */ skipDelay(): void; /** * Cancels the operation if it hasn't already been executed or canceled. The * promise will be rejected. * * As long as the operation has not yet been run, calling cancel() provides a * guarantee that the operation will not be run. */ cancel(reason?: string): void; then: (onfulfilled?: ((value: T) => TResult1 | PromiseLike) | null | undefined, onrejected?: ((reason: any) => TResult2 | PromiseLike) | null | undefined) => Promise; private handleDelayElapsed; private clearTimeout; } /** * Deletes the document referred to by the specified `DocumentReference`. * * The deletion will only be reflected in document reads that occur after the * returned promise resolves. If the client is offline, the * delete fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @param reference - A reference to the document to delete. * @returns A `Promise` resolved once the document has been successfully * deleted from the backend. */ export declare function deleteDoc(reference: DocumentReference): Promise; /** * Returns a sentinel for use with {@link @firebase/firestore/lite#(updateDoc:1)} or * {@link @firebase/firestore/lite#(setDoc:1)} with `{merge: true}` to mark a field for deletion. */ export declare function deleteField(): FieldValue; /** * The direction of sorting in an order by. */ declare const enum Direction { ASCENDING = "asc", DESCENDING = "desc" } /** * Gets a `DocumentReference` instance that refers to the document at the * specified absolute path. * * @param firestore - A reference to the root `Firestore` instance. * @param path - A slash-separated path to a document. * @param pathSegments - Additional path segments that will be applied relative * to the first argument. * @throws If the final path has an odd number of segments and does not point to * a document. * @returns The `DocumentReference` instance. */ export declare function doc(firestore: Firestore, path: string, ...pathSegments: string[]): DocumentReference; /** * Gets a `DocumentReference` instance that refers to a document within * `reference` at the specified relative path. If no path is specified, an * automatically-generated unique ID will be used for the returned * `DocumentReference`. * * @param reference - A reference to a collection. * @param path - A slash-separated path to a document. Has to be omitted to use * auto-genrated IDs. * @param pathSegments - Additional path segments that will be applied relative * to the first argument. * @throws If the final path has an odd number of segments and does not point to * a document. * @returns The `DocumentReference` instance. */ export declare function doc(reference: CollectionReference, path?: string, ...pathSegments: string[]): DocumentReference; /** * Gets a `DocumentReference` instance that refers to a document within * `reference` at the specified relative path. * * @param reference - A reference to a Firestore document. * @param path - A slash-separated path to a document. * @param pathSegments - Additional path segments that will be applied relative * to the first argument. * @throws If the final path has an odd number of segments and does not point to * a document. * @returns The `DocumentReference` instance. */ export declare function doc(reference: DocumentReference, path: string, ...pathSegments: string[]): DocumentReference; /** * Represents a document in Firestore with a key, version, data and whether the * data has local mutations applied to it. */ declare interface Document_2 { /** The key for this document */ readonly key: DocumentKey; /** * The version of this document if it exists or a version at which this * document was guaranteed to not exist. */ readonly version: SnapshotVersion; /** * The timestamp at which this document was read from the remote server. Uses * `SnapshotVersion.min()` for documents created by the user. */ readonly readTime: SnapshotVersion; /** * The timestamp at which the document was created. This value increases * monotonically when a document is deleted then recreated. It can also be * compared to `createTime` of other documents and the `readTime` of a query. */ readonly createTime: SnapshotVersion; /** The underlying data of this document or an empty value if no data exists. */ readonly data: ObjectValue; /** Returns whether local mutations were applied via the mutation queue. */ readonly hasLocalMutations: boolean; /** Returns whether mutations were applied based on a write acknowledgment. */ readonly hasCommittedMutations: boolean; /** * Whether this document had a local mutation applied that has not yet been * acknowledged by Watch. */ readonly hasPendingWrites: boolean; /** * Returns whether this document is valid (i.e. it is an entry in the * RemoteDocumentCache, was created by a mutation or read from the backend). */ isValidDocument(): boolean; /** * Returns whether the document exists and its data is known at the current * version. */ isFoundDocument(): boolean; /** * Returns whether the document is known to not exist at the current version. */ isNoDocument(): boolean; /** * Returns whether the document exists and its data is unknown at the current * version. */ isUnknownDocument(): boolean; isEqual(other: Document_2 | null | undefined): boolean; /** Creates a mutable copy of this document. */ mutableCopy(): MutableDocument; toString(): string; } declare type DocumentComparator = (doc1: Document_2, doc2: Document_2) => number; /** * Document data (for use with {@link @firebase/firestore/lite#(setDoc:1)}) consists of fields mapped to * values. */ export declare interface DocumentData { /** A mapping between a field and its value. */ [field: string]: any; } /** * Returns a special sentinel `FieldPath` to refer to the ID of a document. * It can be used in queries to sort or filter by the document ID. */ export declare function documentId(): FieldPath; /* Excluded from this release type: DocumentKey */ declare type DocumentKeyMap = ObjectMap; declare type DocumentKeySet = SortedSet; declare type DocumentMap = SortedMap; /** * Provides methods to read and write document overlays. * * An overlay is a saved mutation, that gives a local view of a document when * applied to the remote version of the document. * * Each overlay stores the largest batch ID that is included in the overlay, * which allows us to remove the overlay once all batches leading up to it have * been acknowledged. */ declare interface DocumentOverlayCache { /** * Gets the saved overlay mutation for the given document key. * Returns null if there is no overlay for that key. */ getOverlay(transaction: PersistenceTransaction, key: DocumentKey): PersistencePromise; /** * Gets the saved overlay mutation for the given document keys. Skips keys for * which there are no overlays. */ getOverlays(transaction: PersistenceTransaction, keys: DocumentKey[]): PersistencePromise; /** * Saves the given document mutation map to persistence as overlays. * All overlays will have their largest batch id set to `largestBatchId`. */ saveOverlays(transaction: PersistenceTransaction, largestBatchId: number, overlays: MutationMap): PersistencePromise; /** Removes overlays for the given document keys and batch ID. */ removeOverlaysForBatchId(transaction: PersistenceTransaction, documentKeys: DocumentKeySet, batchId: number): PersistencePromise; /** * Returns all saved overlays for the given collection. * * @param transaction - The persistence transaction to use for this operation. * @param collection - The collection path to get the overlays for. * @param sinceBatchId - The minimum batch ID to filter by (exclusive). * Only overlays that contain a change past `sinceBatchId` are returned. * @returns Mapping of each document key in the collection to its overlay. */ getOverlaysForCollection(transaction: PersistenceTransaction, collection: ResourcePath, sinceBatchId: number): PersistencePromise; /** * Returns `count` overlays with a batch ID higher than `sinceBatchId` for the * provided collection group, processed by ascending batch ID. The method * always returns all overlays for a batch even if the last batch contains * more documents than the remaining limit. * * @param transaction - The persistence transaction used for this operation. * @param collectionGroup - The collection group to get the overlays for. * @param sinceBatchId - The minimum batch ID to filter by (exclusive). * Only overlays that contain a change past `sinceBatchId` are returned. * @param count - The number of overlays to return. Can be exceeded if the last * batch contains more entries. * @return Mapping of each document key in the collection group to its overlay. */ getOverlaysForCollectionGroup(transaction: PersistenceTransaction, collectionGroup: string, sinceBatchId: number, count: number): PersistencePromise; } /** * A `DocumentReference` refers to a document location in a Firestore database * and can be used to write, read, or listen to the location. The document at * the referenced location may or may not exist. */ export declare class DocumentReference { /** * If provided, the `FirestoreDataConverter` associated with this instance. */ readonly converter: FirestoreDataConverter | null; readonly _key: DocumentKey; /** The type of this Firestore reference. */ readonly type = "document"; /** * The {@link Firestore} instance the document is in. * This is useful for performing transactions, for example. */ readonly firestore: Firestore; /** @hideconstructor */ constructor(firestore: Firestore, /** * If provided, the `FirestoreDataConverter` associated with this instance. */ converter: FirestoreDataConverter | null, _key: DocumentKey); get _path(): ResourcePath; /** * The document's identifier within its collection. */ get id(): string; /** * A string representing the path of the referenced document (relative * to the root of the database). */ get path(): string; /** * The collection this `DocumentReference` belongs to. */ get parent(): CollectionReference; /** * Applies a custom data converter to this `DocumentReference`, allowing you * to use your own custom model objects with Firestore. When you call {@link * @firebase/firestore/lite#(setDoc:1)}, {@link @firebase/firestore/lite#getDoc}, etc. with the returned `DocumentReference` * instance, the provided converter will convert between Firestore data of * type `NewDbModelType` and your custom type `NewAppModelType`. * * @param converter - Converts objects to and from Firestore. * @returns A `DocumentReference` that uses the provided converter. */ withConverter(converter: FirestoreDataConverter): DocumentReference; /** * Removes the current converter. * * @param converter - `null` removes the current converter. * @returns A `DocumentReference` that does not * use a converter. */ withConverter(converter: null): DocumentReference; } /** * DocumentSet is an immutable (copy-on-write) collection that holds documents * in order specified by the provided comparator. We always add a document key * comparator on top of what is provided to guarantee document equality based on * the key. */ declare class DocumentSet { /** * Returns an empty copy of the existing DocumentSet, using the same * comparator. */ static emptySet(oldSet: DocumentSet): DocumentSet; private comparator; private keyedMap; private sortedSet; /** The default ordering is by key if the comparator is omitted */ constructor(comp?: DocumentComparator); has(key: DocumentKey): boolean; get(key: DocumentKey): Document_2 | null; first(): Document_2 | null; last(): Document_2 | null; isEmpty(): boolean; /** * Returns the index of the provided key in the document set, or -1 if the * document key is not present in the set; */ indexOf(key: DocumentKey): number; get size(): number; /** Iterates documents in order defined by "comparator" */ forEach(cb: (doc: Document_2) => void): void; /** Inserts or updates a document with the same key */ add(doc: Document_2): DocumentSet; /** Deletes a document with a given key */ delete(key: DocumentKey): DocumentSet; isEqual(other: DocumentSet | null | undefined): boolean; toString(): string; private copy; } /** * A `DocumentSnapshot` contains data read from a document in your Firestore * database. The data can be extracted with `.data()` or `.get()` to * get a specific field. * * For a `DocumentSnapshot` that points to a non-existing document, any data * access will return 'undefined'. You can use the `exists()` method to * explicitly verify a document's existence. */ export declare class DocumentSnapshot { _firestore: Firestore; _userDataWriter: AbstractUserDataWriter; _key: DocumentKey; _document: Document_2 | null; _converter: UntypedFirestoreDataConverter | null; /** @hideconstructor protected */ constructor(_firestore: Firestore, _userDataWriter: AbstractUserDataWriter, _key: DocumentKey, _document: Document_2 | null, _converter: UntypedFirestoreDataConverter | null); /** Property of the `DocumentSnapshot` that provides the document's ID. */ get id(): string; /** * The `DocumentReference` for the document included in the `DocumentSnapshot`. */ get ref(): DocumentReference; /** * Signals whether or not the document at the snapshot's location exists. * * @returns true if the document exists. */ exists(): this is QueryDocumentSnapshot; /** * Retrieves all fields in the document as an `Object`. Returns `undefined` if * the document doesn't exist. * * @returns An `Object` containing all fields in the document or `undefined` * if the document doesn't exist. */ data(): AppModelType | undefined; /** * Retrieves the field specified by `fieldPath`. Returns `undefined` if the * document or field doesn't exist. * * @param fieldPath - The path (for example 'foo' or 'foo.bar') to a specific * field. * @returns The data at the specified field location or undefined if no such * field exists in the document. */ get(fieldPath: string | FieldPath): any; } declare type DocumentVersionMap = SortedMap; declare interface DocumentViewChange { type: ChangeType; doc: Document_2; } export { EmulatorMockTokenOptions } /** * Creates a {@link QueryEndAtConstraint} that modifies the result set to end at * the provided document (inclusive). The end position is relative to the order * of the query. The document must contain all of the fields provided in the * orderBy of the query. * * @param snapshot - The snapshot of the document to end at. * @returns A {@link QueryEndAtConstraint} to pass to `query()` */ export declare function endAt(snapshot: DocumentSnapshot): QueryEndAtConstraint; /** * Creates a {@link QueryEndAtConstraint} that modifies the result set to end at * the provided fields relative to the order of the query. The order of the field * values must match the order of the order by clauses of the query. * * @param fieldValues - The field values to end this query at, in order * of the query's order by. * @returns A {@link QueryEndAtConstraint} to pass to `query()` */ export declare function endAt(...fieldValues: unknown[]): QueryEndAtConstraint; /** * Creates a {@link QueryEndAtConstraint} that modifies the result set to end * before the provided document (exclusive). The end position is relative to the * order of the query. The document must contain all of the fields provided in * the orderBy of the query. * * @param snapshot - The snapshot of the document to end before. * @returns A {@link QueryEndAtConstraint} to pass to `query()` */ export declare function endBefore(snapshot: DocumentSnapshot): QueryEndAtConstraint; /** * Creates a {@link QueryEndAtConstraint} that modifies the result set to end * before the provided fields relative to the order of the query. The order of * the field values must match the order of the order by clauses of the query. * * @param fieldValues - The field values to end this query before, in order * of the query's order by. * @returns A {@link QueryEndAtConstraint} to pass to `query()` */ export declare function endBefore(...fieldValues: unknown[]): QueryEndAtConstraint; declare interface Entry { key: K; value: V; } /** * EventManager is responsible for mapping queries to query event emitters. * It handles "fan-out". -- Identical queries will re-use the same watch on the * backend. * * PORTING NOTE: On Web, EventManager `onListen` and `onUnlisten` need to be * assigned to SyncEngine's `listen()` and `unlisten()` API before usage. This * allows users to tree-shake the Watch logic. */ declare interface EventManager { onListen?: (query: Query_2) => Promise; onUnlisten?: (query: Query_2) => Promise; } /** * @license * Copyright 2023 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * Options that configure the SDK’s underlying network transport (WebChannel) * when long-polling is used. * * Note: This interface is "experimental" and is subject to change. * * See `FirestoreSettings.experimentalAutoDetectLongPolling`, * `FirestoreSettings.experimentalForceLongPolling`, and * `FirestoreSettings.experimentalLongPollingOptions`. */ declare interface ExperimentalLongPollingOptions { /** * The desired maximum timeout interval, in seconds, to complete a * long-polling GET response. Valid values are between 5 and 30, inclusive. * Floating point values are allowed and will be rounded to the nearest * millisecond. * * By default, when long-polling is used the "hanging GET" request sent by * the client times out after 30 seconds. To request a different timeout * from the server, set this setting with the desired timeout. * * Changing the default timeout may be useful, for example, if the buffering * proxy that necessitated enabling long-polling in the first place has a * shorter timeout for hanging GET requests, in which case setting the * long-polling timeout to a shorter value, such as 25 seconds, may fix * prematurely-closed hanging GET requests. * For example, see https://github.com/firebase/firebase-js-sdk/issues/6987. */ timeoutSeconds?: number; } declare class FieldFilter extends Filter { readonly field: FieldPath_2; readonly op: Operator; readonly value: Value; protected constructor(field: FieldPath_2, op: Operator, value: Value); /** * Creates a filter based on the provided arguments. */ static create(field: FieldPath_2, op: Operator, value: Value): FieldFilter; private static createKeyFieldInFilter; matches(doc: Document_2): boolean; protected matchesComparison(comparison: number): boolean; isInequality(): boolean; getFlattenedFilters(): readonly FieldFilter[]; getFilters(): Filter[]; } declare type FieldFilterOp = 'OPERATOR_UNSPECIFIED' | 'LESS_THAN' | 'LESS_THAN_OR_EQUAL' | 'GREATER_THAN' | 'GREATER_THAN_OR_EQUAL' | 'EQUAL' | 'NOT_EQUAL' | 'ARRAY_CONTAINS' | 'IN' | 'ARRAY_CONTAINS_ANY' | 'NOT_IN'; /** * An index definition for field indexes in Firestore. * * Every index is associated with a collection. The definition contains a list * of fields and their index kind (which can be `ASCENDING`, `DESCENDING` or * `CONTAINS` for ArrayContains/ArrayContainsAny queries). * * Unlike the backend, the SDK does not differentiate between collection or * collection group-scoped indices. Every index can be used for both single * collection and collection group queries. */ declare class FieldIndex { /** * The index ID. Returns -1 if the index ID is not available (e.g. the index * has not yet been persisted). */ readonly indexId: number; /** The collection ID this index applies to. */ readonly collectionGroup: string; /** The field segments for this index. */ readonly fields: IndexSegment[]; /** Shows how up-to-date the index is for the current user. */ readonly indexState: IndexState; /** An ID for an index that has not yet been added to persistence. */ static UNKNOWN_ID: number; constructor( /** * The index ID. Returns -1 if the index ID is not available (e.g. the index * has not yet been persisted). */ indexId: number, /** The collection ID this index applies to. */ collectionGroup: string, /** The field segments for this index. */ fields: IndexSegment[], /** Shows how up-to-date the index is for the current user. */ indexState: IndexState); } /** * Provides a set of fields that can be used to partially patch a document. * FieldMask is used in conjunction with ObjectValue. * Examples: * foo - Overwrites foo entirely with the provided value. If foo is not * present in the companion ObjectValue, the field is deleted. * foo.bar - Overwrites only the field bar of the object foo. * If foo is not an object, foo is replaced with an object * containing foo */ declare class FieldMask { readonly fields: FieldPath_2[]; constructor(fields: FieldPath_2[]); static empty(): FieldMask; /** * Returns a new FieldMask object that is the result of adding all the given * fields paths to this field mask. */ unionWith(extraFields: FieldPath_2[]): FieldMask; /** * Verifies that `fieldPath` is included by at least one field in this field * mask. * * This is an O(n) operation, where `n` is the size of the field mask. */ covers(fieldPath: FieldPath_2): boolean; isEqual(other: FieldMask): boolean; } /** * A `FieldPath` refers to a field in a document. The path may consist of a * single field name (referring to a top-level field in the document), or a * list of field names (referring to a nested field in the document). * * Create a `FieldPath` by providing field names. If more than one field * name is provided, the path will point to a nested field in a document. */ export declare class FieldPath { /** Internal representation of a Firestore field path. */ readonly _internalPath: FieldPath_2; /** * Creates a `FieldPath` from the provided field names. If more than one field * name is provided, the path will point to a nested field in a document. * * @param fieldNames - A list of field names. */ constructor(...fieldNames: string[]); /** * Returns true if this `FieldPath` is equal to the provided one. * * @param other - The `FieldPath` to compare against. * @returns true if this `FieldPath` is equal to the provided one. */ isEqual(other: FieldPath): boolean; } /* Excluded from this release type: FieldPath_2 */ /** A field path and the TransformOperation to perform upon it. */ declare class FieldTransform { readonly field: FieldPath_2; readonly transform: TransformOperation; constructor(field: FieldPath_2, transform: TransformOperation); } declare type FieldTransformSetToServerValue = 'SERVER_VALUE_UNSPECIFIED' | 'REQUEST_TIME'; /** * Sentinel values that can be used when writing document fields with `set()` * or `update()`. */ export declare abstract class FieldValue { _methodName: string; /** * @param _methodName - The public API endpoint that returns this class. * @hideconstructor */ constructor(_methodName: string); /** Compares `FieldValue`s for equality. */ abstract isEqual(other: FieldValue): boolean; abstract _toFieldTransform(context: ParseContext): FieldTransform | null; } declare abstract class Filter { abstract matches(doc: Document_2): boolean; abstract getFlattenedFilters(): readonly FieldFilter[]; abstract getFilters(): Filter[]; } /* Excluded from this release type: _FirebaseService */ /** * The Cloud Firestore service interface. * * Do not call this constructor directly. Instead, use {@link (getFirestore:1)}. */ export declare class Firestore implements FirestoreService { _authCredentials: CredentialsProvider; _appCheckCredentials: CredentialsProvider; readonly _databaseId: DatabaseId; readonly _app?: FirebaseApp | undefined; /** * Whether it's a Firestore or Firestore Lite instance. */ type: 'firestore-lite' | 'firestore'; readonly _persistenceKey: string; private _settings; private _settingsFrozen; private _terminateTask?; /** @hideconstructor */ constructor(_authCredentials: CredentialsProvider, _appCheckCredentials: CredentialsProvider, _databaseId: DatabaseId, _app?: FirebaseApp | undefined); /** * The {@link @firebase/app#FirebaseApp} associated with this `Firestore` service * instance. */ get app(): FirebaseApp; get _initialized(): boolean; get _terminated(): boolean; _setSettings(settings: PrivateSettings): void; _getSettings(): FirestoreSettingsImpl; _freezeSettings(): FirestoreSettingsImpl; _delete(): Promise; /** Returns a JSON-serializable representation of this `Firestore` instance. */ toJSON(): object; /** * Terminates all components used by this client. Subclasses can override * this method to clean up their own dependencies, but must also call this * method. * * Only ever called once. */ protected _terminate(): Promise; } /** * Converter used by `withConverter()` to transform user objects of type * `AppModelType` into Firestore data of type `DbModelType`. * * Using the converter allows you to specify generic type arguments when * storing and retrieving objects from Firestore. * * In this context, an "AppModel" is a class that is used in an application to * package together related information and functionality. Such a class could, * for example, have properties with complex, nested data types, properties used * for memoization, properties of types not supported by Firestore (such as * `symbol` and `bigint`), and helper functions that perform compound * operations. Such classes are not suitable and/or possible to store into a * Firestore database. Instead, instances of such classes need to be converted * to "plain old JavaScript objects" (POJOs) with exclusively primitive * properties, potentially nested inside other POJOs or arrays of POJOs. In this * context, this type is referred to as the "DbModel" and would be an object * suitable for persisting into Firestore. For convenience, applications can * implement `FirestoreDataConverter` and register the converter with Firestore * objects, such as `DocumentReference` or `Query`, to automatically convert * `AppModel` to `DbModel` when storing into Firestore, and convert `DbModel` * to `AppModel` when retrieving from Firestore. * * @example * * Simple Example * * ```typescript * const numberConverter = { * toFirestore(value: WithFieldValue) { * return { value }; * }, * fromFirestore(snapshot: QueryDocumentSnapshot, options: SnapshotOptions) { * return snapshot.data(options).value as number; * } * }; * * async function simpleDemo(db: Firestore): Promise { * const documentRef = doc(db, 'values/value123').withConverter(numberConverter); * * // converters are used with `setDoc`, `addDoc`, and `getDoc` * await setDoc(documentRef, 42); * const snapshot1 = await getDoc(documentRef); * assertEqual(snapshot1.data(), 42); * * // converters are not used when writing data with `updateDoc` * await updateDoc(documentRef, { value: 999 }); * const snapshot2 = await getDoc(documentRef); * assertEqual(snapshot2.data(), 999); * } * ``` * * Advanced Example * * ```typescript * // The Post class is a model that is used by our application. * // This class may have properties and methods that are specific * // to our application execution, which do not need to be persisted * // to Firestore. * class Post { * constructor( * readonly title: string, * readonly author: string, * readonly lastUpdatedMillis: number * ) {} * toString(): string { * return `${this.title} by ${this.author}`; * } * } * * // The PostDbModel represents how we want our posts to be stored * // in Firestore. This DbModel has different properties (`ttl`, * // `aut`, and `lut`) from the Post class we use in our application. * interface PostDbModel { * ttl: string; * aut: { firstName: string; lastName: string }; * lut: Timestamp; * } * * // The `PostConverter` implements `FirestoreDataConverter` and specifies * // how the Firestore SDK can convert `Post` objects to `PostDbModel` * // objects and vice versa. * class PostConverter implements FirestoreDataConverter { * toFirestore(post: WithFieldValue): WithFieldValue { * return { * ttl: post.title, * aut: this._autFromAuthor(post.author), * lut: this._lutFromLastUpdatedMillis(post.lastUpdatedMillis) * }; * } * * fromFirestore(snapshot: QueryDocumentSnapshot, options: SnapshotOptions): Post { * const data = snapshot.data(options) as PostDbModel; * const author = `${data.aut.firstName} ${data.aut.lastName}`; * return new Post(data.ttl, author, data.lut.toMillis()); * } * * _autFromAuthor( * author: string | FieldValue * ): { firstName: string; lastName: string } | FieldValue { * if (typeof author !== 'string') { * // `author` is a FieldValue, so just return it. * return author; * } * const [firstName, lastName] = author.split(' '); * return {firstName, lastName}; * } * * _lutFromLastUpdatedMillis( * lastUpdatedMillis: number | FieldValue * ): Timestamp | FieldValue { * if (typeof lastUpdatedMillis !== 'number') { * // `lastUpdatedMillis` must be a FieldValue, so just return it. * return lastUpdatedMillis; * } * return Timestamp.fromMillis(lastUpdatedMillis); * } * } * * async function advancedDemo(db: Firestore): Promise { * // Create a `DocumentReference` with a `FirestoreDataConverter`. * const documentRef = doc(db, 'posts/post123').withConverter(new PostConverter()); * * // The `data` argument specified to `setDoc()` is type checked by the * // TypeScript compiler to be compatible with `Post`. Since the `data` * // argument is typed as `WithFieldValue` rather than just `Post`, * // this allows properties of the `data` argument to also be special * // Firestore values that perform server-side mutations, such as * // `arrayRemove()`, `deleteField()`, and `serverTimestamp()`. * await setDoc(documentRef, { * title: 'My Life', * author: 'Foo Bar', * lastUpdatedMillis: serverTimestamp() * }); * * // The TypeScript compiler will fail to compile if the `data` argument to * // `setDoc()` is _not_ compatible with `WithFieldValue`. This * // type checking prevents the caller from specifying objects with incorrect * // properties or property values. * // @ts-expect-error "Argument of type { ttl: string; } is not assignable * // to parameter of type WithFieldValue" * await setDoc(documentRef, { ttl: 'The Title' }); * * // When retrieving a document with `getDoc()` the `DocumentSnapshot` * // object's `data()` method returns a `Post`, rather than a generic object, * // which would have been returned if the `DocumentReference` did _not_ have a * // `FirestoreDataConverter` attached to it. * const snapshot1: DocumentSnapshot = await getDoc(documentRef); * const post1: Post = snapshot1.data()!; * if (post1) { * assertEqual(post1.title, 'My Life'); * assertEqual(post1.author, 'Foo Bar'); * } * * // The `data` argument specified to `updateDoc()` is type checked by the * // TypeScript compiler to be compatible with `PostDbModel`. Note that * // unlike `setDoc()`, whose `data` argument must be compatible with `Post`, * // the `data` argument to `updateDoc()` must be compatible with * // `PostDbModel`. Similar to `setDoc()`, since the `data` argument is typed * // as `WithFieldValue` rather than just `PostDbModel`, this * // allows properties of the `data` argument to also be those special * // Firestore values, like `arrayRemove()`, `deleteField()`, and * // `serverTimestamp()`. * await updateDoc(documentRef, { * 'aut.firstName': 'NewFirstName', * lut: serverTimestamp() * }); * * // The TypeScript compiler will fail to compile if the `data` argument to * // `updateDoc()` is _not_ compatible with `WithFieldValue`. * // This type checking prevents the caller from specifying objects with * // incorrect properties or property values. * // @ts-expect-error "Argument of type { title: string; } is not assignable * // to parameter of type WithFieldValue" * await updateDoc(documentRef, { title: 'New Title' }); * const snapshot2: DocumentSnapshot = await getDoc(documentRef); * const post2: Post = snapshot2.data()!; * if (post2) { * assertEqual(post2.title, 'My Life'); * assertEqual(post2.author, 'NewFirstName Bar'); * } * } * ``` */ export declare interface FirestoreDataConverter { /** * Called by the Firestore SDK to convert a custom model object of type * `AppModelType` into a plain Javascript object (suitable for writing * directly to the Firestore database) of type `DbModelType`. Used with * {@link @firebase/firestore/lite#(setDoc:1)}, * {@link @firebase/firestore/lite#(WriteBatch.set:1)} and * {@link @firebase/firestore/lite#(Transaction.set:1)}. * * The `WithFieldValue` type extends `T` to also allow FieldValues such as * {@link (deleteField:1)} to be used as property values. */ toFirestore(modelObject: WithFieldValue): WithFieldValue; /** * Called by the Firestore SDK to convert a custom model object of type * `AppModelType` into a plain Javascript object (suitable for writing * directly to the Firestore database) of type `DbModelType`. Used with * {@link @firebase/firestore/lite#(setDoc:1)}, * {@link @firebase/firestore/lite#(WriteBatch.set:1)} and * {@link @firebase/firestore/lite#(Transaction.set:1)} with `merge:true` * or `mergeFields`. * * The `PartialWithFieldValue` type extends `Partial` to allow * FieldValues such as {@link (arrayUnion:1)} to be used as property values. * It also supports nested `Partial` by allowing nested fields to be * omitted. */ toFirestore(modelObject: PartialWithFieldValue, options: SetOptions): PartialWithFieldValue; /** * Called by the Firestore SDK to convert Firestore data into an object of * type `AppModelType`. You can access your data by calling: * `snapshot.data()`. * * * Generally, the data returned from `snapshot.data()` can be cast to * `DbModelType`; however, this is not guaranteed because Firestore does not * enforce a schema on the database. For example, writes from a previous * version of the application or writes from another client that did not use a * type converter could have written data with different properties and/or * property types. The implementation will need to choose whether to * gracefully recover from non-conforming data or throw an error. * * @param snapshot - A `QueryDocumentSnapshot` containing your data and * metadata. */ fromFirestore(snapshot: QueryDocumentSnapshot): AppModelType; } /** An error returned by a Firestore operation. */ export declare class FirestoreError extends FirebaseError { /** * The backend error code associated with this error. */ readonly code: FirestoreErrorCode; /** * A custom error description. */ readonly message: string; /** The stack of the error. */ readonly stack?: string; /** @hideconstructor */ constructor( /** * The backend error code associated with this error. */ code: FirestoreErrorCode, /** * A custom error description. */ message: string); } /** * The set of Firestore status codes. The codes are the same at the ones * exposed by gRPC here: * https://github.com/grpc/grpc/blob/master/doc/statuscodes.md * * Possible values: * - 'cancelled': The operation was cancelled (typically by the caller). * - 'unknown': Unknown error or an error from a different error domain. * - 'invalid-argument': Client specified an invalid argument. Note that this * differs from 'failed-precondition'. 'invalid-argument' indicates * arguments that are problematic regardless of the state of the system * (e.g. an invalid field name). * - 'deadline-exceeded': Deadline expired before operation could complete. * For operations that change the state of the system, this error may be * returned even if the operation has completed successfully. For example, * a successful response from a server could have been delayed long enough * for the deadline to expire. * - 'not-found': Some requested document was not found. * - 'already-exists': Some document that we attempted to create already * exists. * - 'permission-denied': The caller does not have permission to execute the * specified operation. * - 'resource-exhausted': Some resource has been exhausted, perhaps a * per-user quota, or perhaps the entire file system is out of space. * - 'failed-precondition': Operation was rejected because the system is not * in a state required for the operation's execution. * - 'aborted': The operation was aborted, typically due to a concurrency * issue like transaction aborts, etc. * - 'out-of-range': Operation was attempted past the valid range. * - 'unimplemented': Operation is not implemented or not supported/enabled. * - 'internal': Internal errors. Means some invariants expected by * underlying system has been broken. If you see one of these errors, * something is very broken. * - 'unavailable': The service is currently unavailable. This is most likely * a transient condition and may be corrected by retrying with a backoff. * - 'data-loss': Unrecoverable data loss or corruption. * - 'unauthenticated': The request does not have valid authentication * credentials for the operation. */ export declare type FirestoreErrorCode = 'cancelled' | 'unknown' | 'invalid-argument' | 'deadline-exceeded' | 'not-found' | 'already-exists' | 'permission-denied' | 'resource-exhausted' | 'failed-precondition' | 'aborted' | 'out-of-range' | 'unimplemented' | 'internal' | 'unavailable' | 'data-loss' | 'unauthenticated'; /** * Union type from all supported SDK cache layer. */ declare type FirestoreLocalCache = MemoryLocalCache | PersistentLocalCache; /** * An interface implemented by FirebaseFirestore that provides compatibility * with the usage in this file. * * This interface mainly exists to remove a cyclic dependency. */ declare interface FirestoreService extends _FirebaseService { _authCredentials: CredentialsProvider; _appCheckCredentials: CredentialsProvider; _persistenceKey: string; _databaseId: DatabaseId; _terminated: boolean; _freezeSettings(): FirestoreSettingsImpl; } /** * A concrete type describing all the values that can be applied via a * user-supplied `FirestoreSettings` object. This is a separate type so that * defaults can be supplied and the value can be checked for equality. */ declare class FirestoreSettingsImpl { /** The hostname to connect to. */ readonly host: string; /** Whether to use SSL when connecting. */ readonly ssl: boolean; readonly cacheSizeBytes: number; readonly experimentalForceLongPolling: boolean; readonly experimentalAutoDetectLongPolling: boolean; readonly experimentalLongPollingOptions: ExperimentalLongPollingOptions; readonly ignoreUndefinedProperties: boolean; readonly useFetchStreams: boolean; readonly localCache?: FirestoreLocalCache; credentials?: any; constructor(settings: PrivateSettings); isEqual(other: FirestoreSettingsImpl): boolean; } declare namespace firestoreV1ApiClientInterfaces { interface ArrayValue { values?: Value[]; } interface BatchGetDocumentsRequest { database?: string; documents?: string[]; mask?: DocumentMask; transaction?: string; newTransaction?: TransactionOptions; readTime?: string; } interface BatchGetDocumentsResponse { found?: Document; missing?: string; transaction?: string; readTime?: string; } interface BeginTransactionRequest { options?: TransactionOptions; } interface BeginTransactionResponse { transaction?: string; } interface CollectionSelector { collectionId?: string; allDescendants?: boolean; } interface CommitRequest { database?: string; writes?: Write[]; transaction?: string; } interface CommitResponse { writeResults?: WriteResult[]; commitTime?: string; } interface CompositeFilter { op?: CompositeFilterOp; filters?: Filter[]; } interface Cursor { values?: Value[]; before?: boolean; } interface Document { name?: string; fields?: ApiClientObjectMap; createTime?: Timestamp_2; updateTime?: Timestamp_2; } interface DocumentChange { document?: Document; targetIds?: number[]; removedTargetIds?: number[]; } interface DocumentDelete { document?: string; removedTargetIds?: number[]; readTime?: Timestamp_2; } interface DocumentMask { fieldPaths?: string[]; } interface DocumentRemove { document?: string; removedTargetIds?: number[]; readTime?: string; } interface DocumentTransform { document?: string; fieldTransforms?: FieldTransform[]; } interface DocumentsTarget { documents?: string[]; } interface Empty { } interface ExistenceFilter { targetId?: number; count?: number; unchangedNames?: BloomFilter; } interface BloomFilter { bits?: BitSequence; hashCount?: number; } interface BitSequence { bitmap?: string | Uint8Array; padding?: number; } interface FieldFilter { field?: FieldReference; op?: FieldFilterOp; value?: Value; } interface FieldReference { fieldPath?: string; } interface FieldTransform { fieldPath?: string; setToServerValue?: FieldTransformSetToServerValue; appendMissingElements?: ArrayValue; removeAllFromArray?: ArrayValue; increment?: Value; } interface Filter { compositeFilter?: CompositeFilter; fieldFilter?: FieldFilter; unaryFilter?: UnaryFilter; } interface Index { name?: string; collectionId?: string; fields?: IndexField[]; state?: IndexState_2; } interface IndexField { fieldPath?: string; mode?: IndexFieldMode; } interface LatLng { latitude?: number; longitude?: number; } interface ListCollectionIdsRequest { pageSize?: number; pageToken?: string; } interface ListCollectionIdsResponse { collectionIds?: string[]; nextPageToken?: string; } interface ListDocumentsResponse { documents?: Document[]; nextPageToken?: string; } interface ListIndexesResponse { indexes?: Index[]; nextPageToken?: string; } interface ListenRequest { addTarget?: Target; removeTarget?: number; labels?: ApiClientObjectMap; } interface ListenResponse { targetChange?: TargetChange; documentChange?: DocumentChange; documentDelete?: DocumentDelete; documentRemove?: DocumentRemove; filter?: ExistenceFilter; } interface MapValue { fields?: ApiClientObjectMap; } interface Operation { name?: string; metadata?: ApiClientObjectMap; done?: boolean; error?: Status; response?: ApiClientObjectMap; } interface Order { field?: FieldReference; direction?: OrderDirection; } interface Precondition { exists?: boolean; updateTime?: Timestamp_2; } interface Projection { fields?: FieldReference[]; } interface QueryTarget { parent?: string; structuredQuery?: StructuredQuery; } interface ReadOnly { readTime?: string; } interface ReadWrite { retryTransaction?: string; } interface RollbackRequest { transaction?: string; } interface RunQueryRequest { parent?: string; structuredQuery?: StructuredQuery; transaction?: string; newTransaction?: TransactionOptions; readTime?: string; } interface RunQueryResponse { transaction?: string; document?: Document; readTime?: string; skippedResults?: number; } interface RunAggregationQueryRequest { parent?: string; structuredAggregationQuery?: StructuredAggregationQuery; transaction?: string; newTransaction?: TransactionOptions; readTime?: string; } interface RunAggregationQueryResponse { result?: AggregationResult; transaction?: string; readTime?: string; } interface AggregationResult { aggregateFields?: ApiClientObjectMap; } interface StructuredAggregationQuery { structuredQuery?: StructuredQuery; aggregations?: Aggregation[]; } interface Aggregation { count?: Count; sum?: Sum; avg?: Avg; alias?: string; } interface Count { upTo?: number; } interface Sum { field?: FieldReference; } interface Avg { field?: FieldReference; } interface Status { code?: number; message?: string; details?: Array>; } interface StructuredQuery { select?: Projection; from?: CollectionSelector[]; where?: Filter; orderBy?: Order[]; startAt?: Cursor; endAt?: Cursor; offset?: number; limit?: number | { value: number; }; } interface Target { query?: QueryTarget; documents?: DocumentsTarget; resumeToken?: string | Uint8Array; readTime?: Timestamp_2; targetId?: number; once?: boolean; expectedCount?: number | { value: number; }; } interface TargetChange { targetChangeType?: TargetChangeTargetChangeType; targetIds?: number[]; cause?: Status; resumeToken?: string | Uint8Array; readTime?: Timestamp_2; } interface TransactionOptions { readOnly?: ReadOnly; readWrite?: ReadWrite; } interface UnaryFilter { op?: UnaryFilterOp; field?: FieldReference; } interface Value { nullValue?: ValueNullValue; booleanValue?: boolean; integerValue?: string | number; doubleValue?: string | number; timestampValue?: Timestamp_2; stringValue?: string; bytesValue?: string | Uint8Array; referenceValue?: string; geoPointValue?: LatLng; arrayValue?: ArrayValue; mapValue?: MapValue; } interface Write { update?: Document; delete?: string; verify?: string; transform?: DocumentTransform; updateMask?: DocumentMask; updateTransforms?: FieldTransform[]; currentDocument?: Precondition; } interface WriteRequest { streamId?: string; writes?: Write[]; streamToken?: string | Uint8Array; labels?: ApiClientObjectMap; } interface WriteResponse { streamId?: string; streamToken?: string | Uint8Array; writeResults?: WriteResult[]; commitTime?: Timestamp_2; } interface WriteResult { updateTime?: Timestamp_2; transformResults?: Value[]; } } /* Excluded from this release type: FirstPartyCredentialsSettings */ /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ declare type FulfilledHandler = ((result: T) => R | PersistencePromise) | null; /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * An immutable object representing a geographic location in Firestore. The * location is represented as latitude/longitude pair. * * Latitude values are in the range of [-90, 90]. * Longitude values are in the range of [-180, 180]. */ export declare class GeoPoint { private _lat; private _long; /** * Creates a new immutable `GeoPoint` object with the provided latitude and * longitude values. * @param latitude - The latitude as number between -90 and 90. * @param longitude - The longitude as number between -180 and 180. */ constructor(latitude: number, longitude: number); /** * The latitude of this `GeoPoint` instance. */ get latitude(): number; /** * The longitude of this `GeoPoint` instance. */ get longitude(): number; /** * Returns true if this `GeoPoint` is equal to the provided one. * * @param other - The `GeoPoint` to compare against. * @returns true if this `GeoPoint` is equal to the provided one. */ isEqual(other: GeoPoint): boolean; /** Returns a JSON-serializable representation of this GeoPoint. */ toJSON(): { latitude: number; longitude: number; }; /** * Actually private to JS consumers of our API, so this function is prefixed * with an underscore. */ _compareTo(other: GeoPoint): number; } /** * Calculates the specified aggregations over the documents in the result * set of the given query without actually downloading the documents. * * Using this function to perform aggregations is efficient because only the * final aggregation values, not the documents' data, are downloaded. This * function can perform aggregations of the documents in cases where the result * set is prohibitively large to download entirely (thousands of documents). * * @param query The query whose result set is aggregated over. * @param aggregateSpec An `AggregateSpec` object that specifies the aggregates * to perform over the result set. The AggregateSpec specifies aliases for each * aggregate, which can be used to retrieve the aggregate result. * @example * ```typescript * const aggregateSnapshot = await getAggregate(query, { * countOfDocs: count(), * totalHours: sum('hours'), * averageScore: average('score') * }); * * const countOfDocs: number = aggregateSnapshot.data().countOfDocs; * const totalHours: number = aggregateSnapshot.data().totalHours; * const averageScore: number | null = aggregateSnapshot.data().averageScore; * ``` */ export declare function getAggregate(query: Query, aggregateSpec: AggregateSpecType): Promise>; /** * Calculates the number of documents in the result set of the given query * without actually downloading the documents. * * Using this function to count the documents is efficient because only the * final count, not the documents' data, is downloaded. This function can * count the documents in cases where the result set is prohibitively large to * download entirely (thousands of documents). * * @param query The query whose result set size is calculated. * @returns A Promise that will be resolved with the count; the count can be * retrieved from `snapshot.data().count`, where `snapshot` is the * `AggregateQuerySnapshot` to which the returned Promise resolves. */ export declare function getCount(query: Query): Promise; }, AppModelType, DbModelType>>; /** * Reads the document referred to by the specified document reference. * * All documents are directly fetched from the server, even if the document was * previously read or modified. Recent modifications are only reflected in the * retrieved `DocumentSnapshot` if they have already been applied by the * backend. If the client is offline, the read fails. If you like to use * caching or see local modifications, please use the full Firestore SDK. * * @param reference - The reference of the document to fetch. * @returns A Promise resolved with a `DocumentSnapshot` containing the current * document contents. */ export declare function getDoc(reference: DocumentReference): Promise>; /** * Executes the query and returns the results as a {@link QuerySnapshot}. * * All queries are executed directly by the server, even if the the query was * previously executed. Recent modifications are only reflected in the retrieved * results if they have already been applied by the backend. If the client is * offline, the operation fails. To see previously cached result and local * modifications, use the full Firestore SDK. * * @param query - The `Query` to execute. * @returns A Promise that will be resolved with the results of the query. */ export declare function getDocs(query: Query): Promise>; /** * Returns the existing default {@link Firestore} instance that is associated with the * default {@link @firebase/app#FirebaseApp}. If no instance exists, initializes a new * instance with default settings. * * @returns The {@link Firestore} instance of the provided app. */ export declare function getFirestore(): Firestore; /** * Returns the existing default {@link Firestore} instance that is associated with the * provided {@link @firebase/app#FirebaseApp}. If no instance exists, initializes a new * instance with default settings. * * @param app - The {@link @firebase/app#FirebaseApp} instance that the returned {@link Firestore} * instance is associated with. * @returns The {@link Firestore} instance of the provided app. */ export declare function getFirestore(app: FirebaseApp): Firestore; /** * Returns the existing {@link Firestore} instance that is associated with the * default {@link @firebase/app#FirebaseApp}. If no instance exists, initializes a new * instance with default settings. * * @param databaseId - The name of the database. * @returns The {@link Firestore} instance of the provided app. * @beta */ export declare function getFirestore(databaseId: string): Firestore; /** * Returns the existing {@link Firestore} instance that is associated with the * provided {@link @firebase/app#FirebaseApp}. If no instance exists, initializes a new * instance with default settings. * * @param app - The {@link @firebase/app#FirebaseApp} instance that the returned {@link Firestore} * instance is associated with. * @param databaseId - The name of the database. * @returns The {@link Firestore} instance of the provided app. * @beta */ export declare function getFirestore(app: FirebaseApp, databaseId: string): Firestore; /** * Returns a special value that can be used with {@link @firebase/firestore/lite#(setDoc:1)} or {@link * @firebase/firestore/lite#(updateDoc:1)} that tells the server to increment the field's current value by * the given value. * * If either the operand or the current field value uses floating point * precision, all arithmetic follows IEEE 754 semantics. If both values are * integers, values outside of JavaScript's safe number range * (`Number.MIN_SAFE_INTEGER` to `Number.MAX_SAFE_INTEGER`) are also subject to * precision loss. Furthermore, once processed by the Firestore backend, all * integer operations are capped between -2^63 and 2^63-1. * * If the current field value is not of type `number`, or if the field does not * yet exist, the transformation sets the field to the given value. * * @param n - The value to increment by. * @returns The `FieldValue` sentinel for use in a call to `setDoc()` or * `updateDoc()` */ export declare function increment(n: number): FieldValue; declare type IndexFieldMode = 'MODE_UNSPECIFIED' | 'ASCENDING' | 'DESCENDING'; /** The type of the index, e.g. for which type of query it can be used. */ declare const enum IndexKind { /** * Ordered index. Can be used for <, <=, ==, >=, >, !=, IN and NOT IN queries. */ ASCENDING = 0, /** * Ordered index. Can be used for <, <=, ==, >=, >, !=, IN and NOT IN queries. */ DESCENDING = 1, /** Contains index. Can be used for ArrayContains and ArrayContainsAny. */ CONTAINS = 2 } /** * Represents a set of indexes that are used to execute queries efficiently. * * Currently the only index is a [collection id] => [parent path] index, used * to execute Collection Group queries. */ declare interface IndexManager { /** * Creates an index entry mapping the collectionId (last segment of the path) * to the parent path (either the containing document location or the empty * path for root-level collections). Index entries can be retrieved via * getCollectionParents(). * * NOTE: Currently we don't remove index entries. If this ends up being an * issue we can devise some sort of GC strategy. */ addToCollectionParentIndex(transaction: PersistenceTransaction, collectionPath: ResourcePath): PersistencePromise; /** * Retrieves all parent locations containing the given collectionId, as a * list of paths (each path being either a document location or the empty * path for a root-level collection). */ getCollectionParents(transaction: PersistenceTransaction, collectionId: string): PersistencePromise; /** * Adds a field path index. * * Values for this index are persisted via the index backfill, which runs * asynchronously in the background. Once the first values are written, * an index can be used to serve partial results for any matching queries. * Any unindexed portion of the database will continue to be served via * collection scons. */ addFieldIndex(transaction: PersistenceTransaction, index: FieldIndex): PersistencePromise; /** Removes the given field index and deletes all index values. */ deleteFieldIndex(transaction: PersistenceTransaction, index: FieldIndex): PersistencePromise; /** Removes all field indexes and deletes all index values. */ deleteAllFieldIndexes(transaction: PersistenceTransaction): PersistencePromise; /** Creates a full matched field index which serves the given target. */ createTargetIndexes(transaction: PersistenceTransaction, target: Target): PersistencePromise; /** * Returns a list of field indexes that correspond to the specified collection * group. * * @param collectionGroup The collection group to get matching field indexes * for. * @return A collection of field indexes for the specified collection group. */ getFieldIndexes(transaction: PersistenceTransaction, collectionGroup: string): PersistencePromise; /** Returns all configured field indexes. */ getFieldIndexes(transaction: PersistenceTransaction): PersistencePromise; /** * Returns the type of index (if any) that can be used to serve the given * target. */ getIndexType(transaction: PersistenceTransaction, target: Target): PersistencePromise; /** * Returns the documents that match the given target based on the provided * index or `null` if the target does not have a matching index. */ getDocumentsMatchingTarget(transaction: PersistenceTransaction, target: Target): PersistencePromise; /** * Returns the next collection group to update. Returns `null` if no group * exists. */ getNextCollectionGroupToUpdate(transaction: PersistenceTransaction): PersistencePromise; /** * Sets the collection group's latest read time. * * This method updates the index offset for all field indices for the * collection group and increments their sequence number. Subsequent calls to * `getNextCollectionGroupToUpdate()` will return a different collection group * (unless only one collection group is configured). */ updateCollectionGroup(transaction: PersistenceTransaction, collectionGroup: string, offset: IndexOffset): PersistencePromise; /** Updates the index entries for the provided documents. */ updateIndexEntries(transaction: PersistenceTransaction, documents: DocumentMap): PersistencePromise; /** * Iterates over all field indexes that are used to serve the given target, * and returns the minimum offset of them all. */ getMinOffset(transaction: PersistenceTransaction, target: Target): PersistencePromise; /** Returns the minimum offset for the given collection group. */ getMinOffsetFromCollectionGroup(transaction: PersistenceTransaction, collectionGroup: string): PersistencePromise; } /** * Stores the latest read time, document and batch ID that were processed for an * index. */ declare class IndexOffset { /** * The latest read time version that has been indexed by Firestore for this * field index. */ readonly readTime: SnapshotVersion; /** * The key of the last document that was indexed for this query. Use * `DocumentKey.empty()` if no document has been indexed. */ readonly documentKey: DocumentKey; readonly largestBatchId: number; constructor( /** * The latest read time version that has been indexed by Firestore for this * field index. */ readTime: SnapshotVersion, /** * The key of the last document that was indexed for this query. Use * `DocumentKey.empty()` if no document has been indexed. */ documentKey: DocumentKey, largestBatchId: number); /** Returns an offset that sorts before all regular offsets. */ static min(): IndexOffset; /** Returns an offset that sorts after all regular offsets. */ static max(): IndexOffset; } /** An index component consisting of field path and index type. */ declare class IndexSegment { /** The field path of the component. */ readonly fieldPath: FieldPath_2; /** The fields sorting order. */ readonly kind: IndexKind; constructor( /** The field path of the component. */ fieldPath: FieldPath_2, /** The fields sorting order. */ kind: IndexKind); } /** * Stores the "high water mark" that indicates how updated the Index is for the * current user. */ declare class IndexState { /** * Indicates when the index was last updated (relative to other indexes). */ readonly sequenceNumber: number; /** The the latest indexed read time, document and batch id. */ readonly offset: IndexOffset; constructor( /** * Indicates when the index was last updated (relative to other indexes). */ sequenceNumber: number, /** The the latest indexed read time, document and batch id. */ offset: IndexOffset); /** The state of an index that has not yet been backfilled. */ static empty(): IndexState; } declare type IndexState_2 = 'STATE_UNSPECIFIED' | 'CREATING' | 'READY' | 'ERROR'; /** Represents the index state as it relates to a particular target. */ declare const enum IndexType { /** Indicates that no index could be found for serving the target. */ NONE = 0, /** * Indicates that only a "partial index" could be found for serving the * target. A partial index is one which does not have a segment for every * filter/orderBy in the target. */ PARTIAL = 1, /** * Indicates that a "full index" could be found for serving the target. A full * index is one which has a segment for every filter/orderBy in the target. */ FULL = 2 } /** * Initializes a new instance of Cloud Firestore with the provided settings. * Can only be called before any other functions, including * {@link (getFirestore:1)}. If the custom settings are empty, this function is * equivalent to calling {@link (getFirestore:1)}. * * @param app - The {@link @firebase/app#FirebaseApp} with which the `Firestore` instance will * be associated. * @param settings - A settings object to configure the `Firestore` instance. * @returns A newly initialized `Firestore` instance. */ export declare function initializeFirestore(app: FirebaseApp, settings: Settings): Firestore; /** * Initializes a new instance of Cloud Firestore with the provided settings. * Can only be called before any other functions, including * {@link (getFirestore:1)}. If the custom settings are empty, this function is * equivalent to calling {@link (getFirestore:1)}. * * @param app - The {@link @firebase/app#FirebaseApp} with which the `Firestore` instance will * be associated. * @param settings - A settings object to configure the `Firestore` instance. * @param databaseId - The name of the database. * @returns A newly initialized `Firestore` instance. * @beta */ export declare function initializeFirestore(app: FirebaseApp, settings: Settings, databaseId?: string): Firestore; /** * This class generates JsonObject values for the Datastore API suitable for * sending to either GRPC stub methods or via the JSON/HTTP REST API. * * The serializer supports both Protobuf.js and Proto3 JSON formats. By * setting `useProto3Json` to true, the serializer will use the Proto3 JSON * format. * * For a description of the Proto3 JSON format check * https://developers.google.com/protocol-buffers/docs/proto3#json * * TODO(klimt): We can remove the databaseId argument if we keep the full * resource name in documents. */ declare class JsonProtoSerializer implements Serializer { readonly databaseId: DatabaseId; readonly useProto3Json: boolean; constructor(databaseId: DatabaseId, useProto3Json: boolean); } /** * Creates a {@link QueryLimitConstraint} that only returns the first matching * documents. * * @param limit - The maximum number of items to return. * @returns The created {@link QueryLimitConstraint}. */ export declare function limit(limit: number): QueryLimitConstraint; /** * Creates a {@link QueryLimitConstraint} that only returns the last matching * documents. * * You must specify at least one `orderBy` clause for `limitToLast` queries, * otherwise an exception will be thrown during execution. * * @param limit - The maximum number of items to return. * @returns The created {@link QueryLimitConstraint}. */ export declare function limitToLast(limit: number): QueryLimitConstraint; declare const enum LimitType { First = "F", Last = "L" } /** LimitType enum. */ declare type LimitType_2 = 'FIRST' | 'LAST'; declare type ListenSequenceNumber = number; declare class LLRBEmptyNode { get key(): never; get value(): never; get color(): never; get left(): never; get right(): never; size: number; copy(key: K | null, value: V | null, color: boolean | null, left: LLRBNode | LLRBEmptyNode | null, right: LLRBNode | LLRBEmptyNode | null): LLRBEmptyNode; insert(key: K, value: V, comparator: Comparator): LLRBNode; remove(key: K, comparator: Comparator): LLRBEmptyNode; isEmpty(): boolean; inorderTraversal(action: (k: K, v: V) => boolean): boolean; reverseTraversal(action: (k: K, v: V) => boolean): boolean; minKey(): K | null; maxKey(): K | null; isRed(): boolean; checkMaxDepth(): boolean; protected check(): 0; } declare class LLRBNode { key: K; value: V; readonly color: boolean; readonly left: LLRBNode | LLRBEmptyNode; readonly right: LLRBNode | LLRBEmptyNode; readonly size: number; static EMPTY: LLRBEmptyNode; static RED: boolean; static BLACK: boolean; constructor(key: K, value: V, color?: boolean, left?: LLRBNode | LLRBEmptyNode, right?: LLRBNode | LLRBEmptyNode); copy(key: K | null, value: V | null, color: boolean | null, left: LLRBNode | LLRBEmptyNode | null, right: LLRBNode | LLRBEmptyNode | null): LLRBNode; isEmpty(): boolean; inorderTraversal(action: (k: K, v: V) => T): T; reverseTraversal(action: (k: K, v: V) => T): T; private min; minKey(): K | null; maxKey(): K | null; insert(key: K, value: V, comparator: Comparator): LLRBNode; private removeMin; remove(key: K, comparator: Comparator): LLRBNode | LLRBEmptyNode; isRed(): boolean; private fixUp; private moveRedLeft; private moveRedRight; private rotateLeft; private rotateRight; private colorFlip; checkMaxDepth(): boolean; protected check(): number; } /** * A readonly view of the local state of all documents we're tracking (i.e. we * have a cached version in remoteDocumentCache or local mutations for the * document). The view is computed by applying the mutations in the * MutationQueue to the RemoteDocumentCache. */ declare class LocalDocumentsView { readonly remoteDocumentCache: RemoteDocumentCache; readonly mutationQueue: MutationQueue; readonly documentOverlayCache: DocumentOverlayCache; readonly indexManager: IndexManager; constructor(remoteDocumentCache: RemoteDocumentCache, mutationQueue: MutationQueue, documentOverlayCache: DocumentOverlayCache, indexManager: IndexManager); /** * Get the local view of the document identified by `key`. * * @returns Local view of the document or null if we don't have any cached * state for it. */ getDocument(transaction: PersistenceTransaction, key: DocumentKey): PersistencePromise; /** * Gets the local view of the documents identified by `keys`. * * If we don't have cached state for a document in `keys`, a NoDocument will * be stored for that key in the resulting set. */ getDocuments(transaction: PersistenceTransaction, keys: DocumentKeySet): PersistencePromise; /** * Similar to `getDocuments`, but creates the local view from the given * `baseDocs` without retrieving documents from the local store. * * @param transaction - The transaction this operation is scoped to. * @param docs - The documents to apply local mutations to get the local views. * @param existenceStateChanged - The set of document keys whose existence state * is changed. This is useful to determine if some documents overlay needs * to be recalculated. */ getLocalViewOfDocuments(transaction: PersistenceTransaction, docs: MutableDocumentMap, existenceStateChanged?: DocumentKeySet): PersistencePromise; /** * Gets the overlayed documents for the given document map, which will include * the local view of those documents and a `FieldMask` indicating which fields * are mutated locally, `null` if overlay is a Set or Delete mutation. */ getOverlayedDocuments(transaction: PersistenceTransaction, docs: MutableDocumentMap): PersistencePromise; /** * Fetches the overlays for {@code docs} and adds them to provided overlay map * if the map does not already contain an entry for the given document key. */ private populateOverlays; /** * Computes the local view for the given documents. * * @param docs - The documents to compute views for. It also has the base * version of the documents. * @param overlays - The overlays that need to be applied to the given base * version of the documents. * @param existenceStateChanged - A set of documents whose existence states * might have changed. This is used to determine if we need to re-calculate * overlays from mutation queues. * @return A map represents the local documents view. */ computeViews(transaction: PersistenceTransaction, docs: MutableDocumentMap, overlays: OverlayMap, existenceStateChanged: DocumentKeySet): PersistencePromise; private recalculateAndSaveOverlays; /** * Recalculates overlays by reading the documents from remote document cache * first, and saves them after they are calculated. */ recalculateAndSaveOverlaysForDocumentKeys(transaction: PersistenceTransaction, documentKeys: DocumentKeySet): PersistencePromise>; /** * Performs a query against the local view of all documents. * * @param transaction - The persistence transaction. * @param query - The query to match documents against. * @param offset - Read time and key to start scanning by (exclusive). * @param context - A optional tracker to keep a record of important details * during database local query execution. */ getDocumentsMatchingQuery(transaction: PersistenceTransaction, query: Query_2, offset: IndexOffset, context?: QueryContext): PersistencePromise; /** * Given a collection group, returns the next documents that follow the provided offset, along * with an updated batch ID. * *

The documents returned by this method are ordered by remote version from the provided * offset. If there are no more remote documents after the provided offset, documents with * mutations in order of batch id from the offset are returned. Since all documents in a batch are * returned together, the total number of documents returned can exceed {@code count}. * * @param transaction * @param collectionGroup The collection group for the documents. * @param offset The offset to index into. * @param count The number of documents to return * @return A LocalWriteResult with the documents that follow the provided offset and the last processed batch id. */ getNextDocuments(transaction: PersistenceTransaction, collectionGroup: string, offset: IndexOffset, count: number): PersistencePromise; private getDocumentsMatchingDocumentQuery; private getDocumentsMatchingCollectionGroupQuery; private getDocumentsMatchingCollectionQuery; } declare interface LocalStore { collectGarbage(garbageCollector: LruGarbageCollector): Promise; /** Manages the list of active field and collection indices. */ indexManager: IndexManager; /** * The "local" view of all documents (layering mutationQueue on top of * remoteDocumentCache). */ localDocuments: LocalDocumentsView; } /** The result of a write to the local store. */ declare interface LocalWriteResult { batchId: BatchId; changes: DocumentMap; } export { LogLevel } declare interface LruGarbageCollector { readonly params: LruParams; collect(txn: PersistenceTransaction, activeTargetIds: ActiveTargets): PersistencePromise; /** Given a percentile of target to collect, returns the number of targets to collect. */ calculateTargetCount(txn: PersistenceTransaction, percentile: number): PersistencePromise; /** Returns the nth sequence number, counting in order from the smallest. */ nthSequenceNumber(txn: PersistenceTransaction, n: number): PersistencePromise; /** * Removes documents that have a sequence number equal to or less than the * upper bound and are not otherwise pinned. */ removeOrphanedDocuments(txn: PersistenceTransaction, upperBound: ListenSequenceNumber): PersistencePromise; getCacheSize(txn: PersistenceTransaction): PersistencePromise; /** * Removes targets with a sequence number equal to or less than the given * upper bound, and removes document associations with those targets. */ removeTargets(txn: PersistenceTransaction, upperBound: ListenSequenceNumber, activeTargetIds: ActiveTargets): PersistencePromise; } declare class LruParams { readonly cacheSizeCollectionThreshold: number; readonly percentileToCollect: number; readonly maximumSequenceNumbersToCollect: number; private static readonly DEFAULT_COLLECTION_PERCENTILE; private static readonly DEFAULT_MAX_SEQUENCE_NUMBERS_TO_COLLECT; static withCacheSize(cacheSize: number): LruParams; static readonly DEFAULT: LruParams; static readonly DISABLED: LruParams; constructor(cacheSizeCollectionThreshold: number, percentileToCollect: number, maximumSequenceNumbersToCollect: number); } /** * Describes the results of a garbage collection run. `didRun` will be set to * `false` if collection was skipped (either it is disabled or the cache size * has not hit the threshold). If collection ran, the other fields will be * filled in with the details of the results. */ declare interface LruResults { readonly didRun: boolean; readonly sequenceNumbersCollected: number; readonly targetsRemoved: number; readonly documentsRemoved: number; } declare type MapValue = firestoreV1ApiClientInterfaces.MapValue; /** * Provides an in-memory cache to the SDK. This is the default cache unless explicitly * configured otherwise. * * To use, create an instance using the factory function {@link memoryLocalCache()}, then * set the instance to `FirestoreSettings.cache` and call `initializeFirestore` using * the settings object. */ declare type MemoryLocalCache = { kind: 'memory'; /* Excluded from this release type: _onlineComponentProvider */ /* Excluded from this release type: _offlineComponentProvider */ }; /** * Provides all components needed for Firestore with in-memory persistence. * Uses EagerGC garbage collection. */ declare class MemoryOfflineComponentProvider implements OfflineComponentProvider { persistence: Persistence; sharedClientState: SharedClientState; localStore: LocalStore; gcScheduler: Scheduler | null; indexBackfillerScheduler: Scheduler | null; synchronizeTabs: boolean; serializer: JsonProtoSerializer; initialize(cfg: ComponentConfiguration): Promise; createGarbageCollectionScheduler(cfg: ComponentConfiguration, localStore: LocalStore): Scheduler | null; createIndexBackfillerScheduler(cfg: ComponentConfiguration, localStore: LocalStore): Scheduler | null; createLocalStore(cfg: ComponentConfiguration): LocalStore; createPersistence(cfg: ComponentConfiguration): Persistence; createSharedClientState(cfg: ComponentConfiguration): SharedClientState; terminate(): Promise; } /** * Represents a document in Firestore with a key, version, data and whether it * has local mutations applied to it. * * Documents can transition between states via `convertToFoundDocument()`, * `convertToNoDocument()` and `convertToUnknownDocument()`. If a document does * not transition to one of these states even after all mutations have been * applied, `isValidDocument()` returns false and the document should be removed * from all views. */ declare class MutableDocument implements Document_2 { readonly key: DocumentKey; private documentType; version: SnapshotVersion; readTime: SnapshotVersion; createTime: SnapshotVersion; data: ObjectValue; private documentState; private constructor(); /** * Creates a document with no known version or data, but which can serve as * base document for mutations. */ static newInvalidDocument(documentKey: DocumentKey): MutableDocument; /** * Creates a new document that is known to exist with the given data at the * given version. */ static newFoundDocument(documentKey: DocumentKey, version: SnapshotVersion, createTime: SnapshotVersion, value: ObjectValue): MutableDocument; /** Creates a new document that is known to not exist at the given version. */ static newNoDocument(documentKey: DocumentKey, version: SnapshotVersion): MutableDocument; /** * Creates a new document that is known to exist at the given version but * whose data is not known (e.g. a document that was updated without a known * base document). */ static newUnknownDocument(documentKey: DocumentKey, version: SnapshotVersion): MutableDocument; /** * Changes the document type to indicate that it exists and that its version * and data are known. */ convertToFoundDocument(version: SnapshotVersion, value: ObjectValue): MutableDocument; /** * Changes the document type to indicate that it doesn't exist at the given * version. */ convertToNoDocument(version: SnapshotVersion): MutableDocument; /** * Changes the document type to indicate that it exists at a given version but * that its data is not known (e.g. a document that was updated without a known * base document). */ convertToUnknownDocument(version: SnapshotVersion): MutableDocument; setHasCommittedMutations(): MutableDocument; setHasLocalMutations(): MutableDocument; setReadTime(readTime: SnapshotVersion): MutableDocument; get hasLocalMutations(): boolean; get hasCommittedMutations(): boolean; get hasPendingWrites(): boolean; isValidDocument(): boolean; isFoundDocument(): boolean; isNoDocument(): boolean; isUnknownDocument(): boolean; isEqual(other: Document_2 | null | undefined): boolean; mutableCopy(): MutableDocument; toString(): string; } /** Miscellaneous collection types / constants. */ declare type MutableDocumentMap = SortedMap; /** * A mutation describes a self-contained change to a document. Mutations can * create, replace, delete, and update subsets of documents. * * Mutations not only act on the value of the document but also its version. * * For local mutations (mutations that haven't been committed yet), we preserve * the existing version for Set and Patch mutations. For Delete mutations, we * reset the version to 0. * * Here's the expected transition table. * * MUTATION APPLIED TO RESULTS IN * * SetMutation Document(v3) Document(v3) * SetMutation NoDocument(v3) Document(v0) * SetMutation InvalidDocument(v0) Document(v0) * PatchMutation Document(v3) Document(v3) * PatchMutation NoDocument(v3) NoDocument(v3) * PatchMutation InvalidDocument(v0) UnknownDocument(v3) * DeleteMutation Document(v3) NoDocument(v0) * DeleteMutation NoDocument(v3) NoDocument(v0) * DeleteMutation InvalidDocument(v0) NoDocument(v0) * * For acknowledged mutations, we use the updateTime of the WriteResponse as * the resulting version for Set and Patch mutations. As deletes have no * explicit update time, we use the commitTime of the WriteResponse for * Delete mutations. * * If a mutation is acknowledged by the backend but fails the precondition check * locally, we transition to an `UnknownDocument` and rely on Watch to send us * the updated version. * * Field transforms are used only with Patch and Set Mutations. We use the * `updateTransforms` message to store transforms, rather than the `transforms`s * messages. * * ## Subclassing Notes * * Every type of mutation needs to implement its own applyToRemoteDocument() and * applyToLocalView() to implement the actual behavior of applying the mutation * to some source document (see `setMutationApplyToRemoteDocument()` for an * example). */ declare abstract class Mutation { abstract readonly type: MutationType; abstract readonly key: DocumentKey; abstract readonly precondition: Precondition; abstract readonly fieldTransforms: FieldTransform[]; /** * Returns a `FieldMask` representing the fields that will be changed by * applying this mutation. Returns `null` if the mutation will overwrite the * entire document. */ abstract getFieldMask(): FieldMask | null; } /** * A batch of mutations that will be sent as one unit to the backend. */ declare class MutationBatch { batchId: BatchId; localWriteTime: Timestamp; baseMutations: Mutation[]; mutations: Mutation[]; /** * @param batchId - The unique ID of this mutation batch. * @param localWriteTime - The original write time of this mutation. * @param baseMutations - Mutations that are used to populate the base * values when this mutation is applied locally. This can be used to locally * overwrite values that are persisted in the remote document cache. Base * mutations are never sent to the backend. * @param mutations - The user-provided mutations in this mutation batch. * User-provided mutations are applied both locally and remotely on the * backend. */ constructor(batchId: BatchId, localWriteTime: Timestamp, baseMutations: Mutation[], mutations: Mutation[]); /** * Applies all the mutations in this MutationBatch to the specified document * to compute the state of the remote document * * @param document - The document to apply mutations to. * @param batchResult - The result of applying the MutationBatch to the * backend. */ applyToRemoteDocument(document: MutableDocument, batchResult: MutationBatchResult): void; /** * Computes the local view of a document given all the mutations in this * batch. * * @param document - The document to apply mutations to. * @param mutatedFields - Fields that have been updated before applying this mutation batch. * @returns A `FieldMask` representing all the fields that are mutated. */ applyToLocalView(document: MutableDocument, mutatedFields: FieldMask | null): FieldMask | null; /** * Computes the local view for all provided documents given the mutations in * this batch. Returns a `DocumentKey` to `Mutation` map which can be used to * replace all the mutation applications. */ applyToLocalDocumentSet(documentMap: OverlayedDocumentMap, documentsWithoutRemoteVersion: DocumentKeySet): MutationMap; keys(): DocumentKeySet; isEqual(other: MutationBatch): boolean; } /** The result of applying a mutation batch to the backend. */ declare class MutationBatchResult { readonly batch: MutationBatch; readonly commitVersion: SnapshotVersion; readonly mutationResults: MutationResult[]; /** * A pre-computed mapping from each mutated document to the resulting * version. */ readonly docVersions: DocumentVersionMap; private constructor(); /** * Creates a new MutationBatchResult for the given batch and results. There * must be one result for each mutation in the batch. This static factory * caches a document=>version mapping (docVersions). */ static from(batch: MutationBatch, commitVersion: SnapshotVersion, results: MutationResult[]): MutationBatchResult; } declare type MutationMap = DocumentKeyMap; /** A queue of mutations to apply to the remote store. */ declare interface MutationQueue { /** Returns true if this queue contains no mutation batches. */ checkEmpty(transaction: PersistenceTransaction): PersistencePromise; /** * Creates a new mutation batch and adds it to this mutation queue. * * @param transaction - The transaction this operation is scoped to. * @param localWriteTime - The original write time of this mutation. * @param baseMutations - Mutations that are used to populate the base values * when this mutation is applied locally. These mutations are used to locally * overwrite values that are persisted in the remote document cache. * @param mutations - The user-provided mutations in this mutation batch. */ addMutationBatch(transaction: PersistenceTransaction, localWriteTime: Timestamp, baseMutations: Mutation[], mutations: Mutation[]): PersistencePromise; /** * Loads the mutation batch with the given batchId. */ lookupMutationBatch(transaction: PersistenceTransaction, batchId: BatchId): PersistencePromise; /** * Gets the first unacknowledged mutation batch after the passed in batchId * in the mutation queue or null if empty. * * @param batchId - The batch to search after, or BATCHID_UNKNOWN for the * first mutation in the queue. * * @returns the next mutation or null if there wasn't one. */ getNextMutationBatchAfterBatchId(transaction: PersistenceTransaction, batchId: BatchId): PersistencePromise; /** * Gets the largest (latest) batch id in mutation queue for the current user * that is pending server response, returns `BATCHID_UNKNOWN` if the queue is * empty. * * @returns the largest batch id in the mutation queue that is not * acknowledged. */ getHighestUnacknowledgedBatchId(transaction: PersistenceTransaction): PersistencePromise; /** Gets all mutation batches in the mutation queue. */ getAllMutationBatches(transaction: PersistenceTransaction): PersistencePromise; /** * Finds all mutation batches that could possibly affect the given * document key. Not all mutations in a batch will necessarily affect the * document key, so when looping through the batch you'll need to check that * the mutation itself matches the key. * * Batches are guaranteed to be in sorted order. * * Note that because of this requirement implementations are free to return * mutation batches that don't contain the document key at all if it's * convenient. */ getAllMutationBatchesAffectingDocumentKey(transaction: PersistenceTransaction, documentKey: DocumentKey): PersistencePromise; /** * Finds all mutation batches that could possibly affect the given set of * document keys. Not all mutations in a batch will necessarily affect each * key, so when looping through the batch you'll need to check that the * mutation itself matches the key. * * Batches are guaranteed to be in sorted order. * * Note that because of this requirement implementations are free to return * mutation batches that don't contain any of the document keys at all if it's * convenient. */ getAllMutationBatchesAffectingDocumentKeys(transaction: PersistenceTransaction, documentKeys: SortedMap): PersistencePromise; /** * Finds all mutation batches that could affect the results for the given * query. Not all mutations in a batch will necessarily affect the query, so * when looping through the batch you'll need to check that the mutation * itself matches the query. * * Batches are guaranteed to be in sorted order. * * Note that because of this requirement implementations are free to return * mutation batches that don't match the query at all if it's convenient. * * NOTE: A PatchMutation does not need to include all fields in the query * filter criteria in order to be a match (but any fields it does contain do * need to match). */ getAllMutationBatchesAffectingQuery(transaction: PersistenceTransaction, query: Query_2): PersistencePromise; /** * Removes the given mutation batch from the queue. This is useful in two * circumstances: * * + Removing an applied mutation from the head of the queue * + Removing a rejected mutation from anywhere in the queue * * Multi-Tab Note: This operation should only be called by the primary client. */ removeMutationBatch(transaction: PersistenceTransaction, batch: MutationBatch): PersistencePromise; /** * Performs a consistency check, examining the mutation queue for any * leaks, if possible. */ performConsistencyCheck(transaction: PersistenceTransaction): PersistencePromise; } /** The result of successfully applying a mutation to the backend. */ declare class MutationResult { /** * The version at which the mutation was committed: * * - For most operations, this is the updateTime in the WriteResult. * - For deletes, the commitTime of the WriteResponse (because deletes are * not stored and have no updateTime). * * Note that these versions can be different: No-op writes will not change * the updateTime even though the commitTime advances. */ readonly version: SnapshotVersion; /** * The resulting fields returned from the backend after a mutation * containing field transforms has been committed. Contains one FieldValue * for each FieldTransform that was in the mutation. * * Will be empty if the mutation did not contain any field transforms. */ readonly transformResults: Array; constructor( /** * The version at which the mutation was committed: * * - For most operations, this is the updateTime in the WriteResult. * - For deletes, the commitTime of the WriteResponse (because deletes are * not stored and have no updateTime). * * Note that these versions can be different: No-op writes will not change * the updateTime even though the commitTime advances. */ version: SnapshotVersion, /** * The resulting fields returned from the backend after a mutation * containing field transforms has been committed. Contains one FieldValue * for each FieldTransform that was in the mutation. * * Will be empty if the mutation did not contain any field transforms. */ transformResults: Array); } declare const enum MutationType { Set = 0, Patch = 1, Delete = 2, Verify = 3 } /** * Represents a Query saved by the SDK in its local storage. */ declare interface NamedQuery { /** The name of the query. */ readonly name: string; /** The underlying query associated with `name`. */ readonly query: Query_2; /** The time at which the results for this query were read. */ readonly readTime: SnapshotVersion; } /** Properties of a NamedQuery. */ declare interface NamedQuery_2 { /** NamedQuery name */ name?: string | null; /** NamedQuery bundledQuery */ bundledQuery?: BundledQuery | null; /** NamedQuery readTime */ readTime?: Timestamp_2 | null; } /** * For each field (e.g. 'bar'), find all nested keys (e.g. {'bar.baz': T1, * 'bar.qux': T2}). Intersect them together to make a single map containing * all possible keys that are all marked as optional */ export declare type NestedUpdateFields> = UnionToIntersection<{ [K in keyof T & string]: ChildUpdateFields; }[keyof T & string]>; /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * A map implementation that uses objects as keys. Objects must have an * associated equals function and must be immutable. Entries in the map are * stored together with the key being produced from the mapKeyFn. This map * automatically handles collisions of keys. */ declare class ObjectMap { private mapKeyFn; private equalsFn; /** * The inner map for a key/value pair. Due to the possibility of collisions we * keep a list of entries that we do a linear search through to find an actual * match. Note that collisions should be rare, so we still expect near * constant time lookups in practice. */ private inner; /** The number of entries stored in the map */ private innerSize; constructor(mapKeyFn: (key: KeyType) => string, equalsFn: (l: KeyType, r: KeyType) => boolean); /** Get a value for this key, or undefined if it does not exist. */ get(key: KeyType): ValueType | undefined; has(key: KeyType): boolean; /** Put this key and value in the map. */ set(key: KeyType, value: ValueType): void; /** * Remove this key from the map. Returns a boolean if anything was deleted. */ delete(key: KeyType): boolean; forEach(fn: (key: KeyType, val: ValueType) => void): void; isEmpty(): boolean; size(): number; } /** * An ObjectValue represents a MapValue in the Firestore Proto and offers the * ability to add and remove fields (via the ObjectValueBuilder). */ declare class ObjectValue { readonly value: { mapValue: MapValue; }; constructor(value: { mapValue: MapValue; }); static empty(): ObjectValue; /** * Returns the value at the given path or null. * * @param path - the path to search * @returns The value at the path or null if the path is not set. */ field(path: FieldPath_2): Value | null; /** * Sets the field to the provided value. * * @param path - The field path to set. * @param value - The value to set. */ set(path: FieldPath_2, value: Value): void; /** * Sets the provided fields to the provided values. * * @param data - A map of fields to values (or null for deletes). */ setAll(data: Map): void; /** * Removes the field at the specified path. If there is no field at the * specified path, nothing is changed. * * @param path - The field path to remove. */ delete(path: FieldPath_2): void; isEqual(other: ObjectValue): boolean; /** * Returns the map that contains the leaf element of `path`. If the parent * entry does not yet exist, or if it is not a map, a new map will be created. */ private getFieldsMap; /** * Modifies `fieldsMap` by adding, replacing or deleting the specified * entries. */ private applyChanges; clone(): ObjectValue; } /** * Initializes and wires components that are needed to interface with the local * cache. Implementations override `initialize()` to provide all components. */ declare interface OfflineComponentProvider { persistence: Persistence; sharedClientState: SharedClientState; localStore: LocalStore; gcScheduler: Scheduler | null; indexBackfillerScheduler: Scheduler | null; synchronizeTabs: boolean; initialize(cfg: ComponentConfiguration): Promise; terminate(): Promise; } /** * Initializes and wires the components that are needed to interface with the * network. */ declare class OnlineComponentProvider { protected localStore: LocalStore; protected sharedClientState: SharedClientState; datastore: Datastore; eventManager: EventManager; remoteStore: RemoteStore; syncEngine: SyncEngine; initialize(offlineComponentProvider: OfflineComponentProvider, cfg: ComponentConfiguration): Promise; createEventManager(cfg: ComponentConfiguration): EventManager; createDatastore(cfg: ComponentConfiguration): Datastore; createRemoteStore(cfg: ComponentConfiguration): RemoteStore; createSyncEngine(cfg: ComponentConfiguration, startAsPrimary: boolean): SyncEngine; terminate(): Promise; } /** * Describes the online state of the Firestore client. Note that this does not * indicate whether or not the remote store is trying to connect or not. This is * primarily used by the View / EventManager code to change their behavior while * offline (e.g. get() calls shouldn't wait for data from the server and * snapshot events should set metadata.isFromCache=true). * * The string values should not be changed since they are persisted in * WebStorage. */ declare const enum OnlineState { /** * The Firestore client is in an unknown online state. This means the client * is either not actively trying to establish a connection or it is currently * trying to establish a connection, but it has not succeeded or failed yet. * Higher-level components should not operate in offline mode. */ Unknown = "Unknown", /** * The client is connected and the connections are healthy. This state is * reached after a successful connection and there has been at least one * successful message received from the backends. */ Online = "Online", /** * The client is either trying to establish a connection but failing, or it * has been explicitly marked offline via a call to disableNetwork(). * Higher-level components should operate in offline mode. */ Offline = "Offline" } declare const enum Operator { LESS_THAN = "<", LESS_THAN_OR_EQUAL = "<=", EQUAL = "==", NOT_EQUAL = "!=", GREATER_THAN = ">", GREATER_THAN_OR_EQUAL = ">=", ARRAY_CONTAINS = "array-contains", IN = "in", NOT_IN = "not-in", ARRAY_CONTAINS_ANY = "array-contains-any" } /** * Creates a new {@link QueryCompositeFilterConstraint} that is a disjunction of * the given filter constraints. A disjunction filter includes a document if it * satisfies any of the given filters. * * @param queryConstraints - Optional. The list of * {@link QueryFilterConstraint}s to perform a disjunction for. These must be * created with calls to {@link where}, {@link or}, or {@link and}. * @returns The newly created {@link QueryCompositeFilterConstraint}. */ export declare function or(...queryConstraints: QueryFilterConstraint[]): QueryCompositeFilterConstraint; /** * An ordering on a field, in some Direction. Direction defaults to ASCENDING. */ declare class OrderBy { readonly field: FieldPath_2; readonly dir: Direction; constructor(field: FieldPath_2, dir?: Direction); } /** * Creates a {@link QueryOrderByConstraint} that sorts the query result by the * specified field, optionally in descending order instead of ascending. * * Note: Documents that do not contain the specified field will not be present * in the query result. * * @param fieldPath - The field to sort by. * @param directionStr - Optional direction to sort by ('asc' or 'desc'). If * not specified, order will be ascending. * @returns The created {@link QueryOrderByConstraint}. */ export declare function orderBy(fieldPath: string | FieldPath, directionStr?: OrderByDirection): QueryOrderByConstraint; /** * The direction of a {@link orderBy} clause is specified as 'desc' or 'asc' * (descending or ascending). */ export declare type OrderByDirection = 'desc' | 'asc'; declare type OrderDirection = 'DIRECTION_UNSPECIFIED' | 'ASCENDING' | 'DESCENDING'; /** * Representation of an overlay computed by Firestore. * * Holds information about a mutation and the largest batch id in Firestore when * the mutation was created. */ declare class Overlay { readonly largestBatchId: number; readonly mutation: Mutation; constructor(largestBatchId: number, mutation: Mutation); getKey(): DocumentKey; isEqual(other: Overlay | null): boolean; toString(): string; } /** * Represents a local view (overlay) of a document, and the fields that are * locally mutated. */ declare class OverlayedDocument { readonly overlayedDocument: Document_2; /** * The fields that are locally mutated by patch mutations. * * If the overlayed document is from set or delete mutations, this is `null`. * If there is no overlay (mutation) for the document, this is an empty `FieldMask`. */ readonly mutatedFields: FieldMask | null; constructor(overlayedDocument: Document_2, /** * The fields that are locally mutated by patch mutations. * * If the overlayed document is from set or delete mutations, this is `null`. * If there is no overlay (mutation) for the document, this is an empty `FieldMask`. */ mutatedFields: FieldMask | null); } declare type OverlayedDocumentMap = DocumentKeyMap; declare type OverlayMap = DocumentKeyMap; declare interface ParseContext { readonly databaseId: DatabaseId; readonly ignoreUndefinedProperties: boolean; } /** The result of parsing document data (e.g. for a setData call). */ declare class ParsedSetData { readonly data: ObjectValue; readonly fieldMask: FieldMask | null; readonly fieldTransforms: FieldTransform[]; constructor(data: ObjectValue, fieldMask: FieldMask | null, fieldTransforms: FieldTransform[]); toMutation(key: DocumentKey, precondition: Precondition): Mutation; } /** The result of parsing "update" data (i.e. for an updateData call). */ declare class ParsedUpdateData { readonly data: ObjectValue; readonly fieldMask: FieldMask; readonly fieldTransforms: FieldTransform[]; constructor(data: ObjectValue, fieldMask: FieldMask, fieldTransforms: FieldTransform[]); toMutation(key: DocumentKey, precondition: Precondition): Mutation; } /** * Similar to Typescript's `Partial`, but allows nested fields to be * omitted and FieldValues to be passed in as property values. */ export declare type PartialWithFieldValue = Partial | (T extends Primitive ? T : T extends {} ? { [K in keyof T]?: PartialWithFieldValue | FieldValue; } : never); /** * Persistence is the lowest-level shared interface to persistent storage in * Firestore. * * Persistence is used to create MutationQueue and RemoteDocumentCache * instances backed by persistence (which might be in-memory or LevelDB). * * Persistence also exposes an API to create and run PersistenceTransactions * against persistence. All read / write operations must be wrapped in a * transaction. Implementations of PersistenceTransaction / Persistence only * need to guarantee that writes made against the transaction are not made to * durable storage until the transaction resolves its PersistencePromise. * Since memory-only storage components do not alter durable storage, they are * free to ignore the transaction. * * This contract is enough to allow the LocalStore be be written * independently of whether or not the stored state actually is durably * persisted. If persistent storage is enabled, writes are grouped together to * avoid inconsistent state that could cause crashes. * * Concretely, when persistent storage is enabled, the persistent versions of * MutationQueue, RemoteDocumentCache, and others (the mutators) will * defer their writes into a transaction. Once the local store has completed * one logical operation, it commits the transaction. * * When persistent storage is disabled, the non-persistent versions of the * mutators ignore the transaction. This short-cut is allowed because * memory-only storage leaves no state so it cannot be inconsistent. * * This simplifies the implementations of the mutators and allows memory-only * implementations to supplement the persistent ones without requiring any * special dual-store implementation of Persistence. The cost is that the * LocalStore needs to be slightly careful about the order of its reads and * writes in order to avoid relying on being able to read back uncommitted * writes. */ declare interface Persistence { /** * Whether or not this persistence instance has been started. */ readonly started: boolean; readonly referenceDelegate: ReferenceDelegate; /** Starts persistence. */ start(): Promise; /** * Releases any resources held during eager shutdown. */ shutdown(): Promise; /** * Registers a listener that gets called when the database receives a * version change event indicating that it has deleted. * * PORTING NOTE: This is only used for Web multi-tab. */ setDatabaseDeletedListener(databaseDeletedListener: () => Promise): void; /** * Adjusts the current network state in the client's metadata, potentially * affecting the primary lease. * * PORTING NOTE: This is only used for Web multi-tab. */ setNetworkEnabled(networkEnabled: boolean): void; /** * Returns a MutationQueue representing the persisted mutations for the * given user. * * Note: The implementation is free to return the same instance every time * this is called for a given user. In particular, the memory-backed * implementation does this to emulate the persisted implementation to the * extent possible (e.g. in the case of uid switching from * sally=>jack=>sally, sally's mutation queue will be preserved). */ getMutationQueue(user: User, indexManager: IndexManager): MutationQueue; /** * Returns a TargetCache representing the persisted cache of targets. * * Note: The implementation is free to return the same instance every time * this is called. In particular, the memory-backed implementation does this * to emulate the persisted implementation to the extent possible. */ getTargetCache(): TargetCache; /** * Returns a RemoteDocumentCache representing the persisted cache of remote * documents. * * Note: The implementation is free to return the same instance every time * this is called. In particular, the memory-backed implementation does this * to emulate the persisted implementation to the extent possible. */ getRemoteDocumentCache(): RemoteDocumentCache; /** * Returns a BundleCache representing the persisted cache of loaded bundles. * * Note: The implementation is free to return the same instance every time * this is called. In particular, the memory-backed implementation does this * to emulate the persisted implementation to the extent possible. */ getBundleCache(): BundleCache; /** * Returns an IndexManager instance that manages our persisted query indexes. * * Note: The implementation is free to return the same instance every time * this is called. In particular, the memory-backed implementation does this * to emulate the persisted implementation to the extent possible. */ getIndexManager(user: User): IndexManager; /** * Returns a DocumentOverlayCache representing the documents that are mutated * locally. */ getDocumentOverlayCache(user: User): DocumentOverlayCache; /** * Performs an operation inside a persistence transaction. Any reads or writes * against persistence must be performed within a transaction. Writes will be * committed atomically once the transaction completes. * * Persistence operations are asynchronous and therefore the provided * transactionOperation must return a PersistencePromise. When it is resolved, * the transaction will be committed and the Promise returned by this method * will resolve. * * @param action - A description of the action performed by this transaction, * used for logging. * @param mode - The underlying mode of the IndexedDb transaction. Can be * 'readonly', 'readwrite' or 'readwrite-primary'. Transactions marked * 'readwrite-primary' can only be executed by the primary client. In this * mode, the transactionOperation will not be run if the primary lease cannot * be acquired and the returned promise will be rejected with a * FAILED_PRECONDITION error. * @param transactionOperation - The operation to run inside a transaction. * @returns A `Promise` that is resolved once the transaction completes. */ runTransaction(action: string, mode: PersistenceTransactionMode, transactionOperation: (transaction: PersistenceTransaction) => PersistencePromise): Promise; } /** * PersistencePromise is essentially a re-implementation of Promise except * it has a .next() method instead of .then() and .next() and .catch() callbacks * are executed synchronously when a PersistencePromise resolves rather than * asynchronously (Promise implementations use setImmediate() or similar). * * This is necessary to interoperate with IndexedDB which will automatically * commit transactions if control is returned to the event loop without * synchronously initiating another operation on the transaction. * * NOTE: .then() and .catch() only allow a single consumer, unlike normal * Promises. */ declare class PersistencePromise { private nextCallback; private catchCallback; private result; private error; private isDone; private callbackAttached; constructor(callback: (resolve: Resolver, reject: Rejector) => void); catch(fn: (error: Error) => R | PersistencePromise): PersistencePromise; next(nextFn?: FulfilledHandler, catchFn?: RejectedHandler): PersistencePromise; toPromise(): Promise; private wrapUserFunction; private wrapSuccess; private wrapFailure; static resolve(): PersistencePromise; static resolve(result: R): PersistencePromise; static reject(error: Error): PersistencePromise; static waitFor(all: { forEach: (cb: (el: PersistencePromise) => void) => void; }): PersistencePromise; /** * Given an array of predicate functions that asynchronously evaluate to a * boolean, implements a short-circuiting `or` between the results. Predicates * will be evaluated until one of them returns `true`, then stop. The final * result will be whether any of them returned `true`. */ static or(predicates: Array<() => PersistencePromise>): PersistencePromise; /** * Given an iterable, call the given function on each element in the * collection and wait for all of the resulting concurrent PersistencePromises * to resolve. */ static forEach(collection: { forEach: (cb: (r: R, s: S) => void) => void; }, f: ((r: R, s: S) => PersistencePromise) | ((r: R) => PersistencePromise)): PersistencePromise; static forEach(collection: { forEach: (cb: (r: R) => void) => void; }, f: (r: R) => PersistencePromise): PersistencePromise; /** * Concurrently map all array elements through asynchronous function. */ static mapArray(array: T[], f: (t: T) => PersistencePromise): PersistencePromise; /** * An alternative to recursive PersistencePromise calls, that avoids * potential memory problems from unbounded chains of promises. * * The `action` will be called repeatedly while `condition` is true. */ static doWhile(condition: () => boolean, action: () => PersistencePromise): PersistencePromise; } /** * A base class representing a persistence transaction, encapsulating both the * transaction's sequence numbers as well as a list of onCommitted listeners. * * When you call Persistence.runTransaction(), it will create a transaction and * pass it to your callback. You then pass it to any method that operates * on persistence. */ declare abstract class PersistenceTransaction { private readonly onCommittedListeners; abstract readonly currentSequenceNumber: ListenSequenceNumber; addOnCommittedListener(listener: () => void): void; raiseOnCommittedEvent(): void; } /** The different modes supported by `Persistence.runTransaction()`. */ declare type PersistenceTransactionMode = 'readonly' | 'readwrite' | 'readwrite-primary'; /** * Provides a persistent cache backed by IndexedDb to the SDK. * * To use, create an instance using the factory function {@link persistentLocalCache()}, then * set the instance to `FirestoreSettings.cache` and call `initializeFirestore` using * the settings object. */ declare type PersistentLocalCache = { kind: 'persistent'; /* Excluded from this release type: _onlineComponentProvider */ /* Excluded from this release type: _offlineComponentProvider */ }; /** * Encodes a precondition for a mutation. This follows the model that the * backend accepts with the special case of an explicit "empty" precondition * (meaning no precondition). */ declare class Precondition { readonly updateTime?: SnapshotVersion | undefined; readonly exists?: boolean | undefined; private constructor(); /** Creates a new empty Precondition. */ static none(): Precondition; /** Creates a new Precondition with an exists flag. */ static exists(exists: boolean): Precondition; /** Creates a new Precondition based on a version a document exists at. */ static updateTime(version: SnapshotVersion): Precondition; /** Returns whether this Precondition is empty. */ get isNone(): boolean; isEqual(other: Precondition): boolean; } /** * These types primarily exist to support the `UpdateData`, * `WithFieldValue`, and `PartialWithFieldValue` types and are not consumed * directly by the end developer. */ /** Primitive types. */ export declare type Primitive = string | number | boolean | undefined | null; /* Excluded from this release type: PrivateSettings */ declare interface ProviderCredentialsSettings { ['type']: 'provider'; ['client']: CredentialsProvider; } /** * A `Query` refers to a query which you can read or listen to. You can also * construct refined `Query` objects by adding filters and ordering. */ export declare class Query { /** * If provided, the `FirestoreDataConverter` associated with this instance. */ readonly converter: FirestoreDataConverter | null; readonly _query: Query_2; /** The type of this Firestore reference. */ readonly type: 'query' | 'collection'; /** * The `Firestore` instance for the Firestore database (useful for performing * transactions, etc.). */ readonly firestore: Firestore; /** @hideconstructor protected */ constructor(firestore: Firestore, /** * If provided, the `FirestoreDataConverter` associated with this instance. */ converter: FirestoreDataConverter | null, _query: Query_2); /** * Removes the current converter. * * @param converter - `null` removes the current converter. * @returns A `Query` that does not use a * converter. */ withConverter(converter: null): Query; /** * Applies a custom data converter to this query, allowing you to use your own * custom model objects with Firestore. When you call {@link getDocs} with * the returned query, the provided converter will convert between Firestore * data of type `NewDbModelType` and your custom type `NewAppModelType`. * * @param converter - Converts objects to and from Firestore. * @returns A `Query` that uses the provided converter. */ withConverter(converter: FirestoreDataConverter): Query; } /** * Creates a new immutable instance of {@link Query} that is extended to also * include additional query constraints. * * @param query - The {@link Query} instance to use as a base for the new * constraints. * @param compositeFilter - The {@link QueryCompositeFilterConstraint} to * apply. Create {@link QueryCompositeFilterConstraint} using {@link and} or * {@link or}. * @param queryConstraints - Additional {@link QueryNonFilterConstraint}s to * apply (e.g. {@link orderBy}, {@link limit}). * @throws if any of the provided query constraints cannot be combined with the * existing or new constraints. */ export declare function query(query: Query, compositeFilter: QueryCompositeFilterConstraint, ...queryConstraints: QueryNonFilterConstraint[]): Query; /** * Creates a new immutable instance of {@link Query} that is extended to also * include additional query constraints. * * @param query - The {@link Query} instance to use as a base for the new * constraints. * @param queryConstraints - The list of {@link QueryConstraint}s to apply. * @throws if any of the provided query constraints cannot be combined with the * existing or new constraints. */ export declare function query(query: Query, ...queryConstraints: QueryConstraint[]): Query; /** * The Query interface defines all external properties of a query. * * QueryImpl implements this interface to provide memoization for `queryNormalizedOrderBy` * and `queryToTarget`. */ declare interface Query_2 { readonly path: ResourcePath; readonly collectionGroup: string | null; readonly explicitOrderBy: OrderBy[]; readonly filters: Filter[]; readonly limit: number | null; readonly limitType: LimitType; readonly startAt: Bound | null; readonly endAt: Bound | null; } /** * A `QueryCompositeFilterConstraint` is used to narrow the set of documents * returned by a Firestore query by performing the logical OR or AND of multiple * {@link QueryFieldFilterConstraint}s or {@link QueryCompositeFilterConstraint}s. * `QueryCompositeFilterConstraint`s are created by invoking {@link or} or * {@link and} and can then be passed to {@link (query:1)} to create a new query * instance that also contains the `QueryCompositeFilterConstraint`. */ export declare class QueryCompositeFilterConstraint extends AppliableConstraint { /** The type of this query constraint */ readonly type: 'or' | 'and'; private readonly _queryConstraints; /* Excluded from this release type: __constructor */ static _create(type: 'or' | 'and', _queryConstraints: QueryFilterConstraint[]): QueryCompositeFilterConstraint; _parse(query: Query): Filter; _apply(query: Query): Query; _getQueryConstraints(): readonly AppliableConstraint[]; _getOperator(): CompositeOperator; } /** * A `QueryConstraint` is used to narrow the set of documents returned by a * Firestore query. `QueryConstraint`s are created by invoking {@link where}, * {@link orderBy}, {@link (startAt:1)}, {@link (startAfter:1)}, {@link * (endBefore:1)}, {@link (endAt:1)}, {@link limit}, {@link limitToLast} and * can then be passed to {@link (query:1)} to create a new query instance that * also contains this `QueryConstraint`. */ export declare abstract class QueryConstraint extends AppliableConstraint { /** The type of this query constraint */ abstract readonly type: QueryConstraintType; /** * Takes the provided {@link Query} and returns a copy of the {@link Query} with this * {@link AppliableConstraint} applied. */ abstract _apply(query: Query): Query; } /** Describes the different query constraints available in this SDK. */ export declare type QueryConstraintType = 'where' | 'orderBy' | 'limit' | 'limitToLast' | 'startAt' | 'startAfter' | 'endAt' | 'endBefore'; /** * @license * Copyright 2023 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * A tracker to keep a record of important details during database local query * execution. */ declare class QueryContext { /** * Counts the number of documents passed through during local query execution. */ private _documentReadCount; get documentReadCount(): number; incrementDocumentReadCount(amount: number): void; } /** * A `QueryDocumentSnapshot` contains data read from a document in your * Firestore database as part of a query. The document is guaranteed to exist * and its data can be extracted with `.data()` or `.get()` to get a * specific field. * * A `QueryDocumentSnapshot` offers the same API surface as a * `DocumentSnapshot`. Since query results contain only existing documents, the * `exists` property will always be true and `data()` will never return * 'undefined'. */ export declare class QueryDocumentSnapshot extends DocumentSnapshot { /** * Retrieves all fields in the document as an `Object`. * * @override * @returns An `Object` containing all fields in the document. */ data(): AppModelType; } /** * A `QueryEndAtConstraint` is used to exclude documents from the end of a * result set returned by a Firestore query. * `QueryEndAtConstraint`s are created by invoking {@link (endAt:1)} or * {@link (endBefore:1)} and can then be passed to {@link (query:1)} to create a new * query instance that also contains this `QueryEndAtConstraint`. */ export declare class QueryEndAtConstraint extends QueryConstraint { /** The type of this query constraint */ readonly type: 'endBefore' | 'endAt'; private readonly _docOrFields; private readonly _inclusive; /* Excluded from this release type: __constructor */ static _create(type: 'endBefore' | 'endAt', _docOrFields: Array>, _inclusive: boolean): QueryEndAtConstraint; _apply(query: Query): Query; } /** * Returns true if the provided queries point to the same collection and apply * the same constraints. * * @param left - A `Query` to compare. * @param right - A `Query` to compare. * @returns true if the references point to the same location in the same * Firestore database. */ export declare function queryEqual(left: Query, right: Query): boolean; /** * A `QueryFieldFilterConstraint` is used to narrow the set of documents returned by * a Firestore query by filtering on one or more document fields. * `QueryFieldFilterConstraint`s are created by invoking {@link where} and can then * be passed to {@link (query:1)} to create a new query instance that also contains * this `QueryFieldFilterConstraint`. */ export declare class QueryFieldFilterConstraint extends QueryConstraint { private readonly _field; private _op; private _value; /** The type of this query constraint */ readonly type = "where"; /* Excluded from this release type: __constructor */ static _create(_field: FieldPath_2, _op: Operator, _value: unknown): QueryFieldFilterConstraint; _apply(query: Query): Query; _parse(query: Query): FieldFilter; } /** * `QueryFilterConstraint` is a helper union type that represents * {@link QueryFieldFilterConstraint} and {@link QueryCompositeFilterConstraint}. */ export declare type QueryFilterConstraint = QueryFieldFilterConstraint | QueryCompositeFilterConstraint; /** * A `QueryLimitConstraint` is used to limit the number of documents returned by * a Firestore query. * `QueryLimitConstraint`s are created by invoking {@link limit} or * {@link limitToLast} and can then be passed to {@link (query:1)} to create a new * query instance that also contains this `QueryLimitConstraint`. */ export declare class QueryLimitConstraint extends QueryConstraint { /** The type of this query constraint */ readonly type: 'limit' | 'limitToLast'; private readonly _limit; private readonly _limitType; /* Excluded from this release type: __constructor */ static _create(type: 'limit' | 'limitToLast', _limit: number, _limitType: LimitType): QueryLimitConstraint; _apply(query: Query): Query; } /** * `QueryNonFilterConstraint` is a helper union type that represents * QueryConstraints which are used to narrow or order the set of documents, * but that do not explicitly filter on a document field. * `QueryNonFilterConstraint`s are created by invoking {@link orderBy}, * {@link (startAt:1)}, {@link (startAfter:1)}, {@link (endBefore:1)}, {@link (endAt:1)}, * {@link limit} or {@link limitToLast} and can then be passed to {@link (query:1)} * to create a new query instance that also contains the `QueryConstraint`. */ export declare type QueryNonFilterConstraint = QueryOrderByConstraint | QueryLimitConstraint | QueryStartAtConstraint | QueryEndAtConstraint; /** * A `QueryOrderByConstraint` is used to sort the set of documents returned by a * Firestore query. `QueryOrderByConstraint`s are created by invoking * {@link orderBy} and can then be passed to {@link (query:1)} to create a new query * instance that also contains this `QueryOrderByConstraint`. * * Note: Documents that do not contain the orderBy field will not be present in * the query result. */ export declare class QueryOrderByConstraint extends QueryConstraint { private readonly _field; private _direction; /** The type of this query constraint */ readonly type = "orderBy"; /* Excluded from this release type: __constructor */ static _create(_field: FieldPath_2, _direction: Direction): QueryOrderByConstraint; _apply(query: Query): Query; } /** * A `QuerySnapshot` contains zero or more `DocumentSnapshot` objects * representing the results of a query. The documents can be accessed as an * array via the `docs` property or enumerated using the `forEach` method. The * number of documents can be determined via the `empty` and `size` * properties. */ export declare class QuerySnapshot { readonly _docs: Array>; /** * The query on which you called {@link getDocs} in order to get this * `QuerySnapshot`. */ readonly query: Query; /** @hideconstructor */ constructor(_query: Query, _docs: Array>); /** An array of all the documents in the `QuerySnapshot`. */ get docs(): Array>; /** The number of documents in the `QuerySnapshot`. */ get size(): number; /** True if there are no documents in the `QuerySnapshot`. */ get empty(): boolean; /** * Enumerates all of the documents in the `QuerySnapshot`. * * @param callback - A callback to be called with a `QueryDocumentSnapshot` for * each document in the snapshot. * @param thisArg - The `this` binding for the callback. */ forEach(callback: (result: QueryDocumentSnapshot) => void, thisArg?: unknown): void; } /** * A `QueryStartAtConstraint` is used to exclude documents from the start of a * result set returned by a Firestore query. * `QueryStartAtConstraint`s are created by invoking {@link (startAt:1)} or * {@link (startAfter:1)} and can then be passed to {@link (query:1)} to create a * new query instance that also contains this `QueryStartAtConstraint`. */ export declare class QueryStartAtConstraint extends QueryConstraint { /** The type of this query constraint */ readonly type: 'startAt' | 'startAfter'; private readonly _docOrFields; private readonly _inclusive; /* Excluded from this release type: __constructor */ static _create(type: 'startAt' | 'startAfter', _docOrFields: Array>, _inclusive: boolean): QueryStartAtConstraint; _apply(query: Query): Query; } /** The different states of a watch target. */ declare type QueryTargetState = 'not-current' | 'current' | 'rejected'; /** * Returns true if the provided references are equal. * * @param left - A reference to compare. * @param right - A reference to compare. * @returns true if the references point to the same location in the same * Firestore database. */ export declare function refEqual(left: DocumentReference | CollectionReference, right: DocumentReference | CollectionReference): boolean; /** * A ReferenceDelegate instance handles all of the hooks into the document-reference lifecycle. This * includes being added to a target, being removed from a target, being subject to mutation, and * being mutated by the user. * * Different implementations may do different things with each of these events. Not every * implementation needs to do something with every lifecycle hook. * * PORTING NOTE: since sequence numbers are attached to transactions in this * client, the ReferenceDelegate does not need to deal in transactional * semantics (onTransactionStarted/Committed()), nor does it need to track and * generate sequence numbers (getCurrentSequenceNumber()). */ declare interface ReferenceDelegate { /** Notify the delegate that the given document was added to a target. */ addReference(txn: PersistenceTransaction, targetId: TargetId, doc: DocumentKey): PersistencePromise; /** Notify the delegate that the given document was removed from a target. */ removeReference(txn: PersistenceTransaction, targetId: TargetId, doc: DocumentKey): PersistencePromise; /** * Notify the delegate that a target was removed. The delegate may, but is not obligated to, * actually delete the target and associated data. */ removeTarget(txn: PersistenceTransaction, targetData: TargetData): PersistencePromise; /** * Notify the delegate that a document may no longer be part of any views or * have any mutations associated. */ markPotentiallyOrphaned(txn: PersistenceTransaction, doc: DocumentKey): PersistencePromise; /** Notify the delegate that a limbo document was updated. */ updateLimboDocument(txn: PersistenceTransaction, doc: DocumentKey): PersistencePromise; } declare type RejectedHandler = ((reason: Error) => R | PersistencePromise) | null; declare type Rejector = (error: Error) => void; /** * Represents cached documents received from the remote backend. * * The cache is keyed by DocumentKey and entries in the cache are * MutableDocuments, meaning we can cache both actual documents as well as * documents that are known to not exist. */ declare interface RemoteDocumentCache { /** Sets the index manager to use for managing the collectionGroup index. */ setIndexManager(indexManager: IndexManager): void; /** * Looks up an entry in the cache. * * @param documentKey - The key of the entry to look up.* * @returns The cached document entry. Returns an invalid document if the * document is not cached. */ getEntry(transaction: PersistenceTransaction, documentKey: DocumentKey): PersistencePromise; /** * Looks up a set of entries in the cache. * * @param documentKeys - The keys of the entries to look up. * @returns The cached document entries indexed by key. If an entry is not * cached, the corresponding key will be mapped to an invalid document. */ getEntries(transaction: PersistenceTransaction, documentKeys: DocumentKeySet): PersistencePromise; /** * Returns the documents matching the given query * * @param query - The query to match documents against. * @param offset - The offset to start the scan at (exclusive). * @param context - A optional tracker to keep a record of important details * during database local query execution. * @returns The set of matching documents. */ getDocumentsMatchingQuery(transaction: PersistenceTransaction, query: Query_2, offset: IndexOffset, mutatedDocs: OverlayMap, context?: QueryContext): PersistencePromise; /** * Looks up the next `limit` documents for a collection group based on the * provided offset. The ordering is based on the document's read time and key. * * @param collectionGroup - The collection group to scan. * @param offset - The offset to start the scan at (exclusive). * @param limit - The maximum number of results to return. * @returns The set of matching documents. */ getAllFromCollectionGroup(transaction: PersistenceTransaction, collectionGroup: string, offset: IndexOffset, limit: number): PersistencePromise; /** * Provides access to add or update the contents of the cache. The buffer * handles proper size accounting for the change. * * Multi-Tab Note: This should only be called by the primary client. * * @param options - Specify `trackRemovals` to create sentinel entries for * removed documents, which allows removals to be tracked by * `getNewDocumentChanges()`. */ newChangeBuffer(options?: { trackRemovals: boolean; }): RemoteDocumentChangeBuffer; /** * Get an estimate of the size of the document cache. Note that for eager * garbage collection, we don't track sizes so this will return 0. */ getSize(transaction: PersistenceTransaction): PersistencePromise; } /** * An in-memory buffer of entries to be written to a RemoteDocumentCache. * It can be used to batch up a set of changes to be written to the cache, but * additionally supports reading entries back with the `getEntry()` method, * falling back to the underlying RemoteDocumentCache if no entry is * buffered. * * Entries added to the cache *must* be read first. This is to facilitate * calculating the size delta of the pending changes. * * PORTING NOTE: This class was implemented then removed from other platforms. * If byte-counting ends up being needed on the other platforms, consider * porting this class as part of that implementation work. */ declare abstract class RemoteDocumentChangeBuffer { protected changes: ObjectMap; private changesApplied; protected abstract getFromCache(transaction: PersistenceTransaction, documentKey: DocumentKey): PersistencePromise; protected abstract getAllFromCache(transaction: PersistenceTransaction, documentKeys: DocumentKeySet): PersistencePromise; protected abstract applyChanges(transaction: PersistenceTransaction): PersistencePromise; /** * Buffers a `RemoteDocumentCache.addEntry()` call. * * You can only modify documents that have already been retrieved via * `getEntry()/getEntries()` (enforced via IndexedDbs `apply()`). */ addEntry(document: MutableDocument): void; /** * Buffers a `RemoteDocumentCache.removeEntry()` call. * * You can only remove documents that have already been retrieved via * `getEntry()/getEntries()` (enforced via IndexedDbs `apply()`). */ removeEntry(key: DocumentKey, readTime: SnapshotVersion): void; /** * Looks up an entry in the cache. The buffered changes will first be checked, * and if no buffered change applies, this will forward to * `RemoteDocumentCache.getEntry()`. * * @param transaction - The transaction in which to perform any persistence * operations. * @param documentKey - The key of the entry to look up. * @returns The cached document or an invalid document if we have nothing * cached. */ getEntry(transaction: PersistenceTransaction, documentKey: DocumentKey): PersistencePromise; /** * Looks up several entries in the cache, forwarding to * `RemoteDocumentCache.getEntry()`. * * @param transaction - The transaction in which to perform any persistence * operations. * @param documentKeys - The keys of the entries to look up. * @returns A map of cached documents, indexed by key. If an entry cannot be * found, the corresponding key will be mapped to an invalid document. */ getEntries(transaction: PersistenceTransaction, documentKeys: DocumentKeySet): PersistencePromise; /** * Applies buffered changes to the underlying RemoteDocumentCache, using * the provided transaction. */ apply(transaction: PersistenceTransaction): PersistencePromise; /** Helper to assert this.changes is not null */ protected assertNotApplied(): void; } /** * An event from the RemoteStore. It is split into targetChanges (changes to the * state or the set of documents in our watched targets) and documentUpdates * (changes to the actual documents). */ declare class RemoteEvent { /** * The snapshot version this event brings us up to, or MIN if not set. */ readonly snapshotVersion: SnapshotVersion; /** * A map from target to changes to the target. See TargetChange. */ readonly targetChanges: Map; /** * A map of targets that is known to be inconsistent, and the purpose for * re-listening. Listens for these targets should be re-established without * resume tokens. */ readonly targetMismatches: SortedMap; /** * A set of which documents have changed or been deleted, along with the * doc's new values (if not deleted). */ readonly documentUpdates: MutableDocumentMap; /** * A set of which document updates are due only to limbo resolution targets. */ readonly resolvedLimboDocuments: DocumentKeySet; constructor( /** * The snapshot version this event brings us up to, or MIN if not set. */ snapshotVersion: SnapshotVersion, /** * A map from target to changes to the target. See TargetChange. */ targetChanges: Map, /** * A map of targets that is known to be inconsistent, and the purpose for * re-listening. Listens for these targets should be re-established without * resume tokens. */ targetMismatches: SortedMap, /** * A set of which documents have changed or been deleted, along with the * doc's new values (if not deleted). */ documentUpdates: MutableDocumentMap, /** * A set of which document updates are due only to limbo resolution targets. */ resolvedLimboDocuments: DocumentKeySet); /** * HACK: Views require RemoteEvents in order to determine whether the view is * CURRENT, but secondary tabs don't receive remote events. So this method is * used to create a synthesized RemoteEvent that can be used to apply a * CURRENT status change to a View, for queries executed in a different tab. */ static createSynthesizedRemoteEventForCurrentChange(targetId: TargetId, current: boolean, resumeToken: ByteString): RemoteEvent; } /** * RemoteStore - An interface to remotely stored data, basically providing a * wrapper around the Datastore that is more reliable for the rest of the * system. * * RemoteStore is responsible for maintaining the connection to the server. * - maintaining a list of active listens. * - reconnecting when the connection is dropped. * - resuming all the active listens on reconnect. * * RemoteStore handles all incoming events from the Datastore. * - listening to the watch stream and repackaging the events as RemoteEvents * - notifying SyncEngine of any changes to the active listens. * * RemoteStore takes writes from other components and handles them reliably. * - pulling pending mutations from LocalStore and sending them to Datastore. * - retrying mutations that failed because of network problems. * - acking mutations to the SyncEngine once they are accepted or rejected. */ declare interface RemoteStore { /** * SyncEngine to notify of watch and write events. This must be set * immediately after construction. */ remoteSyncer: RemoteSyncer; } /** * An interface that describes the actions the RemoteStore needs to perform on * a cooperating synchronization engine. */ declare interface RemoteSyncer { /** * Applies one remote event to the sync engine, notifying any views of the * changes, and releasing any pending mutation batches that would become * visible because of the snapshot version the remote event contains. */ applyRemoteEvent?(remoteEvent: RemoteEvent): Promise; /** * Rejects the listen for the given targetID. This can be triggered by the * backend for any active target. * * @param targetId - The targetID corresponds to one previously initiated by * the user as part of TargetData passed to listen() on RemoteStore. * @param error - A description of the condition that has forced the rejection. * Nearly always this will be an indication that the user is no longer * authorized to see the data matching the target. */ rejectListen?(targetId: TargetId, error: FirestoreError): Promise; /** * Applies the result of a successful write of a mutation batch to the sync * engine, emitting snapshots in any views that the mutation applies to, and * removing the batch from the mutation queue. */ applySuccessfulWrite?(result: MutationBatchResult): Promise; /** * Rejects the batch, removing the batch from the mutation queue, recomputing * the local view of any documents affected by the batch and then, emitting * snapshots with the reverted value. */ rejectFailedWrite?(batchId: BatchId, error: FirestoreError): Promise; /** * Returns the set of remote document keys for the given target ID. This list * includes the documents that were assigned to the target when we received * the last snapshot. */ getRemoteKeysForTarget?(targetId: TargetId): DocumentKeySet; /** * Updates all local state to match the pending mutations for the given user. * May be called repeatedly for the same user. */ handleCredentialChange?(user: User): Promise; } declare type Resolver = (value?: T) => void; /* Excluded from this release type: ResourcePath */ /** * Executes the given `updateFunction` and then attempts to commit the changes * applied within the transaction. If any document read within the transaction * has changed, Cloud Firestore retries the `updateFunction`. If it fails to * commit after 5 attempts, the transaction fails. * * The maximum number of writes allowed in a single transaction is 500. * * @param firestore - A reference to the Firestore database to run this * transaction against. * @param updateFunction - The function to execute within the transaction * context. * @param options - An options object to configure maximum number of attempts to * commit. * @returns If the transaction completed successfully or was explicitly aborted * (the `updateFunction` returned a failed promise), the promise returned by the * `updateFunction `is returned here. Otherwise, if the transaction failed, a * rejected promise with the corresponding failure error is returned. */ export declare function runTransaction(firestore: Firestore, updateFunction: (transaction: Transaction) => Promise, options?: TransactionOptions): Promise; /** * Interface to schedule periodic tasks within SDK. */ declare interface Scheduler { readonly started: boolean; start(): void; stop(): void; } /** Base interface for the Serializer implementation. */ declare interface Serializer { readonly useProto3Json: boolean; } /** * Returns a sentinel used with {@link @firebase/firestore/lite#(setDoc:1)} or {@link @firebase/firestore/lite#(updateDoc:1)} to * include a server-generated timestamp in the written data. */ export declare function serverTimestamp(): FieldValue; declare type ServerTimestampBehavior = 'estimate' | 'previous' | 'none'; /** * Writes to the document referred to by the specified `DocumentReference`. If * the document does not yet exist, it will be created. * * The result of this write will only be reflected in document reads that occur * after the returned promise resolves. If the client is offline, the * write fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @param reference - A reference to the document to write. * @param data - A map of the fields and values for the document. * @throws Error - If the provided input is not a valid Firestore document. * @returns A `Promise` resolved once the data has been successfully written * to the backend. */ export declare function setDoc(reference: DocumentReference, data: WithFieldValue): Promise; /** * Writes to the document referred to by the specified `DocumentReference`. If * the document does not yet exist, it will be created. If you provide `merge` * or `mergeFields`, the provided data can be merged into an existing document. * * The result of this write will only be reflected in document reads that occur * after the returned promise resolves. If the client is offline, the * write fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @param reference - A reference to the document to write. * @param data - A map of the fields and values for the document. * @param options - An object to configure the set behavior. * @throws Error - If the provided input is not a valid Firestore document. * @returns A `Promise` resolved once the data has been successfully written * to the backend. */ export declare function setDoc(reference: DocumentReference, data: PartialWithFieldValue, options: SetOptions): Promise; /** * Sets the verbosity of Cloud Firestore logs (debug, error, or silent). * * @param logLevel - The verbosity you set for activity and error logging. Can * be any of the following values: * *

    *
  • `debug` for the most verbose logging level, primarily for * debugging.
  • *
  • `error` to log errors only.
  • *
  • `silent` to turn off logging.
  • *
*/ export declare function setLogLevel(logLevel: LogLevel): void; /** * An options object that configures the behavior of {@link @firebase/firestore/lite#(setDoc:1)}, {@link * @firebase/firestore/lite#(WriteBatch.set:1)} and {@link @firebase/firestore/lite#(Transaction.set:1)} calls. These calls can be * configured to perform granular merges instead of overwriting the target * documents in their entirety by providing a `SetOptions` with `merge: true`. * * @param merge - Changes the behavior of a `setDoc()` call to only replace the * values specified in its data argument. Fields omitted from the `setDoc()` * call remain untouched. If your input sets any field to an empty map, all * nested fields are overwritten. * @param mergeFields - Changes the behavior of `setDoc()` calls to only replace * the specified field paths. Any field path that is not specified is ignored * and remains untouched. If your input sets any field to an empty map, all * nested fields are overwritten. */ export declare type SetOptions = { readonly merge?: boolean; } | { readonly mergeFields?: Array; }; /** * Specifies custom configurations for your Cloud Firestore instance. * You must set these before invoking any other methods. */ export declare interface Settings { /** The hostname to connect to. */ host?: string; /** Whether to use SSL when connecting. */ ssl?: boolean; /** * Whether to skip nested properties that are set to `undefined` during * object serialization. If set to `true`, these properties are skipped * and not written to Firestore. If set to `false` or omitted, the SDK * throws an exception when it encounters properties of type `undefined`. */ ignoreUndefinedProperties?: boolean; } /** * A `SharedClientState` keeps track of the global state of the mutations * and query targets for all active clients with the same persistence key (i.e. * project ID and FirebaseApp name). It relays local changes to other clients * and updates its local state as new state is observed. * * `SharedClientState` is primarily used for synchronization in Multi-Tab * environments. Each tab is responsible for registering its active query * targets and mutations. `SharedClientState` will then notify the listener * assigned to `.syncEngine` for updates to mutations and queries that * originated in other clients. * * To receive notifications, `.syncEngine` and `.onlineStateHandler` has to be * assigned before calling `start()`. */ declare interface SharedClientState { onlineStateHandler: ((onlineState: OnlineState) => void) | null; sequenceNumberHandler: ((sequenceNumber: ListenSequenceNumber) => void) | null; /** Registers the Mutation Batch ID of a newly pending mutation. */ addPendingMutation(batchId: BatchId): void; /** * Records that a pending mutation has been acknowledged or rejected. * Called by the primary client to notify secondary clients of mutation * results as they come back from the backend. */ updateMutationState(batchId: BatchId, state: 'acknowledged' | 'rejected', error?: FirestoreError): void; /** * Associates a new Query Target ID with the local Firestore client. Returns * the new query state for the query (which can be 'current' if the query is * already associated with another tab). * * If the target id is already associated with local client, the method simply * returns its `QueryTargetState`. */ addLocalQueryTarget(targetId: TargetId): QueryTargetState; /** Removes the Query Target ID association from the local client. */ removeLocalQueryTarget(targetId: TargetId): void; /** Checks whether the target is associated with the local client. */ isLocalQueryTarget(targetId: TargetId): boolean; /** * Processes an update to a query target. * * Called by the primary client to notify secondary clients of document * changes or state transitions that affect the provided query target. */ updateQueryState(targetId: TargetId, state: QueryTargetState, error?: FirestoreError): void; /** * Removes the target's metadata entry. * * Called by the primary client when all clients stopped listening to a query * target. */ clearQueryState(targetId: TargetId): void; /** * Gets the active Query Targets IDs for all active clients. * * The implementation for this may require O(n) runtime, where 'n' is the size * of the result set. */ getAllActiveQueryTargets(): SortedSet; /** * Checks whether the provided target ID is currently being listened to by * any of the active clients. * * The implementation may require O(n*log m) runtime, where 'n' is the number * of clients and 'm' the number of targets. */ isActiveQueryTarget(targetId: TargetId): boolean; /** * Starts the SharedClientState, reads existing client data and registers * listeners for updates to new and existing clients. */ start(): Promise; /** Shuts down the `SharedClientState` and its listeners. */ shutdown(): void; /** * Changes the active user and removes all existing user-specific data. The * user change does not call back into SyncEngine (for example, no mutations * will be marked as removed). */ handleUserChange(user: User, removedBatchIds: BatchId[], addedBatchIds: BatchId[]): void; /** Changes the shared online state of all clients. */ setOnlineState(onlineState: OnlineState): void; writeSequenceNumber(sequenceNumber: ListenSequenceNumber): void; /** * Notifies other clients when remote documents have changed due to loading * a bundle. * * @param collectionGroups The collection groups affected by this bundle. */ notifyBundleLoaded(collectionGroups: Set): void; } /** * Returns true if the provided snapshots are equal. * * @param left - A snapshot to compare. * @param right - A snapshot to compare. * @returns true if the snapshots are equal. */ export declare function snapshotEqual(left: DocumentSnapshot | QuerySnapshot, right: DocumentSnapshot | QuerySnapshot): boolean; /** * A version of a document in Firestore. This corresponds to the version * timestamp, such as update_time or read_time. */ declare class SnapshotVersion { private timestamp; static fromTimestamp(value: Timestamp): SnapshotVersion; static min(): SnapshotVersion; static max(): SnapshotVersion; private constructor(); compareTo(other: SnapshotVersion): number; isEqual(other: SnapshotVersion): boolean; /** Returns a number representation of the version for use in spec tests. */ toMicroseconds(): number; toString(): string; toTimestamp(): Timestamp; } declare class SortedMap { comparator: Comparator; root: LLRBNode | LLRBEmptyNode; constructor(comparator: Comparator, root?: LLRBNode | LLRBEmptyNode); insert(key: K, value: V): SortedMap; remove(key: K): SortedMap; get(key: K): V | null; indexOf(key: K): number; isEmpty(): boolean; get size(): number; minKey(): K | null; maxKey(): K | null; inorderTraversal(action: (k: K, v: V) => T): T; forEach(fn: (k: K, v: V) => void): void; toString(): string; reverseTraversal(action: (k: K, v: V) => T): T; getIterator(): SortedMapIterator; getIteratorFrom(key: K): SortedMapIterator; getReverseIterator(): SortedMapIterator; getReverseIteratorFrom(key: K): SortedMapIterator; } declare class SortedMapIterator { private isReverse; private nodeStack; constructor(node: LLRBNode | LLRBEmptyNode, startKey: K | null, comparator: Comparator, isReverse: boolean); getNext(): Entry; hasNext(): boolean; peek(): Entry | null; } /** * SortedSet is an immutable (copy-on-write) collection that holds elements * in order specified by the provided comparator. * * NOTE: if provided comparator returns 0 for two elements, we consider them to * be equal! */ declare class SortedSet { private comparator; private data; constructor(comparator: (left: T, right: T) => number); has(elem: T): boolean; first(): T | null; last(): T | null; get size(): number; indexOf(elem: T): number; /** Iterates elements in order defined by "comparator" */ forEach(cb: (elem: T) => void): void; /** Iterates over `elem`s such that: range[0] <= elem < range[1]. */ forEachInRange(range: [T, T], cb: (elem: T) => void): void; /** * Iterates over `elem`s such that: start <= elem until false is returned. */ forEachWhile(cb: (elem: T) => boolean, start?: T): void; /** Finds the least element greater than or equal to `elem`. */ firstAfterOrEqual(elem: T): T | null; getIterator(): SortedSetIterator; getIteratorFrom(key: T): SortedSetIterator; /** Inserts or updates an element */ add(elem: T): SortedSet; /** Deletes an element */ delete(elem: T): SortedSet; isEmpty(): boolean; unionWith(other: SortedSet): SortedSet; isEqual(other: SortedSet): boolean; toArray(): T[]; toString(): string; private copy; } declare class SortedSetIterator { private iter; constructor(iter: SortedMapIterator); getNext(): T; hasNext(): boolean; } /** * Creates a {@link QueryStartAtConstraint} that modifies the result set to * start after the provided document (exclusive). The starting position is * relative to the order of the query. The document must contain all of the * fields provided in the orderBy of the query. * * @param snapshot - The snapshot of the document to start after. * @returns A {@link QueryStartAtConstraint} to pass to `query()` */ export declare function startAfter(snapshot: DocumentSnapshot): QueryStartAtConstraint; /** * Creates a {@link QueryStartAtConstraint} that modifies the result set to * start after the provided fields relative to the order of the query. The order * of the field values must match the order of the order by clauses of the query. * * @param fieldValues - The field values to start this query after, in order * of the query's order by. * @returns A {@link QueryStartAtConstraint} to pass to `query()` */ export declare function startAfter(...fieldValues: unknown[]): QueryStartAtConstraint; /** * Creates a {@link QueryStartAtConstraint} that modifies the result set to * start at the provided document (inclusive). The starting position is relative * to the order of the query. The document must contain all of the fields * provided in the `orderBy` of this query. * * @param snapshot - The snapshot of the document to start at. * @returns A {@link QueryStartAtConstraint} to pass to `query()`. */ export declare function startAt(snapshot: DocumentSnapshot): QueryStartAtConstraint; /** * Creates a {@link QueryStartAtConstraint} that modifies the result set to * start at the provided fields relative to the order of the query. The order of * the field values must match the order of the order by clauses of the query. * * @param fieldValues - The field values to start this query at, in order * of the query's order by. * @returns A {@link QueryStartAtConstraint} to pass to `query()`. */ export declare function startAt(...fieldValues: unknown[]): QueryStartAtConstraint; declare type StructuredQuery = firestoreV1ApiClientInterfaces.StructuredQuery; /** * Create an AggregateField object that can be used to compute the sum of * a specified field over a range of documents in the result set of a query. * @param field Specifies the field to sum across the result set. */ export declare function sum(field: string | FieldPath): AggregateField; /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * SyncEngine is the central controller in the client SDK architecture. It is * the glue code between the EventManager, LocalStore, and RemoteStore. Some of * SyncEngine's responsibilities include: * 1. Coordinating client requests and remote events between the EventManager * and the local and remote data stores. * 2. Managing a View object for each query, providing the unified view between * the local and remote data stores. * 3. Notifying the RemoteStore when the LocalStore has new mutations in its * queue that need sending to the backend. * * The SyncEngine’s methods should only ever be called by methods running in the * global async queue. * * PORTING NOTE: On Web, SyncEngine does not have an explicit subscribe() * function. Instead, it directly depends on EventManager's tree-shakeable API * (via `ensureWatchStream()`). */ declare interface SyncEngine { isPrimaryClient: boolean; } /** * A Target represents the WatchTarget representation of a Query, which is used * by the LocalStore and the RemoteStore to keep track of and to execute * backend queries. While a Query can represent multiple Targets, each Targets * maps to a single WatchTarget in RemoteStore and a single TargetData entry * in persistence. */ declare interface Target { readonly path: ResourcePath; readonly collectionGroup: string | null; readonly orderBy: OrderBy[]; readonly filters: Filter[]; readonly limit: number | null; readonly startAt: Bound | null; readonly endAt: Bound | null; } /** * Represents cached targets received from the remote backend. * * The cache is keyed by `Target` and entries in the cache are `TargetData` * instances. */ declare interface TargetCache { /** * A global snapshot version representing the last consistent snapshot we * received from the backend. This is monotonically increasing and any * snapshots received from the backend prior to this version (e.g. for targets * resumed with a resume_token) should be suppressed (buffered) until the * backend has caught up to this snapshot version again. This prevents our * cache from ever going backwards in time. * * This is updated whenever our we get a TargetChange with a read_time and * empty target_ids. */ getLastRemoteSnapshotVersion(transaction: PersistenceTransaction): PersistencePromise; /** * @returns The highest sequence number observed, including any that might be * persisted on-disk. */ getHighestSequenceNumber(transaction: PersistenceTransaction): PersistencePromise; /** * Call provided function with each `TargetData` that we have cached. */ forEachTarget(txn: PersistenceTransaction, f: (q: TargetData) => void): PersistencePromise; /** * Set the highest listen sequence number and optionally updates the * snapshot version of the last consistent snapshot received from the backend * (see getLastRemoteSnapshotVersion() for more details). * * @param highestListenSequenceNumber - The new maximum listen sequence number. * @param lastRemoteSnapshotVersion - The new snapshot version. Optional. */ setTargetsMetadata(transaction: PersistenceTransaction, highestListenSequenceNumber: number, lastRemoteSnapshotVersion?: SnapshotVersion): PersistencePromise; /** * Adds an entry in the cache. * * The cache key is extracted from `targetData.target`. The key must not already * exist in the cache. * * @param targetData - A TargetData instance to put in the cache. */ addTargetData(transaction: PersistenceTransaction, targetData: TargetData): PersistencePromise; /** * Updates an entry in the cache. * * The cache key is extracted from `targetData.target`. The entry must already * exist in the cache, and it will be replaced. * @param targetData - The TargetData to be replaced into the cache. */ updateTargetData(transaction: PersistenceTransaction, targetData: TargetData): PersistencePromise; /** * Removes the cached entry for the given target data. It is an error to remove * a target data that does not exist. * * Multi-Tab Note: This operation should only be called by the primary client. */ removeTargetData(transaction: PersistenceTransaction, targetData: TargetData): PersistencePromise; /** * The number of targets currently in the cache. */ getTargetCount(transaction: PersistenceTransaction): PersistencePromise; /** * Looks up a TargetData entry by target. * * @param target - The query target corresponding to the entry to look up. * @returns The cached TargetData entry, or null if the cache has no entry for * the target. */ getTargetData(transaction: PersistenceTransaction, target: Target): PersistencePromise; /** * Adds the given document keys to cached query results of the given target * ID. * * Multi-Tab Note: This operation should only be called by the primary client. */ addMatchingKeys(transaction: PersistenceTransaction, keys: DocumentKeySet, targetId: TargetId): PersistencePromise; /** * Removes the given document keys from the cached query results of the * given target ID. * * Multi-Tab Note: This operation should only be called by the primary client. */ removeMatchingKeys(transaction: PersistenceTransaction, keys: DocumentKeySet, targetId: TargetId): PersistencePromise; /** * Removes all the keys in the query results of the given target ID. * * Multi-Tab Note: This operation should only be called by the primary client. */ removeMatchingKeysForTargetId(transaction: PersistenceTransaction, targetId: TargetId): PersistencePromise; /** * Returns the document keys that match the provided target ID. */ getMatchingKeysForTargetId(transaction: PersistenceTransaction, targetId: TargetId): PersistencePromise; /** * Returns a new target ID that is higher than any query in the cache. If * there are no queries in the cache, returns the first valid target ID. * Allocated target IDs are persisted and `allocateTargetId()` will never * return the same ID twice. */ allocateTargetId(transaction: PersistenceTransaction): PersistencePromise; containsKey(transaction: PersistenceTransaction, key: DocumentKey): PersistencePromise; } /** * A TargetChange specifies the set of changes for a specific target as part of * a RemoteEvent. These changes track which documents are added, modified or * removed, as well as the target's resume token and whether the target is * marked CURRENT. * The actual changes *to* documents are not part of the TargetChange since * documents may be part of multiple targets. */ declare class TargetChange { /** * An opaque, server-assigned token that allows watching a query to be resumed * after disconnecting without retransmitting all the data that matches the * query. The resume token essentially identifies a point in time from which * the server should resume sending results. */ readonly resumeToken: ByteString; /** * The "current" (synced) status of this target. Note that "current" * has special meaning in the RPC protocol that implies that a target is * both up-to-date and consistent with the rest of the watch stream. */ readonly current: boolean; /** * The set of documents that were newly assigned to this target as part of * this remote event. */ readonly addedDocuments: DocumentKeySet; /** * The set of documents that were already assigned to this target but received * an update during this remote event. */ readonly modifiedDocuments: DocumentKeySet; /** * The set of documents that were removed from this target as part of this * remote event. */ readonly removedDocuments: DocumentKeySet; constructor( /** * An opaque, server-assigned token that allows watching a query to be resumed * after disconnecting without retransmitting all the data that matches the * query. The resume token essentially identifies a point in time from which * the server should resume sending results. */ resumeToken: ByteString, /** * The "current" (synced) status of this target. Note that "current" * has special meaning in the RPC protocol that implies that a target is * both up-to-date and consistent with the rest of the watch stream. */ current: boolean, /** * The set of documents that were newly assigned to this target as part of * this remote event. */ addedDocuments: DocumentKeySet, /** * The set of documents that were already assigned to this target but received * an update during this remote event. */ modifiedDocuments: DocumentKeySet, /** * The set of documents that were removed from this target as part of this * remote event. */ removedDocuments: DocumentKeySet); /** * This method is used to create a synthesized TargetChanges that can be used to * apply a CURRENT status change to a View (for queries executed in a different * tab) or for new queries (to raise snapshots with correct CURRENT status). */ static createSynthesizedTargetChangeForCurrentChange(targetId: TargetId, current: boolean, resumeToken: ByteString): TargetChange; } declare type TargetChangeTargetChangeType = 'NO_CHANGE' | 'ADD' | 'REMOVE' | 'CURRENT' | 'RESET'; /** * An immutable set of metadata that the local store tracks for each target. */ declare class TargetData { /** The target being listened to. */ readonly target: Target; /** * The target ID to which the target corresponds; Assigned by the * LocalStore for user listens and by the SyncEngine for limbo watches. */ readonly targetId: TargetId; /** The purpose of the target. */ readonly purpose: TargetPurpose; /** * The sequence number of the last transaction during which this target data * was modified. */ readonly sequenceNumber: ListenSequenceNumber; /** The latest snapshot version seen for this target. */ readonly snapshotVersion: SnapshotVersion; /** * The maximum snapshot version at which the associated view * contained no limbo documents. */ readonly lastLimboFreeSnapshotVersion: SnapshotVersion; /** * An opaque, server-assigned token that allows watching a target to be * resumed after disconnecting without retransmitting all the data that * matches the target. The resume token essentially identifies a point in * time from which the server should resume sending results. */ readonly resumeToken: ByteString; /** * The number of documents that last matched the query at the resume token or * read time. Documents are counted only when making a listen request with * resume token or read time, otherwise, keep it null. */ readonly expectedCount: number | null; constructor( /** The target being listened to. */ target: Target, /** * The target ID to which the target corresponds; Assigned by the * LocalStore for user listens and by the SyncEngine for limbo watches. */ targetId: TargetId, /** The purpose of the target. */ purpose: TargetPurpose, /** * The sequence number of the last transaction during which this target data * was modified. */ sequenceNumber: ListenSequenceNumber, /** The latest snapshot version seen for this target. */ snapshotVersion?: SnapshotVersion, /** * The maximum snapshot version at which the associated view * contained no limbo documents. */ lastLimboFreeSnapshotVersion?: SnapshotVersion, /** * An opaque, server-assigned token that allows watching a target to be * resumed after disconnecting without retransmitting all the data that * matches the target. The resume token essentially identifies a point in * time from which the server should resume sending results. */ resumeToken?: ByteString, /** * The number of documents that last matched the query at the resume token or * read time. Documents are counted only when making a listen request with * resume token or read time, otherwise, keep it null. */ expectedCount?: number | null); /** Creates a new target data instance with an updated sequence number. */ withSequenceNumber(sequenceNumber: number): TargetData; /** * Creates a new target data instance with an updated resume token and * snapshot version. */ withResumeToken(resumeToken: ByteString, snapshotVersion: SnapshotVersion): TargetData; /** * Creates a new target data instance with an updated expected count. */ withExpectedCount(expectedCount: number): TargetData; /** * Creates a new target data instance with an updated last limbo free * snapshot version number. */ withLastLimboFreeSnapshotVersion(lastLimboFreeSnapshotVersion: SnapshotVersion): TargetData; } /** * A locally-assigned ID used to refer to a target being watched via the * Watch service. */ declare type TargetId = number; /** An enumeration of the different purposes we have for targets. */ declare const enum TargetPurpose { /** A regular, normal query target. */ Listen = "TargetPurposeListen", /** * The query target was used to refill a query after an existence filter * mismatch. */ ExistenceFilterMismatch = "TargetPurposeExistenceFilterMismatch", /** * The query target was used if the query is the result of a false positive in * the bloom filter. */ ExistenceFilterMismatchBloom = "TargetPurposeExistenceFilterMismatchBloom", /** The query target was used to resolve a limbo document. */ LimboResolution = "TargetPurposeLimboResolution" } /** * Terminates the provided `Firestore` instance. * * After calling `terminate()` only the `clearIndexedDbPersistence()` functions * may be used. Any other function will throw a `FirestoreError`. Termination * does not cancel any pending writes, and any promises that are awaiting a * response from the server will not be resolved. * * To restart after termination, create a new instance of `Firestore` with * {@link (getFirestore:1)}. * * Note: Under normal circumstances, calling `terminate()` is not required. This * function is useful only when you want to force this instance to release all of * its resources or in combination with {@link clearIndexedDbPersistence} to * ensure that all local state is destroyed between test runs. * * @param firestore - The `Firestore` instance to terminate. * @returns A `Promise` that is resolved when the instance has been successfully * terminated. */ export declare function terminate(firestore: Firestore): Promise; /** * Wellknown "timer" IDs used when scheduling delayed operations on the * AsyncQueue. These IDs can then be used from tests to check for the presence * of operations or to run them early. * * The string values are used when encoding these timer IDs in JSON spec tests. */ declare const enum TimerId { /** All can be used with runDelayedOperationsEarly() to run all timers. */ All = "all", /** * The following 5 timers are used in persistent_stream.ts for the listen and * write streams. The "Idle" timer is used to close the stream due to * inactivity. The "ConnectionBackoff" timer is used to restart a stream once * the appropriate backoff delay has elapsed. The health check is used to mark * a stream healthy if it has not received an error during its initial setup. */ ListenStreamIdle = "listen_stream_idle", ListenStreamConnectionBackoff = "listen_stream_connection_backoff", WriteStreamIdle = "write_stream_idle", WriteStreamConnectionBackoff = "write_stream_connection_backoff", HealthCheckTimeout = "health_check_timeout", /** * A timer used in online_state_tracker.ts to transition from * OnlineState.Unknown to Offline after a set timeout, rather than waiting * indefinitely for success or failure. */ OnlineStateTimeout = "online_state_timeout", /** * A timer used to update the client metadata in IndexedDb, which is used * to determine the primary leaseholder. */ ClientMetadataRefresh = "client_metadata_refresh", /** A timer used to periodically attempt LRU Garbage collection */ LruGarbageCollection = "lru_garbage_collection", /** * A timer used to retry transactions. Since there can be multiple concurrent * transactions, multiple of these may be in the queue at a given time. */ TransactionRetry = "transaction_retry", /** * A timer used to retry operations scheduled via retryable AsyncQueue * operations. */ AsyncQueueRetry = "async_queue_retry", /** * A timer used to periodically attempt index backfill. */ IndexBackfill = "index_backfill" } /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * A `Timestamp` represents a point in time independent of any time zone or * calendar, represented as seconds and fractions of seconds at nanosecond * resolution in UTC Epoch time. * * It is encoded using the Proleptic Gregorian Calendar which extends the * Gregorian calendar backwards to year one. It is encoded assuming all minutes * are 60 seconds long, i.e. leap seconds are "smeared" so that no leap second * table is needed for interpretation. Range is from 0001-01-01T00:00:00Z to * 9999-12-31T23:59:59.999999999Z. * * For examples and further specifications, refer to the * {@link https://github.com/google/protobuf/blob/master/src/google/protobuf/timestamp.proto | Timestamp definition}. */ export declare class Timestamp { /** * The number of seconds of UTC time since Unix epoch 1970-01-01T00:00:00Z. */ readonly seconds: number; /** * The fractions of a second at nanosecond resolution.* */ readonly nanoseconds: number; /** * Creates a new timestamp with the current date, with millisecond precision. * * @returns a new timestamp representing the current date. */ static now(): Timestamp; /** * Creates a new timestamp from the given date. * * @param date - The date to initialize the `Timestamp` from. * @returns A new `Timestamp` representing the same point in time as the given * date. */ static fromDate(date: Date): Timestamp; /** * Creates a new timestamp from the given number of milliseconds. * * @param milliseconds - Number of milliseconds since Unix epoch * 1970-01-01T00:00:00Z. * @returns A new `Timestamp` representing the same point in time as the given * number of milliseconds. */ static fromMillis(milliseconds: number): Timestamp; /** * Creates a new timestamp. * * @param seconds - The number of seconds of UTC time since Unix epoch * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to * 9999-12-31T23:59:59Z inclusive. * @param nanoseconds - The non-negative fractions of a second at nanosecond * resolution. Negative second values with fractions must still have * non-negative nanoseconds values that count forward in time. Must be * from 0 to 999,999,999 inclusive. */ constructor( /** * The number of seconds of UTC time since Unix epoch 1970-01-01T00:00:00Z. */ seconds: number, /** * The fractions of a second at nanosecond resolution.* */ nanoseconds: number); /** * Converts a `Timestamp` to a JavaScript `Date` object. This conversion * causes a loss of precision since `Date` objects only support millisecond * precision. * * @returns JavaScript `Date` object representing the same point in time as * this `Timestamp`, with millisecond precision. */ toDate(): Date; /** * Converts a `Timestamp` to a numeric timestamp (in milliseconds since * epoch). This operation causes a loss of precision. * * @returns The point in time corresponding to this timestamp, represented as * the number of milliseconds since Unix epoch 1970-01-01T00:00:00Z. */ toMillis(): number; _compareTo(other: Timestamp): number; /** * Returns true if this `Timestamp` is equal to the provided one. * * @param other - The `Timestamp` to compare against. * @returns true if this `Timestamp` is equal to the provided one. */ isEqual(other: Timestamp): boolean; /** Returns a textual representation of this `Timestamp`. */ toString(): string; /** Returns a JSON-serializable representation of this `Timestamp`. */ toJSON(): { seconds: number; nanoseconds: number; }; /** * Converts this object to a primitive string, which allows `Timestamp` objects * to be compared using the `>`, `<=`, `>=` and `>` operators. */ valueOf(): string; } declare type Timestamp_2 = string | { seconds?: string | number; nanos?: number; }; declare interface Token { /** Type of token. */ type: TokenType; /** * The user with which the token is associated (used for persisting user * state on disk, etc.). * This will be null for Tokens of the type 'AppCheck'. */ user?: User; /** Header values to set for this token */ headers: Map; } declare type TokenType = 'OAuth' | 'FirstParty' | 'AppCheck'; /** * A reference to a transaction. * * The `Transaction` object passed to a transaction's `updateFunction` provides * the methods to read and write data within the transaction context. See * {@link runTransaction}. */ export declare class Transaction { protected readonly _firestore: Firestore; private readonly _transaction; private readonly _dataReader; /** @hideconstructor */ constructor(_firestore: Firestore, _transaction: Transaction_2); /** * Reads the document referenced by the provided {@link DocumentReference}. * * @param documentRef - A reference to the document to be read. * @returns A `DocumentSnapshot` with the read data. */ get(documentRef: DocumentReference): Promise>; /** * Writes to the document referred to by the provided {@link * DocumentReference}. If the document does not exist yet, it will be created. * * @param documentRef - A reference to the document to be set. * @param data - An object of the fields and values for the document. * @throws Error - If the provided input is not a valid Firestore document. * @returns This `Transaction` instance. Used for chaining method calls. */ set(documentRef: DocumentReference, data: WithFieldValue): this; /** * Writes to the document referred to by the provided {@link * DocumentReference}. If the document does not exist yet, it will be created. * If you provide `merge` or `mergeFields`, the provided data can be merged * into an existing document. * * @param documentRef - A reference to the document to be set. * @param data - An object of the fields and values for the document. * @param options - An object to configure the set behavior. * @throws Error - If the provided input is not a valid Firestore document. * @returns This `Transaction` instance. Used for chaining method calls. */ set(documentRef: DocumentReference, data: PartialWithFieldValue, options: SetOptions): this; /** * Updates fields in the document referred to by the provided {@link * DocumentReference}. The update will fail if applied to a document that does * not exist. * * @param documentRef - A reference to the document to be updated. * @param data - An object containing the fields and values with which to * update the document. Fields can contain dots to reference nested fields * within the document. * @throws Error - If the provided input is not valid Firestore data. * @returns This `Transaction` instance. Used for chaining method calls. */ update(documentRef: DocumentReference, data: UpdateData): this; /** * Updates fields in the document referred to by the provided {@link * DocumentReference}. The update will fail if applied to a document that does * not exist. * * Nested fields can be updated by providing dot-separated field path * strings or by providing `FieldPath` objects. * * @param documentRef - A reference to the document to be updated. * @param field - The first field to update. * @param value - The first value. * @param moreFieldsAndValues - Additional key/value pairs. * @throws Error - If the provided input is not valid Firestore data. * @returns This `Transaction` instance. Used for chaining method calls. */ update(documentRef: DocumentReference, field: string | FieldPath, value: unknown, ...moreFieldsAndValues: unknown[]): this; /** * Deletes the document referred to by the provided {@link DocumentReference}. * * @param documentRef - A reference to the document to be deleted. * @returns This `Transaction` instance. Used for chaining method calls. */ delete(documentRef: DocumentReference): this; } /** * Internal transaction object responsible for accumulating the mutations to * perform and the base versions for any documents read. */ declare class Transaction_2 { private datastore; private readVersions; private mutations; private committed; /** * A deferred usage error that occurred previously in this transaction that * will cause the transaction to fail once it actually commits. */ private lastTransactionError; /** * Set of documents that have been written in the transaction. * * When there's more than one write to the same key in a transaction, any * writes after the first are handled differently. */ private writtenDocs; constructor(datastore: Datastore); lookup(keys: DocumentKey[]): Promise; set(key: DocumentKey, data: ParsedSetData): void; update(key: DocumentKey, data: ParsedUpdateData): void; delete(key: DocumentKey): void; commit(): Promise; private recordVersion; /** * Returns the version of this document when it was read in this transaction, * as a precondition, or no precondition if it was not read. */ private precondition; /** * Returns the precondition for a document if the operation is an update. */ private preconditionForUpdate; private write; private ensureCommitNotCalled; } /** * @license * Copyright 2022 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * Options to customize transaction behavior. */ export declare interface TransactionOptions { /** Maximum number of attempts to commit, after which transaction fails. Default is 5. */ readonly maxAttempts?: number; } /** Used to represent a field transform on a mutation. */ declare class TransformOperation { private _; } declare type UnaryFilterOp = 'OPERATOR_UNSPECIFIED' | 'IS_NAN' | 'IS_NULL' | 'IS_NOT_NAN' | 'IS_NOT_NULL'; /** * Given a union type `U = T1 | T2 | ...`, returns an intersected type * `(T1 & T2 & ...)`. * * Uses distributive conditional types and inference from conditional types. * This works because multiple candidates for the same type variable in * contra-variant positions causes an intersection type to be inferred. * https://www.typescriptlang.org/docs/handbook/advanced-types.html#type-inference-in-conditional-types * https://stackoverflow.com/questions/50374908/transform-union-type-to-intersection-type */ export declare type UnionToIntersection = (U extends unknown ? (k: U) => void : never) extends (k: infer I) => void ? I : never; /** * An untyped Firestore Data Converter interface that is shared between the * lite, firestore-exp and classic SDK. */ declare interface UntypedFirestoreDataConverter { toFirestore(modelObject: WithFieldValue): WithFieldValue; toFirestore(modelObject: PartialWithFieldValue, options: SetOptions_2): PartialWithFieldValue; fromFirestore(snapshot: unknown, options?: unknown): AppModelType; } /** * Update data (for use with {@link (updateDoc:1)}) that consists of field paths * (e.g. 'foo' or 'foo.baz') mapped to values. Fields that contain dots * reference nested fields within the document. FieldValues can be passed in * as property values. */ export declare type UpdateData = T extends Primitive ? T : T extends {} ? { [K in keyof T]?: UpdateData | FieldValue; } & NestedUpdateFields : Partial; /** * Updates fields in the document referred to by the specified * `DocumentReference`. The update will fail if applied to a document that does * not exist. * * The result of this update will only be reflected in document reads that occur * after the returned promise resolves. If the client is offline, the * update fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @param reference - A reference to the document to update. * @param data - An object containing the fields and values with which to * update the document. Fields can contain dots to reference nested fields * within the document. * @throws Error - If the provided input is not valid Firestore data. * @returns A `Promise` resolved once the data has been successfully written * to the backend. */ export declare function updateDoc(reference: DocumentReference, data: UpdateData): Promise; /** * Updates fields in the document referred to by the specified * `DocumentReference` The update will fail if applied to a document that does * not exist. * * Nested fields can be updated by providing dot-separated field path * strings or by providing `FieldPath` objects. * * The result of this update will only be reflected in document reads that occur * after the returned promise resolves. If the client is offline, the * update fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @param reference - A reference to the document to update. * @param field - The first field to update. * @param value - The first value. * @param moreFieldsAndValues - Additional key value pairs. * @throws Error - If the provided input is not valid Firestore data. * @returns A `Promise` resolved once the data has been successfully written * to the backend. */ export declare function updateDoc(reference: DocumentReference, field: string | FieldPath, value: unknown, ...moreFieldsAndValues: unknown[]): Promise; /** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * Simple wrapper around a nullable UID. Mostly exists to make code more * readable. */ declare class User { readonly uid: string | null; /** A user with a null UID. */ static readonly UNAUTHENTICATED: User; static readonly GOOGLE_CREDENTIALS: User; static readonly FIRST_PARTY: User; static readonly MOCK_USER: User; constructor(uid: string | null); isAuthenticated(): boolean; /** * Returns a key representing this user, suitable for inclusion in a * dictionary. */ toKey(): string; isEqual(otherUser: User): boolean; } declare type Value = firestoreV1ApiClientInterfaces.Value; declare type ValueNullValue = 'NULL_VALUE'; declare class ViewSnapshot { readonly query: Query_2; readonly docs: DocumentSet; readonly oldDocs: DocumentSet; readonly docChanges: DocumentViewChange[]; readonly mutatedKeys: DocumentKeySet; readonly fromCache: boolean; readonly syncStateChanged: boolean; readonly excludesMetadataChanges: boolean; readonly hasCachedResults: boolean; constructor(query: Query_2, docs: DocumentSet, oldDocs: DocumentSet, docChanges: DocumentViewChange[], mutatedKeys: DocumentKeySet, fromCache: boolean, syncStateChanged: boolean, excludesMetadataChanges: boolean, hasCachedResults: boolean); /** Returns a view snapshot as if all documents in the snapshot were added. */ static fromInitialDocuments(query: Query_2, documents: DocumentSet, mutatedKeys: DocumentKeySet, fromCache: boolean, hasCachedResults: boolean): ViewSnapshot; get hasPendingWrites(): boolean; isEqual(other: ViewSnapshot): boolean; } /** * Creates a {@link QueryFieldFilterConstraint} that enforces that documents * must contain the specified field and that the value should satisfy the * relation constraint provided. * * @param fieldPath - The path to compare * @param opStr - The operation string (e.g "<", "<=", "==", "<", * "<=", "!="). * @param value - The value for comparison * @returns The created {@link QueryFieldFilterConstraint}. */ export declare function where(fieldPath: string | FieldPath, opStr: WhereFilterOp, value: unknown): QueryFieldFilterConstraint; /** * Filter conditions in a {@link where} clause are specified using the * strings '<', '<=', '==', '!=', '>=', '>', 'array-contains', 'in', * 'array-contains-any', and 'not-in'. */ export declare type WhereFilterOp = '<' | '<=' | '==' | '!=' | '>=' | '>' | 'array-contains' | 'in' | 'array-contains-any' | 'not-in'; /** * Allows FieldValues to be passed in as a property value while maintaining * type safety. */ export declare type WithFieldValue = T | (T extends Primitive ? T : T extends {} ? { [K in keyof T]: WithFieldValue | FieldValue; } : never); /** * A write batch, used to perform multiple writes as a single atomic unit. * * A `WriteBatch` object can be acquired by calling {@link writeBatch}. It * provides methods for adding writes to the write batch. None of the writes * will be committed (or visible locally) until {@link WriteBatch.commit} is * called. */ export declare class WriteBatch { private readonly _firestore; private readonly _commitHandler; private readonly _dataReader; private _mutations; private _committed; /** @hideconstructor */ constructor(_firestore: Firestore, _commitHandler: (m: Mutation[]) => Promise); /** * Writes to the document referred to by the provided {@link * DocumentReference}. If the document does not exist yet, it will be created. * * @param documentRef - A reference to the document to be set. * @param data - An object of the fields and values for the document. * @returns This `WriteBatch` instance. Used for chaining method calls. */ set(documentRef: DocumentReference, data: WithFieldValue): WriteBatch; /** * Writes to the document referred to by the provided {@link * DocumentReference}. If the document does not exist yet, it will be created. * If you provide `merge` or `mergeFields`, the provided data can be merged * into an existing document. * * @param documentRef - A reference to the document to be set. * @param data - An object of the fields and values for the document. * @param options - An object to configure the set behavior. * @throws Error - If the provided input is not a valid Firestore document. * @returns This `WriteBatch` instance. Used for chaining method calls. */ set(documentRef: DocumentReference, data: PartialWithFieldValue, options: SetOptions): WriteBatch; /** * Updates fields in the document referred to by the provided {@link * DocumentReference}. The update will fail if applied to a document that does * not exist. * * @param documentRef - A reference to the document to be updated. * @param data - An object containing the fields and values with which to * update the document. Fields can contain dots to reference nested fields * within the document. * @throws Error - If the provided input is not valid Firestore data. * @returns This `WriteBatch` instance. Used for chaining method calls. */ update(documentRef: DocumentReference, data: UpdateData): WriteBatch; /** * Updates fields in the document referred to by this {@link * DocumentReference}. The update will fail if applied to a document that does * not exist. * * Nested fields can be update by providing dot-separated field path strings * or by providing `FieldPath` objects. * * @param documentRef - A reference to the document to be updated. * @param field - The first field to update. * @param value - The first value. * @param moreFieldsAndValues - Additional key value pairs. * @throws Error - If the provided input is not valid Firestore data. * @returns This `WriteBatch` instance. Used for chaining method calls. */ update(documentRef: DocumentReference, field: string | FieldPath, value: unknown, ...moreFieldsAndValues: unknown[]): WriteBatch; /** * Deletes the document referred to by the provided {@link DocumentReference}. * * @param documentRef - A reference to the document to be deleted. * @returns This `WriteBatch` instance. Used for chaining method calls. */ delete(documentRef: DocumentReference): WriteBatch; /** * Commits all of the writes in this write batch as a single atomic unit. * * The result of these writes will only be reflected in document reads that * occur after the returned promise resolves. If the client is offline, the * write fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @returns A `Promise` resolved once all of the writes in the batch have been * successfully written to the backend as an atomic unit (note that it won't * resolve while you're offline). */ commit(): Promise; private _verifyNotCommitted; } /** * Creates a write batch, used for performing multiple writes as a single * atomic operation. The maximum number of writes allowed in a single WriteBatch * is 500. * * The result of these writes will only be reflected in document reads that * occur after the returned promise resolves. If the client is offline, the * write fails. If you would like to see local modifications or buffer writes * until the client is online, use the full Firestore SDK. * * @returns A `WriteBatch` that can be used to atomically execute multiple * writes. */ export declare function writeBatch(firestore: Firestore): WriteBatch; export { }