1 line
1.1 MiB
1 line
1.1 MiB
{"version":3,"file":"index.memory.esm2017.js","sources":["../src/core/version.ts","../src/auth/user.ts","../src/util/error.ts","../src/api/credentials.ts","../src/api/timestamp.ts","../src/core/snapshot_version.ts","../src/model/path.ts","../src/model/document_key.ts","../src/util/obj.ts","../src/util/byte_string.ts","../src/util/types.ts","../src/model/server_timestamps.ts","../src/model/values.ts","../src/model/transform_operation.ts","../src/model/mutation.ts","../src/model/object_value.ts","../src/model/document.ts","../src/core/target.ts","../src/core/query.ts","../src/local/target_data.ts","../src/remote/existence_filter.ts","../src/remote/rpc_error.ts","../src/util/sorted_map.ts","../src/util/sorted_set.ts","../src/model/collections.ts","../src/model/document_set.ts","../src/core/view_snapshot.ts","../src/remote/remote_event.ts","../src/remote/watch_change.ts","../src/remote/serializer.ts","../src/platform/platform.ts","../src/util/log.ts","../src/util/assert.ts","../src/util/misc.ts","../src/core/database_info.ts","../src/util/obj_map.ts","../src/model/mutation_batch.ts","../src/local/persistence_promise.ts","../src/local/local_documents_view.ts","../src/local/local_view_changes.ts","../src/core/listen_sequence.ts","../src/util/promise.ts","../src/remote/backoff.ts","../src/local/memory_index_manager.ts","../src/core/target_id_generator.ts","../src/local/simple_db.ts","../src/util/async_queue.ts","../src/local/lru_garbage_collector.ts","../src/local/local_store.ts","../src/local/persistence.ts","../src/local/reference_set.ts","../src/util/input_validation.ts","../src/api/blob.ts","../src/api/field_path.ts","../src/api/field_value.ts","../src/api/geo_point.ts","../src/api/user_data_reader.ts","../src/remote/persistent_stream.ts","../src/remote/datastore.ts","../src/core/transaction.ts","../src/remote/online_state_tracker.ts","../src/remote/remote_store.ts","../src/local/shared_client_state.ts","../src/core/view.ts","../src/core/transaction_runner.ts","../src/core/sync_engine.ts","../src/core/event_manager.ts","../src/local/index_free_query_engine.ts","../src/local/memory_mutation_queue.ts","../src/local/memory_remote_document_cache.ts","../src/local/remote_document_change_buffer.ts","../src/local/memory_target_cache.ts","../src/local/memory_persistence.ts","../src/core/component_provider.ts","../src/core/firestore_client.ts","../src/util/async_observer.ts","../src/api/observer.ts","../src/api/user_data_writer.ts","../src/api/database.ts","../src/util/api.ts","../src/platform/config.ts","../src/remote/connectivity_monitor_noop.ts","../src/platform_browser/browser_connectivity_monitor.ts","../src/remote/stream_bridge.ts","../src/platform_browser/webchannel_connection.ts","../src/platform_browser/browser_init.ts","../src/platform_browser/browser_platform.ts","../index.memory.ts"],"sourcesContent":["/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport firebase from '@firebase/app';\n\n/** The semver (www.semver.org) version of the SDK. */\nexport const SDK_VERSION = firebase.SDK_VERSION;\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\n/**\n * Simple wrapper around a nullable UID. Mostly exists to make code more\n * readable.\n */\nexport class User {\n /** A user with a null UID. */\n static readonly UNAUTHENTICATED = new User(null);\n\n // TODO(mikelehen): Look into getting a proper uid-equivalent for\n // non-FirebaseAuth providers.\n static readonly GOOGLE_CREDENTIALS = new User('google-credentials-uid');\n static readonly FIRST_PARTY = new User('first-party-uid');\n\n constructor(readonly uid: string | null) {}\n\n isAuthenticated(): boolean {\n return this.uid != null;\n }\n\n /**\n * Returns a key representing this user, suitable for inclusion in a\n * dictionary.\n */\n toKey(): string {\n if (this.isAuthenticated()) {\n return 'uid:' + this.uid;\n } else {\n return 'anonymous-user';\n }\n }\n\n isEqual(otherUser: User): boolean {\n return otherUser.uid === this.uid;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as firestore from '@firebase/firestore-types';\n\n/**\n * Error Codes describing the different ways Firestore can fail. These come\n * directly from GRPC.\n */\nexport type Code = firestore.FirestoreErrorCode;\n\nexport const Code = {\n // Causes are copied from:\n // https://github.com/grpc/grpc/blob/bceec94ea4fc5f0085d81235d8e1c06798dc341a/include/grpc%2B%2B/impl/codegen/status_code_enum.h\n /** Not an error; returned on success. */\n OK: 'ok' as Code,\n\n /** The operation was cancelled (typically by the caller). */\n CANCELLED: 'cancelled' as Code,\n\n /** Unknown error or an error from a different error domain. */\n UNKNOWN: 'unknown' as Code,\n\n /**\n * Client specified an invalid argument. Note that this differs from\n * FAILED_PRECONDITION. INVALID_ARGUMENT indicates arguments that are\n * problematic regardless of the state of the system (e.g., a malformed file\n * name).\n */\n INVALID_ARGUMENT: 'invalid-argument' as Code,\n\n /**\n * Deadline expired before operation could complete. For operations that\n * change the state of the system, this error may be returned even if the\n * operation has completed successfully. For example, a successful response\n * from a server could have been delayed long enough for the deadline to\n * expire.\n */\n DEADLINE_EXCEEDED: 'deadline-exceeded' as Code,\n\n /** Some requested entity (e.g., file or directory) was not found. */\n NOT_FOUND: 'not-found' as Code,\n\n /**\n * Some entity that we attempted to create (e.g., file or directory) already\n * exists.\n */\n ALREADY_EXISTS: 'already-exists' as Code,\n\n /**\n * The caller does not have permission to execute the specified operation.\n * PERMISSION_DENIED must not be used for rejections caused by exhausting\n * some resource (use RESOURCE_EXHAUSTED instead for those errors).\n * PERMISSION_DENIED must not be used if the caller can not be identified\n * (use UNAUTHENTICATED instead for those errors).\n */\n PERMISSION_DENIED: 'permission-denied' as Code,\n\n /**\n * The request does not have valid authentication credentials for the\n * operation.\n */\n UNAUTHENTICATED: 'unauthenticated' as Code,\n\n /**\n * Some resource has been exhausted, perhaps a per-user quota, or perhaps the\n * entire file system is out of space.\n */\n RESOURCE_EXHAUSTED: 'resource-exhausted' as Code,\n\n /**\n * Operation was rejected because the system is not in a state required for\n * the operation's execution. For example, directory to be deleted may be\n * non-empty, an rmdir operation is applied to a non-directory, etc.\n *\n * A litmus test that may help a service implementor in deciding\n * between FAILED_PRECONDITION, ABORTED, and UNAVAILABLE:\n * (a) Use UNAVAILABLE if the client can retry just the failing call.\n * (b) Use ABORTED if the client should retry at a higher-level\n * (e.g., restarting a read-modify-write sequence).\n * (c) Use FAILED_PRECONDITION if the client should not retry until\n * the system state has been explicitly fixed. E.g., if an \"rmdir\"\n * fails because the directory is non-empty, FAILED_PRECONDITION\n * should be returned since the client should not retry unless\n * they have first fixed up the directory by deleting files from it.\n * (d) Use FAILED_PRECONDITION if the client performs conditional\n * REST Get/Update/Delete on a resource and the resource on the\n * server does not match the condition. E.g., conflicting\n * read-modify-write on the same resource.\n */\n FAILED_PRECONDITION: 'failed-precondition' as Code,\n\n /**\n * The operation was aborted, typically due to a concurrency issue like\n * sequencer check failures, transaction aborts, etc.\n *\n * See litmus test above for deciding between FAILED_PRECONDITION, ABORTED,\n * and UNAVAILABLE.\n */\n ABORTED: 'aborted' as Code,\n\n /**\n * Operation was attempted past the valid range. E.g., seeking or reading\n * past end of file.\n *\n * Unlike INVALID_ARGUMENT, this error indicates a problem that may be fixed\n * if the system state changes. For example, a 32-bit file system will\n * generate INVALID_ARGUMENT if asked to read at an offset that is not in the\n * range [0,2^32-1], but it will generate OUT_OF_RANGE if asked to read from\n * an offset past the current file size.\n *\n * There is a fair bit of overlap between FAILED_PRECONDITION and\n * OUT_OF_RANGE. We recommend using OUT_OF_RANGE (the more specific error)\n * when it applies so that callers who are iterating through a space can\n * easily look for an OUT_OF_RANGE error to detect when they are done.\n */\n OUT_OF_RANGE: 'out-of-range' as Code,\n\n /** Operation is not implemented or not supported/enabled in this service. */\n UNIMPLEMENTED: 'unimplemented' as Code,\n\n /**\n * Internal errors. Means some invariants expected by underlying System has\n * been broken. If you see one of these errors, Something is very broken.\n */\n INTERNAL: 'internal' as Code,\n\n /**\n * The service is currently unavailable. This is a most likely a transient\n * condition and may be corrected by retrying with a backoff.\n *\n * See litmus test above for deciding between FAILED_PRECONDITION, ABORTED,\n * and UNAVAILABLE.\n */\n UNAVAILABLE: 'unavailable' as Code,\n\n /** Unrecoverable data loss or corruption. */\n DATA_LOSS: 'data-loss' as Code\n};\n\n/**\n * An error class used for Firestore-generated errors. Ideally we should be\n * using FirebaseError, but integrating with it is overly arduous at the moment,\n * so we define our own compatible error class (with a `name` of 'FirebaseError'\n * and compatible `code` and `message` fields.)\n */\nexport class FirestoreError extends Error implements firestore.FirestoreError {\n name = 'FirebaseError';\n stack?: string;\n\n constructor(readonly code: Code, readonly message: string) {\n super(message);\n\n // HACK: We write a toString property directly because Error is not a real\n // class and so inheritance does not work correctly. We could alternatively\n // do the same \"back-door inheritance\" trick that FirebaseError does.\n this.toString = () => `${this.name}: [code=${this.code}]: ${this.message}`;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { User } from '../auth/user';\nimport { hardAssert, debugAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport {\n FirebaseAuthInternal,\n FirebaseAuthInternalName\n} from '@firebase/auth-interop-types';\nimport { Provider } from '@firebase/component';\n\n// TODO(mikelehen): This should be split into multiple files and probably\n// moved to an auth/ folder to match other platforms.\n\nexport interface FirstPartyCredentialsSettings {\n type: 'gapi';\n client: unknown;\n sessionIndex: string;\n}\n\nexport interface ProviderCredentialsSettings {\n type: 'provider';\n client: CredentialsProvider;\n}\n\n/** Settings for private credentials */\nexport type CredentialsSettings =\n | FirstPartyCredentialsSettings\n | ProviderCredentialsSettings;\n\nexport type TokenType = 'OAuth' | 'FirstParty';\nexport interface Token {\n /** Type of token. */\n type: TokenType;\n\n /**\n * The user with which the token is associated (used for persisting user\n * state on disk, etc.).\n */\n user: User;\n\n /** Extra header values to be passed along with a request */\n authHeaders: { [header: string]: string };\n}\n\nexport class OAuthToken implements Token {\n type = 'OAuth' as TokenType;\n authHeaders: { [header: string]: string };\n constructor(value: string, public user: User) {\n this.authHeaders = {};\n // Set the headers using Object Literal notation to avoid minification\n this.authHeaders['Authorization'] = `Bearer ${value}`;\n }\n}\n\n/**\n * A Listener for credential change events. The listener should fetch a new\n * token and may need to invalidate other state if the current user has also\n * changed.\n */\nexport type CredentialChangeListener = (user: User) => void;\n\n/**\n * Provides methods for getting the uid and token for the current user and\n * listening for changes.\n */\nexport interface CredentialsProvider {\n /** Requests a token for the current user. */\n getToken(): Promise<Token | null>;\n\n /**\n * Marks the last retrieved token as invalid, making the next GetToken request\n * force-refresh the token.\n */\n invalidateToken(): void;\n\n /**\n * Specifies a listener to be notified of credential changes\n * (sign-in / sign-out, token changes). It is immediately called once with the\n * initial user.\n */\n setChangeListener(changeListener: CredentialChangeListener): void;\n\n /** Removes the previously-set change listener. */\n removeChangeListener(): void;\n}\n\n/** A CredentialsProvider that always yields an empty token. */\nexport class EmptyCredentialsProvider implements CredentialsProvider {\n /**\n * Stores the listener registered with setChangeListener()\n * This isn't actually necessary since the UID never changes, but we use this\n * to verify the listen contract is adhered to in tests.\n */\n private changeListener: CredentialChangeListener | null = null;\n\n getToken(): Promise<Token | null> {\n return Promise.resolve<Token | null>(null);\n }\n\n invalidateToken(): void {}\n\n setChangeListener(changeListener: CredentialChangeListener): void {\n debugAssert(\n !this.changeListener,\n 'Can only call setChangeListener() once.'\n );\n this.changeListener = changeListener;\n // Fire with initial user.\n changeListener(User.UNAUTHENTICATED);\n }\n\n removeChangeListener(): void {\n debugAssert(\n this.changeListener !== null,\n 'removeChangeListener() when no listener registered'\n );\n this.changeListener = null;\n }\n}\n\nexport class FirebaseCredentialsProvider implements CredentialsProvider {\n /**\n * The auth token listener registered with FirebaseApp, retained here so we\n * can unregister it.\n */\n private tokenListener: ((token: string | null) => void) | null = null;\n\n /** Tracks the current User. */\n private currentUser: User = User.UNAUTHENTICATED;\n private receivedInitialUser: boolean = false;\n\n /**\n * Counter used to detect if the token changed while a getToken request was\n * outstanding.\n */\n private tokenCounter = 0;\n\n /** The listener registered with setChangeListener(). */\n private changeListener: CredentialChangeListener | null = null;\n\n private forceRefresh = false;\n\n private auth: FirebaseAuthInternal | null;\n\n constructor(authProvider: Provider<FirebaseAuthInternalName>) {\n this.tokenListener = () => {\n this.tokenCounter++;\n this.currentUser = this.getUser();\n this.receivedInitialUser = true;\n if (this.changeListener) {\n this.changeListener(this.currentUser);\n }\n };\n\n this.tokenCounter = 0;\n\n this.auth = authProvider.getImmediate({ optional: true });\n\n if (this.auth) {\n this.auth.addAuthTokenListener(this.tokenListener!);\n } else {\n // if auth is not available, invoke tokenListener once with null token\n this.tokenListener(null);\n authProvider.get().then(\n auth => {\n this.auth = auth;\n if (this.tokenListener) {\n // tokenListener can be removed by removeChangeListener()\n this.auth.addAuthTokenListener(this.tokenListener);\n }\n },\n () => {\n /* this.authProvider.get() never rejects */\n }\n );\n }\n }\n\n getToken(): Promise<Token | null> {\n debugAssert(\n this.tokenListener != null,\n 'getToken cannot be called after listener removed.'\n );\n\n // Take note of the current value of the tokenCounter so that this method\n // can fail (with an ABORTED error) if there is a token change while the\n // request is outstanding.\n const initialTokenCounter = this.tokenCounter;\n const forceRefresh = this.forceRefresh;\n this.forceRefresh = false;\n\n if (!this.auth) {\n return Promise.resolve(null);\n }\n\n return this.auth.getToken(forceRefresh).then(tokenData => {\n // Cancel the request since the token changed while the request was\n // outstanding so the response is potentially for a previous user (which\n // user, we can't be sure).\n if (this.tokenCounter !== initialTokenCounter) {\n throw new FirestoreError(\n Code.ABORTED,\n 'getToken aborted due to token change.'\n );\n } else {\n if (tokenData) {\n hardAssert(\n typeof tokenData.accessToken === 'string',\n 'Invalid tokenData returned from getToken():' + tokenData\n );\n return new OAuthToken(tokenData.accessToken, this.currentUser);\n } else {\n return null;\n }\n }\n });\n }\n\n invalidateToken(): void {\n this.forceRefresh = true;\n }\n\n setChangeListener(changeListener: CredentialChangeListener): void {\n debugAssert(\n !this.changeListener,\n 'Can only call setChangeListener() once.'\n );\n this.changeListener = changeListener;\n\n // Fire the initial event\n if (this.receivedInitialUser) {\n changeListener(this.currentUser);\n }\n }\n\n removeChangeListener(): void {\n debugAssert(\n this.tokenListener != null,\n 'removeChangeListener() called twice'\n );\n debugAssert(\n this.changeListener !== null,\n 'removeChangeListener() called when no listener registered'\n );\n\n if (this.auth) {\n this.auth.removeAuthTokenListener(this.tokenListener!);\n }\n this.tokenListener = null;\n this.changeListener = null;\n }\n\n // Auth.getUid() can return null even with a user logged in. It is because\n // getUid() is synchronous, but the auth code populating Uid is asynchronous.\n // This method should only be called in the AuthTokenListener callback\n // to guarantee to get the actual user.\n private getUser(): User {\n const currentUid = this.auth && this.auth.getUid();\n hardAssert(\n currentUid === null || typeof currentUid === 'string',\n 'Received invalid UID: ' + currentUid\n );\n return new User(currentUid);\n }\n}\n\n// Manual type definition for the subset of Gapi we use.\ninterface Gapi {\n auth: {\n getAuthHeaderValueForFirstParty: (\n userIdentifiers: Array<{ [key: string]: string }>\n ) => string | null;\n };\n}\n\n/*\n * FirstPartyToken provides a fresh token each time its value\n * is requested, because if the token is too old, requests will be rejected.\n * Technically this may no longer be necessary since the SDK should gracefully\n * recover from unauthenticated errors (see b/33147818 for context), but it's\n * safer to keep the implementation as-is.\n */\nexport class FirstPartyToken implements Token {\n type = 'FirstParty' as TokenType;\n user = User.FIRST_PARTY;\n\n constructor(private gapi: Gapi, private sessionIndex: string) {}\n\n get authHeaders(): { [header: string]: string } {\n const headers: { [header: string]: string } = {\n 'X-Goog-AuthUser': this.sessionIndex\n };\n const authHeader = this.gapi.auth.getAuthHeaderValueForFirstParty([]);\n if (authHeader) {\n headers['Authorization'] = authHeader;\n }\n return headers;\n }\n}\n\n/*\n * Provides user credentials required for the Firestore JavaScript SDK\n * to authenticate the user, using technique that is only available\n * to applications hosted by Google.\n */\nexport class FirstPartyCredentialsProvider implements CredentialsProvider {\n constructor(private gapi: Gapi, private sessionIndex: string) {}\n\n getToken(): Promise<Token | null> {\n return Promise.resolve(new FirstPartyToken(this.gapi, this.sessionIndex));\n }\n\n setChangeListener(changeListener: CredentialChangeListener): void {\n // Fire with initial uid.\n changeListener(User.FIRST_PARTY);\n }\n\n removeChangeListener(): void {}\n\n invalidateToken(): void {}\n}\n\n/**\n * Builds a CredentialsProvider depending on the type of\n * the credentials passed in.\n */\nexport function makeCredentialsProvider(\n credentials?: CredentialsSettings\n): CredentialsProvider {\n if (!credentials) {\n return new EmptyCredentialsProvider();\n }\n\n switch (credentials.type) {\n case 'gapi':\n const client = credentials.client as Gapi;\n // Make sure this really is a Gapi client.\n hardAssert(\n !!(\n typeof client === 'object' &&\n client !== null &&\n client['auth'] &&\n client['auth']['getAuthHeaderValueForFirstParty']\n ),\n 'unexpected gapi interface'\n );\n return new FirstPartyCredentialsProvider(\n client,\n credentials.sessionIndex || '0'\n );\n\n case 'provider':\n return credentials.client;\n\n default:\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'makeCredentialsProvider failed due to invalid credential type'\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Code, FirestoreError } from '../util/error';\nimport { primitiveComparator } from '../util/misc';\n\n// The earlist date supported by Firestore timestamps (0001-01-01T00:00:00Z).\nconst MIN_SECONDS = -62135596800;\n\nexport class Timestamp {\n static now(): Timestamp {\n return Timestamp.fromMillis(Date.now());\n }\n\n static fromDate(date: Date): Timestamp {\n return Timestamp.fromMillis(date.getTime());\n }\n\n static fromMillis(milliseconds: number): Timestamp {\n const seconds = Math.floor(milliseconds / 1000);\n const nanos = (milliseconds - seconds * 1000) * 1e6;\n return new Timestamp(seconds, nanos);\n }\n\n constructor(readonly seconds: number, readonly nanoseconds: number) {\n if (nanoseconds < 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Timestamp nanoseconds out of range: ' + nanoseconds\n );\n }\n if (nanoseconds >= 1e9) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Timestamp nanoseconds out of range: ' + nanoseconds\n );\n }\n if (seconds < MIN_SECONDS) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Timestamp seconds out of range: ' + seconds\n );\n }\n // This will break in the year 10,000.\n if (seconds >= 253402300800) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Timestamp seconds out of range: ' + seconds\n );\n }\n }\n\n toDate(): Date {\n return new Date(this.toMillis());\n }\n\n toMillis(): number {\n return this.seconds * 1000 + this.nanoseconds / 1e6;\n }\n\n _compareTo(other: Timestamp): number {\n if (this.seconds === other.seconds) {\n return primitiveComparator(this.nanoseconds, other.nanoseconds);\n }\n return primitiveComparator(this.seconds, other.seconds);\n }\n\n isEqual(other: Timestamp): boolean {\n return (\n other.seconds === this.seconds && other.nanoseconds === this.nanoseconds\n );\n }\n\n toString(): string {\n return (\n 'Timestamp(seconds=' +\n this.seconds +\n ', nanoseconds=' +\n this.nanoseconds +\n ')'\n );\n }\n\n valueOf(): string {\n // This method returns a string of the form <seconds>.<nanoseconds> where <seconds> is\n // translated to have a non-negative value and both <seconds> and <nanoseconds> are left-padded\n // with zeroes to be a consistent length. Strings with this format then have a lexiographical\n // ordering that matches the expected ordering. The <seconds> translation is done to avoid\n // having a leading negative sign (i.e. a leading '-' character) in its string representation,\n // which would affect its lexiographical ordering.\n const adjustedSeconds = this.seconds - MIN_SECONDS;\n // Note: Up to 12 decimal digits are required to represent all valid 'seconds' values.\n const formattedSeconds = String(adjustedSeconds).padStart(12, '0');\n const formattedNanoseconds = String(this.nanoseconds).padStart(9, '0');\n return formattedSeconds + '.' + formattedNanoseconds;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Timestamp } from '../api/timestamp';\n\n/**\n * A version of a document in Firestore. This corresponds to the version\n * timestamp, such as update_time or read_time.\n */\nexport class SnapshotVersion {\n static fromTimestamp(value: Timestamp): SnapshotVersion {\n return new SnapshotVersion(value);\n }\n\n static min(): SnapshotVersion {\n return new SnapshotVersion(new Timestamp(0, 0));\n }\n\n private constructor(private timestamp: Timestamp) {}\n\n compareTo(other: SnapshotVersion): number {\n return this.timestamp._compareTo(other.timestamp);\n }\n\n isEqual(other: SnapshotVersion): boolean {\n return this.timestamp.isEqual(other.timestamp);\n }\n\n /** Returns a number representation of the version for use in spec tests. */\n toMicroseconds(): number {\n // Convert to microseconds.\n return this.timestamp.seconds * 1e6 + this.timestamp.nanoseconds / 1000;\n }\n\n toString(): string {\n return 'SnapshotVersion(' + this.timestamp.toString() + ')';\n }\n\n toTimestamp(): Timestamp {\n return this.timestamp;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert, fail } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\n\nexport const DOCUMENT_KEY_NAME = '__name__';\n\n/**\n * Path represents an ordered sequence of string segments.\n */\nabstract class BasePath<B extends BasePath<B>> {\n private segments: string[];\n private offset: number;\n private len: number;\n\n constructor(segments: string[], offset?: number, length?: number) {\n if (offset === undefined) {\n offset = 0;\n } else if (offset > segments.length) {\n fail('offset ' + offset + ' out of range ' + segments.length);\n }\n\n if (length === undefined) {\n length = segments.length - offset;\n } else if (length > segments.length - offset) {\n fail('length ' + length + ' out of range ' + (segments.length - offset));\n }\n this.segments = segments;\n this.offset = offset;\n this.len = length;\n }\n\n /**\n * Abstract constructor method to construct an instance of B with the given\n * parameters.\n */\n protected abstract construct(\n segments: string[],\n offset?: number,\n length?: number\n ): B;\n\n /**\n * Returns a String representation.\n *\n * Implementing classes are required to provide deterministic implementations as\n * the String representation is used to obtain canonical Query IDs.\n */\n abstract toString(): string;\n\n get length(): number {\n return this.len;\n }\n\n isEqual(other: B): boolean {\n return BasePath.comparator(this, other) === 0;\n }\n\n child(nameOrPath: string | B): B {\n const segments = this.segments.slice(this.offset, this.limit());\n if (nameOrPath instanceof BasePath) {\n nameOrPath.forEach(segment => {\n segments.push(segment);\n });\n } else {\n segments.push(nameOrPath);\n }\n return this.construct(segments);\n }\n\n /** The index of one past the last segment of the path. */\n private limit(): number {\n return this.offset + this.length;\n }\n\n popFirst(size?: number): B {\n size = size === undefined ? 1 : size;\n debugAssert(\n this.length >= size,\n \"Can't call popFirst() with less segments\"\n );\n return this.construct(\n this.segments,\n this.offset + size,\n this.length - size\n );\n }\n\n popLast(): B {\n debugAssert(!this.isEmpty(), \"Can't call popLast() on empty path\");\n return this.construct(this.segments, this.offset, this.length - 1);\n }\n\n firstSegment(): string {\n debugAssert(!this.isEmpty(), \"Can't call firstSegment() on empty path\");\n return this.segments[this.offset];\n }\n\n lastSegment(): string {\n return this.get(this.length - 1);\n }\n\n get(index: number): string {\n debugAssert(index < this.length, 'Index out of range');\n return this.segments[this.offset + index];\n }\n\n isEmpty(): boolean {\n return this.length === 0;\n }\n\n isPrefixOf(other: this): boolean {\n if (other.length < this.length) {\n return false;\n }\n\n for (let i = 0; i < this.length; i++) {\n if (this.get(i) !== other.get(i)) {\n return false;\n }\n }\n\n return true;\n }\n\n isImmediateParentOf(potentialChild: this): boolean {\n if (this.length + 1 !== potentialChild.length) {\n return false;\n }\n\n for (let i = 0; i < this.length; i++) {\n if (this.get(i) !== potentialChild.get(i)) {\n return false;\n }\n }\n\n return true;\n }\n\n forEach(fn: (segment: string) => void): void {\n for (let i = this.offset, end = this.limit(); i < end; i++) {\n fn(this.segments[i]);\n }\n }\n\n toArray(): string[] {\n return this.segments.slice(this.offset, this.limit());\n }\n\n static comparator<T extends BasePath<T>>(\n p1: BasePath<T>,\n p2: BasePath<T>\n ): number {\n const len = Math.min(p1.length, p2.length);\n for (let i = 0; i < len; i++) {\n const left = p1.get(i);\n const right = p2.get(i);\n if (left < right) {\n return -1;\n }\n if (left > right) {\n return 1;\n }\n }\n if (p1.length < p2.length) {\n return -1;\n }\n if (p1.length > p2.length) {\n return 1;\n }\n return 0;\n }\n}\n\n/**\n * A slash-separated path for navigating resources (documents and collections)\n * within Firestore.\n */\nexport class ResourcePath extends BasePath<ResourcePath> {\n protected construct(\n segments: string[],\n offset?: number,\n length?: number\n ): ResourcePath {\n return new ResourcePath(segments, offset, length);\n }\n\n canonicalString(): string {\n // NOTE: The client is ignorant of any path segments containing escape\n // sequences (e.g. __id123__) and just passes them through raw (they exist\n // for legacy reasons and should not be used frequently).\n\n return this.toArray().join('/');\n }\n\n toString(): string {\n return this.canonicalString();\n }\n\n /**\n * Creates a resource path from the given slash-delimited string.\n */\n static fromString(path: string): ResourcePath {\n // NOTE: The client is ignorant of any path segments containing escape\n // sequences (e.g. __id123__) and just passes them through raw (they exist\n // for legacy reasons and should not be used frequently).\n\n if (path.indexOf('//') >= 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid path (${path}). Paths must not contain // in them.`\n );\n }\n\n // We may still have an empty segment at the beginning or end if they had a\n // leading or trailing slash (which we allow).\n const segments = path.split('/').filter(segment => segment.length > 0);\n\n return new ResourcePath(segments);\n }\n\n static EMPTY_PATH = new ResourcePath([]);\n}\n\nconst identifierRegExp = /^[_a-zA-Z][_a-zA-Z0-9]*$/;\n\n/** A dot-separated path for navigating sub-objects within a document. */\nexport class FieldPath extends BasePath<FieldPath> {\n protected construct(\n segments: string[],\n offset?: number,\n length?: number\n ): FieldPath {\n return new FieldPath(segments, offset, length);\n }\n\n /**\n * Returns true if the string could be used as a segment in a field path\n * without escaping.\n */\n private static isValidIdentifier(segment: string): boolean {\n return identifierRegExp.test(segment);\n }\n\n canonicalString(): string {\n return this.toArray()\n .map(str => {\n str = str.replace('\\\\', '\\\\\\\\').replace('`', '\\\\`');\n if (!FieldPath.isValidIdentifier(str)) {\n str = '`' + str + '`';\n }\n return str;\n })\n .join('.');\n }\n\n toString(): string {\n return this.canonicalString();\n }\n\n /**\n * Returns true if this field references the key of a document.\n */\n isKeyField(): boolean {\n return this.length === 1 && this.get(0) === DOCUMENT_KEY_NAME;\n }\n\n /**\n * The field designating the key of a document.\n */\n static keyField(): FieldPath {\n return new FieldPath([DOCUMENT_KEY_NAME]);\n }\n\n /**\n * Parses a field string from the given server-formatted string.\n *\n * - Splitting the empty string is not allowed (for now at least).\n * - Empty segments within the string (e.g. if there are two consecutive\n * separators) are not allowed.\n *\n * TODO(b/37244157): we should make this more strict. Right now, it allows\n * non-identifier path components, even if they aren't escaped.\n */\n static fromServerFormat(path: string): FieldPath {\n const segments: string[] = [];\n let current = '';\n let i = 0;\n\n const addCurrentSegment = (): void => {\n if (current.length === 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid field path (${path}). Paths must not be empty, begin ` +\n `with '.', end with '.', or contain '..'`\n );\n }\n segments.push(current);\n current = '';\n };\n\n let inBackticks = false;\n\n while (i < path.length) {\n const c = path[i];\n if (c === '\\\\') {\n if (i + 1 === path.length) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Path has trailing escape character: ' + path\n );\n }\n const next = path[i + 1];\n if (!(next === '\\\\' || next === '.' || next === '`')) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Path has invalid escape sequence: ' + path\n );\n }\n current += next;\n i += 2;\n } else if (c === '`') {\n inBackticks = !inBackticks;\n i++;\n } else if (c === '.' && !inBackticks) {\n addCurrentSegment();\n i++;\n } else {\n current += c;\n i++;\n }\n }\n addCurrentSegment();\n\n if (inBackticks) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Unterminated ` in path: ' + path\n );\n }\n\n return new FieldPath(segments);\n }\n\n static EMPTY_PATH = new FieldPath([]);\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert } from '../util/assert';\n\nimport { ResourcePath } from './path';\n\nexport class DocumentKey {\n constructor(readonly path: ResourcePath) {\n debugAssert(\n DocumentKey.isDocumentKey(path),\n 'Invalid DocumentKey with an odd number of segments: ' +\n path.toArray().join('/')\n );\n }\n\n static fromName(name: string): DocumentKey {\n return new DocumentKey(ResourcePath.fromString(name).popFirst(5));\n }\n\n /** Returns true if the document is in the specified collectionId. */\n hasCollectionId(collectionId: string): boolean {\n return (\n this.path.length >= 2 &&\n this.path.get(this.path.length - 2) === collectionId\n );\n }\n\n isEqual(other: DocumentKey | null): boolean {\n return (\n other !== null && ResourcePath.comparator(this.path, other.path) === 0\n );\n }\n\n toString(): string {\n return this.path.toString();\n }\n\n static EMPTY = new DocumentKey(new ResourcePath([]));\n\n static comparator(k1: DocumentKey, k2: DocumentKey): number {\n return ResourcePath.comparator(k1.path, k2.path);\n }\n\n static isDocumentKey(path: ResourcePath): boolean {\n return path.length % 2 === 0;\n }\n\n /**\n * Creates and returns a new document key with the given segments.\n *\n * @param segments The segments of the path to the document\n * @return A new instance of DocumentKey\n */\n static fromSegments(segments: string[]): DocumentKey {\n return new DocumentKey(new ResourcePath(segments.slice()));\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert } from './assert';\n\nexport interface Dict<V> {\n [stringKey: string]: V;\n}\n\nexport function objectSize<V>(obj: object): number {\n let count = 0;\n for (const key in obj) {\n if (Object.prototype.hasOwnProperty.call(obj, key)) {\n count++;\n }\n }\n return count;\n}\n\nexport function forEach<V>(\n obj: Dict<V>,\n fn: (key: string, val: V) => void\n): void {\n for (const key in obj) {\n if (Object.prototype.hasOwnProperty.call(obj, key)) {\n fn(key, obj[key]);\n }\n }\n}\n\nexport function isEmpty<V>(obj: Dict<V>): boolean {\n debugAssert(\n obj != null && typeof obj === 'object',\n 'isEmpty() expects object parameter.'\n );\n for (const key in obj) {\n if (Object.prototype.hasOwnProperty.call(obj, key)) {\n return false;\n }\n }\n return true;\n}\n","/**\n * @license\n * Copyright 2020 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { PlatformSupport } from '../platform/platform';\nimport { primitiveComparator } from './misc';\n\n/**\n * Immutable class that represents a \"proto\" byte string.\n *\n * Proto byte strings can either be Base64-encoded strings or Uint8Arrays when\n * sent on the wire. This class abstracts away this differentiation by holding\n * the proto byte string in a common class that must be converted into a string\n * before being sent as a proto.\n */\nexport class ByteString {\n static readonly EMPTY_BYTE_STRING = new ByteString('');\n\n private constructor(private readonly binaryString: string) {}\n\n static fromBase64String(base64: string): ByteString {\n const binaryString = PlatformSupport.getPlatform().atob(base64);\n return new ByteString(binaryString);\n }\n\n static fromUint8Array(array: Uint8Array): ByteString {\n const binaryString = binaryStringFromUint8Array(array);\n return new ByteString(binaryString);\n }\n\n toBase64(): string {\n return PlatformSupport.getPlatform().btoa(this.binaryString);\n }\n\n toUint8Array(): Uint8Array {\n return uint8ArrayFromBinaryString(this.binaryString);\n }\n\n approximateByteSize(): number {\n return this.binaryString.length * 2;\n }\n\n compareTo(other: ByteString): number {\n return primitiveComparator(this.binaryString, other.binaryString);\n }\n\n isEqual(other: ByteString): boolean {\n return this.binaryString === other.binaryString;\n }\n}\n\n/**\n * Helper function to convert an Uint8array to a binary string.\n */\nexport function binaryStringFromUint8Array(array: Uint8Array): string {\n let binaryString = '';\n for (let i = 0; i < array.length; ++i) {\n binaryString += String.fromCharCode(array[i]);\n }\n return binaryString;\n}\n\n/**\n * Helper function to convert a binary string to an Uint8Array.\n */\nexport function uint8ArrayFromBinaryString(binaryString: string): Uint8Array {\n const buffer = new Uint8Array(binaryString.length);\n for (let i = 0; i < binaryString.length; i++) {\n buffer[i] = binaryString.charCodeAt(i);\n }\n return buffer;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\n// An Object whose keys and values are strings.\nexport interface StringMap {\n [key: string]: string;\n}\n\n/**\n * Returns whether a variable is either undefined or null.\n */\nexport function isNullOrUndefined(value: unknown): value is null | undefined {\n return value === null || value === undefined;\n}\n\n/** Returns whether the value represents -0. */\nexport function isNegativeZero(value: number): boolean {\n // Detect if the value is -0.0. Based on polyfill from\n // https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Object/is\n return value === -0 && 1 / value === 1 / -0;\n}\n\n/**\n * Returns whether a value is an integer and in the safe integer range\n * @param value The value to test for being an integer and in the safe range\n */\nexport function isSafeInteger(value: unknown): boolean {\n return (\n typeof value === 'number' &&\n Number.isInteger(value) &&\n !isNegativeZero(value) &&\n value <= Number.MAX_SAFE_INTEGER &&\n value >= Number.MIN_SAFE_INTEGER\n );\n}\n","/**\n * @license\n * Copyright 2020 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\nimport { Timestamp } from '../api/timestamp';\nimport { normalizeTimestamp } from './values';\n\n/**\n * Represents a locally-applied ServerTimestamp.\n *\n * Server Timestamps are backed by MapValues that contain an internal field\n * `__type__` with a value of `server_timestamp`. The previous value and local\n * write time are stored in its `__previous_value__` and `__local_write_time__`\n * fields respectively.\n *\n * Notes:\n * - ServerTimestampValue instances are created as the result of applying a\n * TransformMutation (see TransformMutation.applyTo()). They can only exist in\n * the local view of a document. Therefore they do not need to be parsed or\n * serialized.\n * - When evaluated locally (e.g. for snapshot.data()), they by default\n * evaluate to `null`. This behavior can be configured by passing custom\n * FieldValueOptions to value().\n * - With respect to other ServerTimestampValues, they sort by their\n * localWriteTime.\n */\n\nconst SERVER_TIMESTAMP_SENTINEL = 'server_timestamp';\nconst TYPE_KEY = '__type__';\nconst PREVIOUS_VALUE_KEY = '__previous_value__';\nconst LOCAL_WRITE_TIME_KEY = '__local_write_time__';\n\nexport function isServerTimestamp(value: api.Value | null): boolean {\n const type = (value?.mapValue?.fields || {})[TYPE_KEY]?.stringValue;\n return type === SERVER_TIMESTAMP_SENTINEL;\n}\n\n/**\n * Creates a new ServerTimestamp proto value (using the internal format).\n */\nexport function serverTimestamp(\n localWriteTime: Timestamp,\n previousValue: api.Value | null\n): api.Value {\n const mapValue: api.MapValue = {\n fields: {\n [TYPE_KEY]: {\n stringValue: SERVER_TIMESTAMP_SENTINEL\n },\n [LOCAL_WRITE_TIME_KEY]: {\n timestampValue: {\n seconds: localWriteTime.seconds,\n nanos: localWriteTime.nanoseconds\n }\n }\n }\n };\n\n if (previousValue) {\n mapValue.fields![PREVIOUS_VALUE_KEY] = previousValue;\n }\n\n return { mapValue };\n}\n\n/**\n * Returns the value of the field before this ServerTimestamp was set.\n *\n * Preserving the previous values allows the user to display the last resoled\n * value until the backend responds with the timestamp.\n */\nexport function getPreviousValue(value: api.Value): api.Value | null {\n const previousValue = value.mapValue!.fields![PREVIOUS_VALUE_KEY];\n\n if (isServerTimestamp(previousValue)) {\n return getPreviousValue(previousValue);\n }\n return previousValue;\n}\n\n/**\n * Returns the local time at which this timestamp was first set.\n */\nexport function getLocalWriteTime(value: api.Value): Timestamp {\n const localWriteTime = normalizeTimestamp(\n value.mapValue!.fields![LOCAL_WRITE_TIME_KEY].timestampValue!\n );\n return new Timestamp(localWriteTime.seconds, localWriteTime.nanos);\n}\n","/**\n * @license\n * Copyright 2020 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { TypeOrder } from './object_value';\nimport { fail, hardAssert } from '../util/assert';\nimport { forEach, objectSize } from '../util/obj';\nimport { ByteString } from '../util/byte_string';\nimport { isNegativeZero } from '../util/types';\nimport { DocumentKey } from './document_key';\nimport { arrayEquals, primitiveComparator } from '../util/misc';\nimport { DatabaseId } from '../core/database_info';\nimport {\n getLocalWriteTime,\n getPreviousValue,\n isServerTimestamp\n} from './server_timestamps';\n\n// A RegExp matching ISO 8601 UTC timestamps with optional fraction.\nconst ISO_TIMESTAMP_REG_EXP = new RegExp(\n /^\\d{4}-\\d\\d-\\d\\dT\\d\\d:\\d\\d:\\d\\d(?:\\.(\\d+))?Z$/\n);\n\n/** Extracts the backend's type order for the provided value. */\nexport function typeOrder(value: api.Value): TypeOrder {\n if ('nullValue' in value) {\n return TypeOrder.NullValue;\n } else if ('booleanValue' in value) {\n return TypeOrder.BooleanValue;\n } else if ('integerValue' in value || 'doubleValue' in value) {\n return TypeOrder.NumberValue;\n } else if ('timestampValue' in value) {\n return TypeOrder.TimestampValue;\n } else if ('stringValue' in value) {\n return TypeOrder.StringValue;\n } else if ('bytesValue' in value) {\n return TypeOrder.BlobValue;\n } else if ('referenceValue' in value) {\n return TypeOrder.RefValue;\n } else if ('geoPointValue' in value) {\n return TypeOrder.GeoPointValue;\n } else if ('arrayValue' in value) {\n return TypeOrder.ArrayValue;\n } else if ('mapValue' in value) {\n if (isServerTimestamp(value)) {\n return TypeOrder.ServerTimestampValue;\n }\n return TypeOrder.ObjectValue;\n } else {\n return fail('Invalid value type: ' + JSON.stringify(value));\n }\n}\n\n/** Tests `left` and `right` for equality based on the backend semantics. */\nexport function valueEquals(left: api.Value, right: api.Value): boolean {\n const leftType = typeOrder(left);\n const rightType = typeOrder(right);\n if (leftType !== rightType) {\n return false;\n }\n\n switch (leftType) {\n case TypeOrder.NullValue:\n return true;\n case TypeOrder.BooleanValue:\n return left.booleanValue === right.booleanValue;\n case TypeOrder.ServerTimestampValue:\n return getLocalWriteTime(left).isEqual(getLocalWriteTime(right));\n case TypeOrder.TimestampValue:\n return timestampEquals(left, right);\n case TypeOrder.StringValue:\n return left.stringValue === right.stringValue;\n case TypeOrder.BlobValue:\n return blobEquals(left, right);\n case TypeOrder.RefValue:\n return left.referenceValue === right.referenceValue;\n case TypeOrder.GeoPointValue:\n return geoPointEquals(left, right);\n case TypeOrder.NumberValue:\n return numberEquals(left, right);\n case TypeOrder.ArrayValue:\n return arrayEquals(\n left.arrayValue!.values || [],\n right.arrayValue!.values || [],\n valueEquals\n );\n case TypeOrder.ObjectValue:\n return objectEquals(left, right);\n default:\n return fail('Unexpected value type: ' + JSON.stringify(left));\n }\n}\n\nfunction timestampEquals(left: api.Value, right: api.Value): boolean {\n if (\n typeof left.timestampValue === 'string' &&\n typeof right.timestampValue === 'string' &&\n left.timestampValue.length === right.timestampValue.length\n ) {\n // Use string equality for ISO 8601 timestamps\n return left.timestampValue === right.timestampValue;\n }\n\n const leftTimestamp = normalizeTimestamp(left.timestampValue!);\n const rightTimestamp = normalizeTimestamp(right.timestampValue!);\n return (\n leftTimestamp.seconds === rightTimestamp.seconds &&\n leftTimestamp.nanos === rightTimestamp.nanos\n );\n}\n\nfunction geoPointEquals(left: api.Value, right: api.Value): boolean {\n return (\n normalizeNumber(left.geoPointValue!.latitude) ===\n normalizeNumber(right.geoPointValue!.latitude) &&\n normalizeNumber(left.geoPointValue!.longitude) ===\n normalizeNumber(right.geoPointValue!.longitude)\n );\n}\n\nfunction blobEquals(left: api.Value, right: api.Value): boolean {\n return normalizeByteString(left.bytesValue!).isEqual(\n normalizeByteString(right.bytesValue!)\n );\n}\n\nexport function numberEquals(left: api.Value, right: api.Value): boolean {\n if ('integerValue' in left && 'integerValue' in right) {\n return (\n normalizeNumber(left.integerValue) === normalizeNumber(right.integerValue)\n );\n } else if ('doubleValue' in left && 'doubleValue' in right) {\n const n1 = normalizeNumber(left.doubleValue!);\n const n2 = normalizeNumber(right.doubleValue!);\n\n if (n1 === n2) {\n return isNegativeZero(n1) === isNegativeZero(n2);\n } else {\n return isNaN(n1) && isNaN(n2);\n }\n }\n\n return false;\n}\n\nfunction objectEquals(left: api.Value, right: api.Value): boolean {\n const leftMap = left.mapValue!.fields || {};\n const rightMap = right.mapValue!.fields || {};\n\n if (objectSize(leftMap) !== objectSize(rightMap)) {\n return false;\n }\n\n for (const key in leftMap) {\n if (leftMap.hasOwnProperty(key)) {\n if (\n rightMap[key] === undefined ||\n !valueEquals(leftMap[key], rightMap[key])\n ) {\n return false;\n }\n }\n }\n return true;\n}\n\n/** Returns true if the ArrayValue contains the specified element. */\nexport function arrayValueContains(\n haystack: api.ArrayValue,\n needle: api.Value\n): boolean {\n return (\n (haystack.values || []).find(v => valueEquals(v, needle)) !== undefined\n );\n}\n\nexport function valueCompare(left: api.Value, right: api.Value): number {\n const leftType = typeOrder(left);\n const rightType = typeOrder(right);\n\n if (leftType !== rightType) {\n return primitiveComparator(leftType, rightType);\n }\n\n switch (leftType) {\n case TypeOrder.NullValue:\n return 0;\n case TypeOrder.BooleanValue:\n return primitiveComparator(left.booleanValue!, right.booleanValue!);\n case TypeOrder.NumberValue:\n return compareNumbers(left, right);\n case TypeOrder.TimestampValue:\n return compareTimestamps(left.timestampValue!, right.timestampValue!);\n case TypeOrder.ServerTimestampValue:\n return compareTimestamps(\n getLocalWriteTime(left),\n getLocalWriteTime(right)\n );\n case TypeOrder.StringValue:\n return primitiveComparator(left.stringValue!, right.stringValue!);\n case TypeOrder.BlobValue:\n return compareBlobs(left.bytesValue!, right.bytesValue!);\n case TypeOrder.RefValue:\n return compareReferences(left.referenceValue!, right.referenceValue!);\n case TypeOrder.GeoPointValue:\n return compareGeoPoints(left.geoPointValue!, right.geoPointValue!);\n case TypeOrder.ArrayValue:\n return compareArrays(left.arrayValue!, right.arrayValue!);\n case TypeOrder.ObjectValue:\n return compareMaps(left.mapValue!, right.mapValue!);\n default:\n throw fail('Invalid value type: ' + leftType);\n }\n}\n\nfunction compareNumbers(left: api.Value, right: api.Value): number {\n const leftNumber = normalizeNumber(left.integerValue || left.doubleValue);\n const rightNumber = normalizeNumber(right.integerValue || right.doubleValue);\n\n if (leftNumber < rightNumber) {\n return -1;\n } else if (leftNumber > rightNumber) {\n return 1;\n } else if (leftNumber === rightNumber) {\n return 0;\n } else {\n // one or both are NaN.\n if (isNaN(leftNumber)) {\n return isNaN(rightNumber) ? 0 : -1;\n } else {\n return 1;\n }\n }\n}\n\nfunction compareTimestamps(left: api.Timestamp, right: api.Timestamp): number {\n if (\n typeof left === 'string' &&\n typeof right === 'string' &&\n left.length === right.length\n ) {\n return primitiveComparator(left, right);\n }\n\n const leftTimestamp = normalizeTimestamp(left);\n const rightTimestamp = normalizeTimestamp(right);\n\n const comparison = primitiveComparator(\n leftTimestamp.seconds,\n rightTimestamp.seconds\n );\n if (comparison !== 0) {\n return comparison;\n }\n return primitiveComparator(leftTimestamp.nanos, rightTimestamp.nanos);\n}\n\nfunction compareReferences(leftPath: string, rightPath: string): number {\n const leftSegments = leftPath.split('/');\n const rightSegments = rightPath.split('/');\n for (let i = 0; i < leftSegments.length && i < rightSegments.length; i++) {\n const comparison = primitiveComparator(leftSegments[i], rightSegments[i]);\n if (comparison !== 0) {\n return comparison;\n }\n }\n return primitiveComparator(leftSegments.length, rightSegments.length);\n}\n\nfunction compareGeoPoints(left: api.LatLng, right: api.LatLng): number {\n const comparison = primitiveComparator(\n normalizeNumber(left.latitude),\n normalizeNumber(right.latitude)\n );\n if (comparison !== 0) {\n return comparison;\n }\n return primitiveComparator(\n normalizeNumber(left.longitude),\n normalizeNumber(right.longitude)\n );\n}\n\nfunction compareBlobs(\n left: string | Uint8Array,\n right: string | Uint8Array\n): number {\n const leftBytes = normalizeByteString(left);\n const rightBytes = normalizeByteString(right);\n return leftBytes.compareTo(rightBytes);\n}\n\nfunction compareArrays(left: api.ArrayValue, right: api.ArrayValue): number {\n const leftArray = left.values || [];\n const rightArray = right.values || [];\n\n for (let i = 0; i < leftArray.length && i < rightArray.length; ++i) {\n const compare = valueCompare(leftArray[i], rightArray[i]);\n if (compare) {\n return compare;\n }\n }\n return primitiveComparator(leftArray.length, rightArray.length);\n}\n\nfunction compareMaps(left: api.MapValue, right: api.MapValue): number {\n const leftMap = left.fields || {};\n const leftKeys = Object.keys(leftMap);\n const rightMap = right.fields || {};\n const rightKeys = Object.keys(rightMap);\n\n // Even though MapValues are likely sorted correctly based on their insertion\n // order (e.g. when received from the backend), local modifications can bring\n // elements out of order. We need to re-sort the elements to ensure that\n // canonical IDs are independent of insertion order.\n leftKeys.sort();\n rightKeys.sort();\n\n for (let i = 0; i < leftKeys.length && i < rightKeys.length; ++i) {\n const keyCompare = primitiveComparator(leftKeys[i], rightKeys[i]);\n if (keyCompare !== 0) {\n return keyCompare;\n }\n const compare = valueCompare(leftMap[leftKeys[i]], rightMap[rightKeys[i]]);\n if (compare !== 0) {\n return compare;\n }\n }\n\n return primitiveComparator(leftKeys.length, rightKeys.length);\n}\n\n/**\n * Generates the canonical ID for the provided field value (as used in Target\n * serialization).\n */\nexport function canonicalId(value: api.Value): string {\n return canonifyValue(value);\n}\n\nfunction canonifyValue(value: api.Value): string {\n if ('nullValue' in value) {\n return 'null';\n } else if ('booleanValue' in value) {\n return '' + value.booleanValue!;\n } else if ('integerValue' in value) {\n return '' + value.integerValue!;\n } else if ('doubleValue' in value) {\n return '' + value.doubleValue!;\n } else if ('timestampValue' in value) {\n return canonifyTimestamp(value.timestampValue!);\n } else if ('stringValue' in value) {\n return value.stringValue!;\n } else if ('bytesValue' in value) {\n return canonifyByteString(value.bytesValue!);\n } else if ('referenceValue' in value) {\n return canonifyReference(value.referenceValue!);\n } else if ('geoPointValue' in value) {\n return canonifyGeoPoint(value.geoPointValue!);\n } else if ('arrayValue' in value) {\n return canonifyArray(value.arrayValue!);\n } else if ('mapValue' in value) {\n return canonifyMap(value.mapValue!);\n } else {\n return fail('Invalid value type: ' + JSON.stringify(value));\n }\n}\n\nfunction canonifyByteString(byteString: string | Uint8Array): string {\n return normalizeByteString(byteString).toBase64();\n}\n\nfunction canonifyTimestamp(timestamp: api.Timestamp): string {\n const normalizedTimestamp = normalizeTimestamp(timestamp);\n return `time(${normalizedTimestamp.seconds},${normalizedTimestamp.nanos})`;\n}\n\nfunction canonifyGeoPoint(geoPoint: api.LatLng): string {\n return `geo(${geoPoint.latitude},${geoPoint.longitude})`;\n}\n\nfunction canonifyReference(referenceValue: string): string {\n return DocumentKey.fromName(referenceValue).toString();\n}\n\nfunction canonifyMap(mapValue: api.MapValue): string {\n // Iteration order in JavaScript is not guaranteed. To ensure that we generate\n // matching canonical IDs for identical maps, we need to sort the keys.\n const sortedKeys = Object.keys(mapValue.fields || {}).sort();\n\n let result = '{';\n let first = true;\n for (const key of sortedKeys) {\n if (!first) {\n result += ',';\n } else {\n first = false;\n }\n result += `${key}:${canonifyValue(mapValue.fields![key])}`;\n }\n return result + '}';\n}\n\nfunction canonifyArray(arrayValue: api.ArrayValue): string {\n let result = '[';\n let first = true;\n for (const value of arrayValue.values || []) {\n if (!first) {\n result += ',';\n } else {\n first = false;\n }\n result += canonifyValue(value);\n }\n return result + ']';\n}\n\n/**\n * Returns an approximate (and wildly inaccurate) in-memory size for the field\n * value.\n *\n * The memory size takes into account only the actual user data as it resides\n * in memory and ignores object overhead.\n */\nexport function estimateByteSize(value: api.Value): number {\n switch (typeOrder(value)) {\n case TypeOrder.NullValue:\n return 4;\n case TypeOrder.BooleanValue:\n return 4;\n case TypeOrder.NumberValue:\n return 8;\n case TypeOrder.TimestampValue:\n // Timestamps are made up of two distinct numbers (seconds + nanoseconds)\n return 16;\n case TypeOrder.ServerTimestampValue:\n const previousValue = getPreviousValue(value);\n return previousValue ? 16 + estimateByteSize(previousValue) : 16;\n case TypeOrder.StringValue:\n // See https://developer.mozilla.org/en-US/docs/Web/JavaScript/Data_structures:\n // \"JavaScript's String type is [...] a set of elements of 16-bit unsigned\n // integer values\"\n return value.stringValue!.length * 2;\n case TypeOrder.BlobValue:\n return normalizeByteString(value.bytesValue!).approximateByteSize();\n case TypeOrder.RefValue:\n return value.referenceValue!.length;\n case TypeOrder.GeoPointValue:\n // GeoPoints are made up of two distinct numbers (latitude + longitude)\n return 16;\n case TypeOrder.ArrayValue:\n return estimateArrayByteSize(value.arrayValue!);\n case TypeOrder.ObjectValue:\n return estimateMapByteSize(value.mapValue!);\n default:\n throw fail('Invalid value type: ' + JSON.stringify(value));\n }\n}\n\nfunction estimateMapByteSize(mapValue: api.MapValue): number {\n let size = 0;\n forEach(mapValue.fields || {}, (key, val) => {\n size += key.length + estimateByteSize(val);\n });\n return size;\n}\n\nfunction estimateArrayByteSize(arrayValue: api.ArrayValue): number {\n return (arrayValue.values || []).reduce(\n (previousSize, value) => previousSize + estimateByteSize(value),\n 0\n );\n}\n\n/**\n * Converts the possible Proto values for a timestamp value into a \"seconds and\n * nanos\" representation.\n */\nexport function normalizeTimestamp(\n date: api.Timestamp\n): { seconds: number; nanos: number } {\n hardAssert(!!date, 'Cannot normalize null or undefined timestamp.');\n\n // The json interface (for the browser) will return an iso timestamp string,\n // while the proto js library (for node) will return a\n // google.protobuf.Timestamp instance.\n if (typeof date === 'string') {\n // The date string can have higher precision (nanos) than the Date class\n // (millis), so we do some custom parsing here.\n\n // Parse the nanos right out of the string.\n let nanos = 0;\n const fraction = ISO_TIMESTAMP_REG_EXP.exec(date);\n hardAssert(!!fraction, 'invalid timestamp: ' + date);\n if (fraction[1]) {\n // Pad the fraction out to 9 digits (nanos).\n let nanoStr = fraction[1];\n nanoStr = (nanoStr + '000000000').substr(0, 9);\n nanos = Number(nanoStr);\n }\n\n // Parse the date to get the seconds.\n const parsedDate = new Date(date);\n const seconds = Math.floor(parsedDate.getTime() / 1000);\n\n return { seconds, nanos };\n } else {\n // TODO(b/37282237): Use strings for Proto3 timestamps\n // assert(!this.options.useProto3Json,\n // 'The timestamp instance format requires Proto JS.');\n const seconds = normalizeNumber(date.seconds);\n const nanos = normalizeNumber(date.nanos);\n return { seconds, nanos };\n }\n}\n\n/**\n * Converts the possible Proto types for numbers into a JavaScript number.\n * Returns 0 if the value is not numeric.\n */\nexport function normalizeNumber(value: number | string | undefined): number {\n // TODO(bjornick): Handle int64 greater than 53 bits.\n if (typeof value === 'number') {\n return value;\n } else if (typeof value === 'string') {\n return Number(value);\n } else {\n return 0;\n }\n}\n\n/** Converts the possible Proto types for Blobs into a ByteString. */\nexport function normalizeByteString(blob: string | Uint8Array): ByteString {\n if (typeof blob === 'string') {\n return ByteString.fromBase64String(blob);\n } else {\n return ByteString.fromUint8Array(blob);\n }\n}\n\n/** Returns a reference value for the provided database and key. */\nexport function refValue(databaseId: DatabaseId, key: DocumentKey): api.Value {\n return {\n referenceValue: `projects/${databaseId.projectId}/databases/${\n databaseId.database\n }/documents/${key.path.canonicalString()}`\n };\n}\n\n/** Returns true if `value` is an IntegerValue . */\nexport function isInteger(\n value?: api.Value | null\n): value is { integerValue: string | number } {\n return !!value && 'integerValue' in value;\n}\n\n/** Returns true if `value` is a DoubleValue. */\nexport function isDouble(\n value?: api.Value | null\n): value is { doubleValue: string | number } {\n return !!value && 'doubleValue' in value;\n}\n\n/** Returns true if `value` is either an IntegerValue or a DoubleValue. */\nexport function isNumber(value?: api.Value | null): boolean {\n return isInteger(value) || isDouble(value);\n}\n\n/** Returns true if `value` is an ArrayValue. */\nexport function isArray(\n value?: api.Value | null\n): value is { arrayValue: api.ArrayValue } {\n return !!value && 'arrayValue' in value;\n}\n\n/** Returns true if `value` is a ReferenceValue. */\nexport function isReferenceValue(\n value?: api.Value | null\n): value is { referenceValue: string } {\n return !!value && 'referenceValue' in value;\n}\n\n/** Returns true if `value` is a NullValue. */\nexport function isNullValue(\n value?: api.Value | null\n): value is { nullValue: 'NULL_VALUE' } {\n return !!value && 'nullValue' in value;\n}\n\n/** Returns true if `value` is NaN. */\nexport function isNanValue(\n value?: api.Value | null\n): value is { doubleValue: 'NaN' | number } {\n return !!value && 'doubleValue' in value && isNaN(Number(value.doubleValue));\n}\n\n/** Returns true if `value` is a MapValue. */\nexport function isMapValue(\n value?: api.Value | null\n): value is { mapValue: api.MapValue } {\n return !!value && 'mapValue' in value;\n}\n","/**\n * @license\n * Copyright 2018 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { Timestamp } from '../api/timestamp';\nimport { debugAssert } from '../util/assert';\nimport { JsonProtoSerializer } from '../remote/serializer';\nimport {\n valueEquals,\n isArray,\n isInteger,\n isNumber,\n normalizeNumber\n} from './values';\nimport { serverTimestamp } from './server_timestamps';\nimport { arrayEquals } from '../util/misc';\n\n/** Represents a transform within a TransformMutation. */\nexport interface TransformOperation {\n /**\n * Computes the local transform result against the provided `previousValue`,\n * optionally using the provided localWriteTime.\n */\n applyToLocalView(\n previousValue: api.Value | null,\n localWriteTime: Timestamp\n ): api.Value;\n\n /**\n * Computes a final transform result after the transform has been acknowledged\n * by the server, potentially using the server-provided transformResult.\n */\n applyToRemoteDocument(\n previousValue: api.Value | null,\n transformResult: api.Value | null\n ): api.Value;\n\n /**\n * If this transform operation is not idempotent, returns the base value to\n * persist for this transform. If a base value is returned, the transform\n * operation is always applied to this base value, even if document has\n * already been updated.\n *\n * Base values provide consistent behavior for non-idempotent transforms and\n * allow us to return the same latency-compensated value even if the backend\n * has already applied the transform operation. The base value is null for\n * idempotent transforms, as they can be re-played even if the backend has\n * already applied them.\n *\n * @return a base value to store along with the mutation, or null for\n * idempotent transforms.\n */\n computeBaseValue(previousValue: api.Value | null): api.Value | null;\n\n isEqual(other: TransformOperation): boolean;\n}\n\n/** Transforms a value into a server-generated timestamp. */\nexport class ServerTimestampTransform implements TransformOperation {\n private constructor() {}\n static instance = new ServerTimestampTransform();\n\n applyToLocalView(\n previousValue: api.Value | null,\n localWriteTime: Timestamp\n ): api.Value {\n return serverTimestamp(localWriteTime!, previousValue);\n }\n\n applyToRemoteDocument(\n previousValue: api.Value | null,\n transformResult: api.Value | null\n ): api.Value {\n return transformResult!;\n }\n\n computeBaseValue(previousValue: api.Value | null): api.Value | null {\n return null; // Server timestamps are idempotent and don't require a base value.\n }\n\n isEqual(other: TransformOperation): boolean {\n return other instanceof ServerTimestampTransform;\n }\n}\n\n/** Transforms an array value via a union operation. */\nexport class ArrayUnionTransformOperation implements TransformOperation {\n constructor(readonly elements: api.Value[]) {}\n\n applyToLocalView(\n previousValue: api.Value | null,\n localWriteTime: Timestamp\n ): api.Value {\n return this.apply(previousValue);\n }\n\n applyToRemoteDocument(\n previousValue: api.Value | null,\n transformResult: api.Value | null\n ): api.Value {\n // The server just sends null as the transform result for array operations,\n // so we have to calculate a result the same as we do for local\n // applications.\n return this.apply(previousValue);\n }\n\n private apply(previousValue: api.Value | null): api.Value {\n const values = coercedFieldValuesArray(previousValue);\n for (const toUnion of this.elements) {\n if (!values.some(element => valueEquals(element, toUnion))) {\n values.push(toUnion);\n }\n }\n return { arrayValue: { values } };\n }\n\n computeBaseValue(previousValue: api.Value | null): api.Value | null {\n return null; // Array transforms are idempotent and don't require a base value.\n }\n\n isEqual(other: TransformOperation): boolean {\n return (\n other instanceof ArrayUnionTransformOperation &&\n arrayEquals(this.elements, other.elements, valueEquals)\n );\n }\n}\n\n/** Transforms an array value via a remove operation. */\nexport class ArrayRemoveTransformOperation implements TransformOperation {\n constructor(readonly elements: api.Value[]) {}\n\n applyToLocalView(\n previousValue: api.Value | null,\n localWriteTime: Timestamp\n ): api.Value {\n return this.apply(previousValue);\n }\n\n applyToRemoteDocument(\n previousValue: api.Value | null,\n transformResult: api.Value | null\n ): api.Value {\n // The server just sends null as the transform result for array operations,\n // so we have to calculate a result the same as we do for local\n // applications.\n return this.apply(previousValue);\n }\n\n private apply(previousValue: api.Value | null): api.Value {\n let values = coercedFieldValuesArray(previousValue);\n for (const toRemove of this.elements) {\n values = values.filter(element => !valueEquals(element, toRemove));\n }\n return { arrayValue: { values } };\n }\n\n computeBaseValue(previousValue: api.Value | null): api.Value | null {\n return null; // Array transforms are idempotent and don't require a base value.\n }\n\n isEqual(other: TransformOperation): boolean {\n return (\n other instanceof ArrayRemoveTransformOperation &&\n arrayEquals(this.elements, other.elements, valueEquals)\n );\n }\n}\n\n/**\n * Implements the backend semantics for locally computed NUMERIC_ADD (increment)\n * transforms. Converts all field values to integers or doubles, but unlike the\n * backend does not cap integer values at 2^63. Instead, JavaScript number\n * arithmetic is used and precision loss can occur for values greater than 2^53.\n */\nexport class NumericIncrementTransformOperation implements TransformOperation {\n constructor(\n private readonly serializer: JsonProtoSerializer,\n readonly operand: api.Value\n ) {\n debugAssert(\n isNumber(operand),\n 'NumericIncrementTransform transform requires a NumberValue'\n );\n }\n\n applyToLocalView(\n previousValue: api.Value | null,\n localWriteTime: Timestamp\n ): api.Value {\n // PORTING NOTE: Since JavaScript's integer arithmetic is limited to 53 bit\n // precision and resolves overflows by reducing precision, we do not\n // manually cap overflows at 2^63.\n const baseValue = this.computeBaseValue(previousValue);\n const sum = this.asNumber(baseValue) + this.asNumber(this.operand);\n if (isInteger(baseValue) && isInteger(this.operand)) {\n return this.serializer.toInteger(sum);\n } else {\n return this.serializer.toDouble(sum);\n }\n }\n\n applyToRemoteDocument(\n previousValue: api.Value | null,\n transformResult: api.Value | null\n ): api.Value {\n debugAssert(\n transformResult !== null,\n \"Didn't receive transformResult for NUMERIC_ADD transform\"\n );\n return transformResult;\n }\n\n /**\n * Inspects the provided value, returning the provided value if it is already\n * a NumberValue, otherwise returning a coerced value of 0.\n */\n computeBaseValue(previousValue: api.Value | null): api.Value {\n return isNumber(previousValue) ? previousValue! : { integerValue: 0 };\n }\n\n isEqual(other: TransformOperation): boolean {\n return (\n other instanceof NumericIncrementTransformOperation &&\n valueEquals(this.operand, other.operand)\n );\n }\n\n private asNumber(value: api.Value): number {\n return normalizeNumber(value.integerValue || value.doubleValue);\n }\n}\n\nfunction coercedFieldValuesArray(value: api.Value | null): api.Value[] {\n return isArray(value) && value.arrayValue.values\n ? value.arrayValue.values.slice()\n : [];\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { Timestamp } from '../api/timestamp';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { debugAssert, fail, hardAssert } from '../util/assert';\n\nimport {\n Document,\n MaybeDocument,\n NoDocument,\n UnknownDocument\n} from './document';\nimport { DocumentKey } from './document_key';\nimport { ObjectValue, ObjectValueBuilder } from './object_value';\nimport { FieldPath } from './path';\nimport { TransformOperation } from './transform_operation';\nimport { arrayEquals } from '../util/misc';\n\n/**\n * Provides a set of fields that can be used to partially patch a document.\n * FieldMask is used in conjunction with ObjectValue.\n * Examples:\n * foo - Overwrites foo entirely with the provided value. If foo is not\n * present in the companion ObjectValue, the field is deleted.\n * foo.bar - Overwrites only the field bar of the object foo.\n * If foo is not an object, foo is replaced with an object\n * containing foo\n */\nexport class FieldMask {\n constructor(readonly fields: FieldPath[]) {\n // TODO(dimond): validation of FieldMask\n // Sort the field mask to support `FieldMask.isEqual()` and assert below.\n fields.sort(FieldPath.comparator);\n debugAssert(\n !fields.some((v, i) => i !== 0 && v.isEqual(fields[i - 1])),\n 'FieldMask contains field that is not unique: ' +\n fields.find((v, i) => i !== 0 && v.isEqual(fields[i - 1]))!\n );\n }\n\n /**\n * Verifies that `fieldPath` is included by at least one field in this field\n * mask.\n *\n * This is an O(n) operation, where `n` is the size of the field mask.\n */\n covers(fieldPath: FieldPath): boolean {\n for (const fieldMaskPath of this.fields) {\n if (fieldMaskPath.isPrefixOf(fieldPath)) {\n return true;\n }\n }\n return false;\n }\n\n isEqual(other: FieldMask): boolean {\n return arrayEquals(this.fields, other.fields, (l, r) => l.isEqual(r));\n }\n}\n\n/** A field path and the TransformOperation to perform upon it. */\nexport class FieldTransform {\n constructor(\n readonly field: FieldPath,\n readonly transform: TransformOperation\n ) {}\n\n isEqual(other: FieldTransform): boolean {\n return (\n this.field.isEqual(other.field) && this.transform.isEqual(other.transform)\n );\n }\n}\n\n/** The result of successfully applying a mutation to the backend. */\nexport class MutationResult {\n constructor(\n /**\n * The version at which the mutation was committed:\n *\n * - For most operations, this is the updateTime in the WriteResult.\n * - For deletes, the commitTime of the WriteResponse (because deletes are\n * not stored and have no updateTime).\n *\n * Note that these versions can be different: No-op writes will not change\n * the updateTime even though the commitTime advances.\n */\n readonly version: SnapshotVersion,\n /**\n * The resulting fields returned from the backend after a\n * TransformMutation has been committed. Contains one FieldValue for each\n * FieldTransform that was in the mutation.\n *\n * Will be null if the mutation was not a TransformMutation.\n */\n readonly transformResults: Array<api.Value | null> | null\n ) {}\n}\n\nexport const enum MutationType {\n Set,\n Patch,\n Transform,\n Delete,\n Verify\n}\n\n/**\n * Encodes a precondition for a mutation. This follows the model that the\n * backend accepts with the special case of an explicit \"empty\" precondition\n * (meaning no precondition).\n */\nexport class Precondition {\n private constructor(\n readonly updateTime?: SnapshotVersion,\n readonly exists?: boolean\n ) {\n debugAssert(\n updateTime === undefined || exists === undefined,\n 'Precondition can specify \"exists\" or \"updateTime\" but not both'\n );\n }\n\n /** Creates a new empty Precondition. */\n static none(): Precondition {\n return new Precondition();\n }\n\n /** Creates a new Precondition with an exists flag. */\n static exists(exists: boolean): Precondition {\n return new Precondition(undefined, exists);\n }\n\n /** Creates a new Precondition based on a version a document exists at. */\n static updateTime(version: SnapshotVersion): Precondition {\n return new Precondition(version);\n }\n\n /** Returns whether this Precondition is empty. */\n get isNone(): boolean {\n return this.updateTime === undefined && this.exists === undefined;\n }\n\n /**\n * Returns true if the preconditions is valid for the given document\n * (or null if no document is available).\n */\n isValidFor(maybeDoc: MaybeDocument | null): boolean {\n if (this.updateTime !== undefined) {\n return (\n maybeDoc instanceof Document &&\n maybeDoc.version.isEqual(this.updateTime)\n );\n } else if (this.exists !== undefined) {\n return this.exists === maybeDoc instanceof Document;\n } else {\n debugAssert(this.isNone, 'Precondition should be empty');\n return true;\n }\n }\n\n isEqual(other: Precondition): boolean {\n return (\n this.exists === other.exists &&\n (this.updateTime\n ? !!other.updateTime && this.updateTime.isEqual(other.updateTime)\n : !other.updateTime)\n );\n }\n}\n\n/**\n * A mutation describes a self-contained change to a document. Mutations can\n * create, replace, delete, and update subsets of documents.\n *\n * Mutations not only act on the value of the document but also its version.\n *\n * For local mutations (mutations that haven't been committed yet), we preserve\n * the existing version for Set, Patch, and Transform mutations. For Delete\n * mutations, we reset the version to 0.\n *\n * Here's the expected transition table.\n *\n * MUTATION APPLIED TO RESULTS IN\n *\n * SetMutation Document(v3) Document(v3)\n * SetMutation NoDocument(v3) Document(v0)\n * SetMutation null Document(v0)\n * PatchMutation Document(v3) Document(v3)\n * PatchMutation NoDocument(v3) NoDocument(v3)\n * PatchMutation null null\n * TransformMutation Document(v3) Document(v3)\n * TransformMutation NoDocument(v3) NoDocument(v3)\n * TransformMutation null null\n * DeleteMutation Document(v3) NoDocument(v0)\n * DeleteMutation NoDocument(v3) NoDocument(v0)\n * DeleteMutation null NoDocument(v0)\n *\n * For acknowledged mutations, we use the updateTime of the WriteResponse as\n * the resulting version for Set, Patch, and Transform mutations. As deletes\n * have no explicit update time, we use the commitTime of the WriteResponse for\n * Delete mutations.\n *\n * If a mutation is acknowledged by the backend but fails the precondition check\n * locally, we return an `UnknownDocument` and rely on Watch to send us the\n * updated version.\n *\n * Note that TransformMutations don't create Documents (in the case of being\n * applied to a NoDocument), even though they would on the backend. This is\n * because the client always combines the TransformMutation with a SetMutation\n * or PatchMutation and we only want to apply the transform if the prior\n * mutation resulted in a Document (always true for a SetMutation, but not\n * necessarily for a PatchMutation).\n *\n * ## Subclassing Notes\n *\n * Subclasses of Mutation need to implement applyToRemoteDocument() and\n * applyToLocalView() to implement the actual behavior of applying the mutation\n * to some source document.\n */\nexport abstract class Mutation {\n abstract readonly type: MutationType;\n abstract readonly key: DocumentKey;\n abstract readonly precondition: Precondition;\n\n /**\n * Applies this mutation to the given MaybeDocument or null for the purposes\n * of computing a new remote document. If the input document doesn't match the\n * expected state (e.g. it is null or outdated), an `UnknownDocument` can be\n * returned.\n *\n * @param maybeDoc The document to mutate. The input document can be null if\n * the client has no knowledge of the pre-mutation state of the document.\n * @param mutationResult The result of applying the mutation from the backend.\n * @return The mutated document. The returned document may be an\n * UnknownDocument if the mutation could not be applied to the locally\n * cached base document.\n */\n abstract applyToRemoteDocument(\n maybeDoc: MaybeDocument | null,\n mutationResult: MutationResult\n ): MaybeDocument;\n\n /**\n * Applies this mutation to the given MaybeDocument or null for the purposes\n * of computing the new local view of a document. Both the input and returned\n * documents can be null.\n *\n * @param maybeDoc The document to mutate. The input document can be null if\n * the client has no knowledge of the pre-mutation state of the document.\n * @param baseDoc The state of the document prior to this mutation batch. The\n * input document can be null if the client has no knowledge of the\n * pre-mutation state of the document.\n * @param localWriteTime A timestamp indicating the local write time of the\n * batch this mutation is a part of.\n * @return The mutated document. The returned document may be null, but only\n * if maybeDoc was null and the mutation would not create a new document.\n */\n abstract applyToLocalView(\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null,\n localWriteTime: Timestamp\n ): MaybeDocument | null;\n\n /**\n * If this mutation is not idempotent, returns the base value to persist with\n * this mutation. If a base value is returned, the mutation is always applied\n * to this base value, even if document has already been updated.\n *\n * The base value is a sparse object that consists of only the document\n * fields for which this mutation contains a non-idempotent transformation\n * (e.g. a numeric increment). The provided value guarantees consistent\n * behavior for non-idempotent transforms and allow us to return the same\n * latency-compensated value even if the backend has already applied the\n * mutation. The base value is null for idempotent mutations, as they can be\n * re-played even if the backend has already applied them.\n *\n * @return a base value to store along with the mutation, or null for\n * idempotent mutations.\n */\n abstract extractBaseValue(maybeDoc: MaybeDocument | null): ObjectValue | null;\n\n abstract isEqual(other: Mutation): boolean;\n\n protected verifyKeyMatches(maybeDoc: MaybeDocument | null): void {\n if (maybeDoc != null) {\n debugAssert(\n maybeDoc.key.isEqual(this.key),\n 'Can only apply a mutation to a document with the same key'\n );\n }\n }\n\n /**\n * Returns the version from the given document for use as the result of a\n * mutation. Mutations are defined to return the version of the base document\n * only if it is an existing document. Deleted and unknown documents have a\n * post-mutation version of SnapshotVersion.min().\n */\n protected static getPostMutationVersion(\n maybeDoc: MaybeDocument | null\n ): SnapshotVersion {\n if (maybeDoc instanceof Document) {\n return maybeDoc.version;\n } else {\n return SnapshotVersion.min();\n }\n }\n}\n\n/**\n * A mutation that creates or replaces the document at the given key with the\n * object value contents.\n */\nexport class SetMutation extends Mutation {\n constructor(\n readonly key: DocumentKey,\n readonly value: ObjectValue,\n readonly precondition: Precondition\n ) {\n super();\n }\n\n readonly type: MutationType = MutationType.Set;\n\n applyToRemoteDocument(\n maybeDoc: MaybeDocument | null,\n mutationResult: MutationResult\n ): MaybeDocument {\n this.verifyKeyMatches(maybeDoc);\n\n debugAssert(\n mutationResult.transformResults == null,\n 'Transform results received by SetMutation.'\n );\n\n // Unlike applyToLocalView, if we're applying a mutation to a remote\n // document the server has accepted the mutation so the precondition must\n // have held.\n\n const version = mutationResult.version;\n return new Document(this.key, version, this.value, {\n hasCommittedMutations: true\n });\n }\n\n applyToLocalView(\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null,\n localWriteTime: Timestamp\n ): MaybeDocument | null {\n this.verifyKeyMatches(maybeDoc);\n\n if (!this.precondition.isValidFor(maybeDoc)) {\n return maybeDoc;\n }\n\n const version = Mutation.getPostMutationVersion(maybeDoc);\n return new Document(this.key, version, this.value, {\n hasLocalMutations: true\n });\n }\n\n extractBaseValue(maybeDoc: MaybeDocument | null): null {\n return null;\n }\n\n isEqual(other: Mutation): boolean {\n return (\n other instanceof SetMutation &&\n this.key.isEqual(other.key) &&\n this.value.isEqual(other.value) &&\n this.precondition.isEqual(other.precondition)\n );\n }\n}\n\n/**\n * A mutation that modifies fields of the document at the given key with the\n * given values. The values are applied through a field mask:\n *\n * * When a field is in both the mask and the values, the corresponding field\n * is updated.\n * * When a field is in neither the mask nor the values, the corresponding\n * field is unmodified.\n * * When a field is in the mask but not in the values, the corresponding field\n * is deleted.\n * * When a field is not in the mask but is in the values, the values map is\n * ignored.\n */\nexport class PatchMutation extends Mutation {\n constructor(\n readonly key: DocumentKey,\n readonly data: ObjectValue,\n readonly fieldMask: FieldMask,\n readonly precondition: Precondition\n ) {\n super();\n }\n\n readonly type: MutationType = MutationType.Patch;\n\n applyToRemoteDocument(\n maybeDoc: MaybeDocument | null,\n mutationResult: MutationResult\n ): MaybeDocument {\n this.verifyKeyMatches(maybeDoc);\n\n debugAssert(\n mutationResult.transformResults == null,\n 'Transform results received by PatchMutation.'\n );\n\n if (!this.precondition.isValidFor(maybeDoc)) {\n // Since the mutation was not rejected, we know that the precondition\n // matched on the backend. We therefore must not have the expected version\n // of the document in our cache and return an UnknownDocument with the\n // known updateTime.\n return new UnknownDocument(this.key, mutationResult.version);\n }\n\n const newData = this.patchDocument(maybeDoc);\n return new Document(this.key, mutationResult.version, newData, {\n hasCommittedMutations: true\n });\n }\n\n applyToLocalView(\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null,\n localWriteTime: Timestamp\n ): MaybeDocument | null {\n this.verifyKeyMatches(maybeDoc);\n\n if (!this.precondition.isValidFor(maybeDoc)) {\n return maybeDoc;\n }\n\n const version = Mutation.getPostMutationVersion(maybeDoc);\n const newData = this.patchDocument(maybeDoc);\n return new Document(this.key, version, newData, {\n hasLocalMutations: true\n });\n }\n\n extractBaseValue(maybeDoc: MaybeDocument | null): null {\n return null;\n }\n\n isEqual(other: Mutation): boolean {\n return (\n other instanceof PatchMutation &&\n this.key.isEqual(other.key) &&\n this.fieldMask.isEqual(other.fieldMask) &&\n this.precondition.isEqual(other.precondition)\n );\n }\n\n /**\n * Patches the data of document if available or creates a new document. Note\n * that this does not check whether or not the precondition of this patch\n * holds.\n */\n private patchDocument(maybeDoc: MaybeDocument | null): ObjectValue {\n let data: ObjectValue;\n if (maybeDoc instanceof Document) {\n data = maybeDoc.data();\n } else {\n data = ObjectValue.empty();\n }\n return this.patchObject(data);\n }\n\n private patchObject(data: ObjectValue): ObjectValue {\n const builder = new ObjectValueBuilder(data);\n this.fieldMask.fields.forEach(fieldPath => {\n if (!fieldPath.isEmpty()) {\n const newValue = this.data.field(fieldPath);\n if (newValue !== null) {\n builder.set(fieldPath, newValue);\n } else {\n builder.delete(fieldPath);\n }\n }\n });\n return builder.build();\n }\n}\n\n/**\n * A mutation that modifies specific fields of the document with transform\n * operations. Currently the only supported transform is a server timestamp, but\n * IP Address, increment(n), etc. could be supported in the future.\n *\n * It is somewhat similar to a PatchMutation in that it patches specific fields\n * and has no effect when applied to a null or NoDocument (see comment on\n * Mutation for rationale).\n */\nexport class TransformMutation extends Mutation {\n readonly type: MutationType = MutationType.Transform;\n\n // NOTE: We set a precondition of exists: true as a safety-check, since we\n // always combine TransformMutations with a SetMutation or PatchMutation which\n // (if successful) should end up with an existing document.\n readonly precondition = Precondition.exists(true);\n\n constructor(\n readonly key: DocumentKey,\n readonly fieldTransforms: FieldTransform[]\n ) {\n super();\n }\n\n applyToRemoteDocument(\n maybeDoc: MaybeDocument | null,\n mutationResult: MutationResult\n ): MaybeDocument {\n this.verifyKeyMatches(maybeDoc);\n\n hardAssert(\n mutationResult.transformResults != null,\n 'Transform results missing for TransformMutation.'\n );\n\n if (!this.precondition.isValidFor(maybeDoc)) {\n // Since the mutation was not rejected, we know that the precondition\n // matched on the backend. We therefore must not have the expected version\n // of the document in our cache and return an UnknownDocument with the\n // known updateTime.\n return new UnknownDocument(this.key, mutationResult.version);\n }\n\n const doc = this.requireDocument(maybeDoc);\n const transformResults = this.serverTransformResults(\n maybeDoc,\n mutationResult.transformResults!\n );\n\n const version = mutationResult.version;\n const newData = this.transformObject(doc.data(), transformResults);\n return new Document(this.key, version, newData, {\n hasCommittedMutations: true\n });\n }\n\n applyToLocalView(\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null,\n localWriteTime: Timestamp\n ): MaybeDocument | null {\n this.verifyKeyMatches(maybeDoc);\n\n if (!this.precondition.isValidFor(maybeDoc)) {\n return maybeDoc;\n }\n\n const doc = this.requireDocument(maybeDoc);\n const transformResults = this.localTransformResults(\n localWriteTime,\n maybeDoc,\n baseDoc\n );\n const newData = this.transformObject(doc.data(), transformResults);\n return new Document(this.key, doc.version, newData, {\n hasLocalMutations: true\n });\n }\n\n extractBaseValue(maybeDoc: MaybeDocument | null): ObjectValue | null {\n let baseObject: ObjectValueBuilder | null = null;\n for (const fieldTransform of this.fieldTransforms) {\n const existingValue =\n maybeDoc instanceof Document\n ? maybeDoc.field(fieldTransform.field)\n : undefined;\n const coercedValue = fieldTransform.transform.computeBaseValue(\n existingValue || null\n );\n\n if (coercedValue != null) {\n if (baseObject == null) {\n baseObject = new ObjectValueBuilder().set(\n fieldTransform.field,\n coercedValue\n );\n } else {\n baseObject = baseObject.set(fieldTransform.field, coercedValue);\n }\n }\n }\n return baseObject ? baseObject.build() : null;\n }\n\n isEqual(other: Mutation): boolean {\n return (\n other instanceof TransformMutation &&\n this.key.isEqual(other.key) &&\n arrayEquals(this.fieldTransforms, other.fieldTransforms, (l, r) =>\n l.isEqual(r)\n ) &&\n this.precondition.isEqual(other.precondition)\n );\n }\n\n /**\n * Asserts that the given MaybeDocument is actually a Document and verifies\n * that it matches the key for this mutation. Since we only support\n * transformations with precondition exists this method is guaranteed to be\n * safe.\n */\n private requireDocument(maybeDoc: MaybeDocument | null): Document {\n debugAssert(\n maybeDoc instanceof Document,\n 'Unknown MaybeDocument type ' + maybeDoc\n );\n debugAssert(\n maybeDoc.key.isEqual(this.key),\n 'Can only transform a document with the same key'\n );\n return maybeDoc;\n }\n\n /**\n * Creates a list of \"transform results\" (a transform result is a field value\n * representing the result of applying a transform) for use after a\n * TransformMutation has been acknowledged by the server.\n *\n * @param baseDoc The document prior to applying this mutation batch.\n * @param serverTransformResults The transform results received by the server.\n * @return The transform results list.\n */\n private serverTransformResults(\n baseDoc: MaybeDocument | null,\n serverTransformResults: Array<api.Value | null>\n ): api.Value[] {\n const transformResults: api.Value[] = [];\n hardAssert(\n this.fieldTransforms.length === serverTransformResults.length,\n `server transform result count (${serverTransformResults.length}) ` +\n `should match field transform count (${this.fieldTransforms.length})`\n );\n\n for (let i = 0; i < serverTransformResults.length; i++) {\n const fieldTransform = this.fieldTransforms[i];\n const transform = fieldTransform.transform;\n let previousValue: api.Value | null = null;\n if (baseDoc instanceof Document) {\n previousValue = baseDoc.field(fieldTransform.field);\n }\n transformResults.push(\n transform.applyToRemoteDocument(\n previousValue,\n serverTransformResults[i]\n )\n );\n }\n return transformResults;\n }\n\n /**\n * Creates a list of \"transform results\" (a transform result is a field value\n * representing the result of applying a transform) for use when applying a\n * TransformMutation locally.\n *\n * @param localWriteTime The local time of the transform mutation (used to\n * generate ServerTimestampValues).\n * @param maybeDoc The current state of the document after applying all\n * previous mutations.\n * @param baseDoc The document prior to applying this mutation batch.\n * @return The transform results list.\n */\n private localTransformResults(\n localWriteTime: Timestamp,\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null\n ): api.Value[] {\n const transformResults: api.Value[] = [];\n for (const fieldTransform of this.fieldTransforms) {\n const transform = fieldTransform.transform;\n\n let previousValue: api.Value | null = null;\n if (maybeDoc instanceof Document) {\n previousValue = maybeDoc.field(fieldTransform.field);\n }\n\n if (previousValue === null && baseDoc instanceof Document) {\n // If the current document does not contain a value for the mutated\n // field, use the value that existed before applying this mutation\n // batch. This solves an edge case where a PatchMutation clears the\n // values in a nested map before the TransformMutation is applied.\n previousValue = baseDoc.field(fieldTransform.field);\n }\n\n transformResults.push(\n transform.applyToLocalView(previousValue, localWriteTime)\n );\n }\n return transformResults;\n }\n\n private transformObject(\n data: ObjectValue,\n transformResults: api.Value[]\n ): ObjectValue {\n debugAssert(\n transformResults.length === this.fieldTransforms.length,\n 'TransformResults length mismatch.'\n );\n\n const builder = new ObjectValueBuilder(data);\n for (let i = 0; i < this.fieldTransforms.length; i++) {\n const fieldTransform = this.fieldTransforms[i];\n const fieldPath = fieldTransform.field;\n builder.set(fieldPath, transformResults[i]);\n }\n return builder.build();\n }\n}\n\n/** A mutation that deletes the document at the given key. */\nexport class DeleteMutation extends Mutation {\n constructor(readonly key: DocumentKey, readonly precondition: Precondition) {\n super();\n }\n\n readonly type: MutationType = MutationType.Delete;\n\n applyToRemoteDocument(\n maybeDoc: MaybeDocument | null,\n mutationResult: MutationResult\n ): MaybeDocument {\n this.verifyKeyMatches(maybeDoc);\n\n debugAssert(\n mutationResult.transformResults == null,\n 'Transform results received by DeleteMutation.'\n );\n\n // Unlike applyToLocalView, if we're applying a mutation to a remote\n // document the server has accepted the mutation so the precondition must\n // have held.\n\n return new NoDocument(this.key, mutationResult.version, {\n hasCommittedMutations: true\n });\n }\n\n applyToLocalView(\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null,\n localWriteTime: Timestamp\n ): MaybeDocument | null {\n this.verifyKeyMatches(maybeDoc);\n\n if (!this.precondition.isValidFor(maybeDoc)) {\n return maybeDoc;\n }\n\n if (maybeDoc) {\n debugAssert(\n maybeDoc.key.isEqual(this.key),\n 'Can only apply mutation to document with same key'\n );\n }\n return new NoDocument(this.key, SnapshotVersion.min());\n }\n\n extractBaseValue(maybeDoc: MaybeDocument | null): null {\n return null;\n }\n\n isEqual(other: Mutation): boolean {\n return (\n other instanceof DeleteMutation &&\n this.key.isEqual(other.key) &&\n this.precondition.isEqual(other.precondition)\n );\n }\n}\n\n/**\n * A mutation that verifies the existence of the document at the given key with\n * the provided precondition.\n *\n * The `verify` operation is only used in Transactions, and this class serves\n * primarily to facilitate serialization into protos.\n */\nexport class VerifyMutation extends Mutation {\n constructor(readonly key: DocumentKey, readonly precondition: Precondition) {\n super();\n }\n\n readonly type: MutationType = MutationType.Verify;\n\n applyToRemoteDocument(\n maybeDoc: MaybeDocument | null,\n mutationResult: MutationResult\n ): MaybeDocument {\n fail('VerifyMutation should only be used in Transactions.');\n }\n\n applyToLocalView(\n maybeDoc: MaybeDocument | null,\n baseDoc: MaybeDocument | null,\n localWriteTime: Timestamp\n ): MaybeDocument | null {\n fail('VerifyMutation should only be used in Transactions.');\n }\n\n extractBaseValue(maybeDoc: MaybeDocument | null): null {\n fail('VerifyMutation should only be used in Transactions.');\n }\n\n isEqual(other: Mutation): boolean {\n return (\n other instanceof VerifyMutation &&\n this.key.isEqual(other.key) &&\n this.precondition.isEqual(other.precondition)\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { debugAssert } from '../util/assert';\nimport { FieldMask } from './mutation';\nimport { FieldPath } from './path';\nimport { isServerTimestamp } from './server_timestamps';\nimport { valueEquals, isMapValue, typeOrder } from './values';\nimport { forEach } from '../util/obj';\n\nexport interface JsonObject<T> {\n [name: string]: T;\n}\n\nexport const enum TypeOrder {\n // This order is based on the backend's ordering, but modified to support\n // server timestamps.\n NullValue = 0,\n BooleanValue = 1,\n NumberValue = 2,\n TimestampValue = 3,\n ServerTimestampValue = 4,\n StringValue = 5,\n BlobValue = 6,\n RefValue = 7,\n GeoPointValue = 8,\n ArrayValue = 9,\n ObjectValue = 10\n}\n\n/**\n * An ObjectValue represents a MapValue in the Firestore Proto and offers the\n * ability to add and remove fields (via the ObjectValueBuilder).\n */\nexport class ObjectValue {\n constructor(public readonly proto: { mapValue: api.MapValue }) {\n debugAssert(\n !isServerTimestamp(proto),\n 'ServerTimestamps should be converted to ServerTimestampValue'\n );\n }\n\n static empty(): ObjectValue {\n return new ObjectValue({ mapValue: {} });\n }\n\n /**\n * Returns the value at the given path or null.\n *\n * @param path the path to search\n * @return The value at the path or if there it doesn't exist.\n */\n field(path: FieldPath): api.Value | null {\n if (path.isEmpty()) {\n return this.proto;\n } else {\n let value: api.Value = this.proto;\n for (let i = 0; i < path.length - 1; ++i) {\n if (!value.mapValue!.fields) {\n return null;\n }\n value = value.mapValue!.fields[path.get(i)];\n if (!isMapValue(value)) {\n return null;\n }\n }\n\n value = (value.mapValue!.fields || {})[path.lastSegment()];\n return value || null;\n }\n }\n\n isEqual(other: ObjectValue): boolean {\n return valueEquals(this.proto, other.proto);\n }\n}\n\n/**\n * An Overlay, which contains an update to apply. Can either be Value proto, a\n * map of Overlay values (to represent additional nesting at the given key) or\n * `null` (to represent field deletes).\n */\ntype Overlay = Map<string, Overlay> | api.Value | null;\n\n/**\n * An ObjectValueBuilder provides APIs to set and delete fields from an\n * ObjectValue.\n */\nexport class ObjectValueBuilder {\n /** A map that contains the accumulated changes in this builder. */\n private overlayMap = new Map<string, Overlay>();\n\n /**\n * @param baseObject The object to mutate.\n */\n constructor(private readonly baseObject: ObjectValue = ObjectValue.empty()) {}\n\n /**\n * Sets the field to the provided value.\n *\n * @param path The field path to set.\n * @param value The value to set.\n * @return The current Builder instance.\n */\n set(path: FieldPath, value: api.Value): ObjectValueBuilder {\n debugAssert(\n !path.isEmpty(),\n 'Cannot set field for empty path on ObjectValue'\n );\n this.setOverlay(path, value);\n return this;\n }\n\n /**\n * Removes the field at the specified path. If there is no field at the\n * specified path, nothing is changed.\n *\n * @param path The field path to remove.\n * @return The current Builder instance.\n */\n delete(path: FieldPath): ObjectValueBuilder {\n debugAssert(\n !path.isEmpty(),\n 'Cannot delete field for empty path on ObjectValue'\n );\n this.setOverlay(path, null);\n return this;\n }\n\n /**\n * Adds `value` to the overlay map at `path`. Creates nested map entries if\n * needed.\n */\n private setOverlay(path: FieldPath, value: api.Value | null): void {\n let currentLevel = this.overlayMap;\n\n for (let i = 0; i < path.length - 1; ++i) {\n const currentSegment = path.get(i);\n let currentValue = currentLevel.get(currentSegment);\n\n if (currentValue instanceof Map) {\n // Re-use a previously created map\n currentLevel = currentValue;\n } else if (\n currentValue &&\n typeOrder(currentValue) === TypeOrder.ObjectValue\n ) {\n // Convert the existing Protobuf MapValue into a map\n currentValue = new Map<string, Overlay>(\n Object.entries(currentValue.mapValue!.fields || {})\n );\n currentLevel.set(currentSegment, currentValue);\n currentLevel = currentValue;\n } else {\n // Create an empty map to represent the current nesting level\n currentValue = new Map<string, Overlay>();\n currentLevel.set(currentSegment, currentValue);\n currentLevel = currentValue;\n }\n }\n\n currentLevel.set(path.lastSegment(), value);\n }\n\n /** Returns an ObjectValue with all mutations applied. */\n build(): ObjectValue {\n const mergedResult = this.applyOverlay(\n FieldPath.EMPTY_PATH,\n this.overlayMap\n );\n if (mergedResult != null) {\n return new ObjectValue(mergedResult);\n } else {\n return this.baseObject;\n }\n }\n\n /**\n * Applies any overlays from `currentOverlays` that exist at `currentPath`\n * and returns the merged data at `currentPath` (or null if there were no\n * changes).\n *\n * @param currentPath The path at the current nesting level. Can be set to\n * FieldValue.EMPTY_PATH to represent the root.\n * @param currentOverlays The overlays at the current nesting level in the\n * same format as `overlayMap`.\n * @return The merged data at `currentPath` or null if no modifications\n * were applied.\n */\n private applyOverlay(\n currentPath: FieldPath,\n currentOverlays: Map<string, Overlay>\n ): { mapValue: api.MapValue } | null {\n let modified = false;\n\n const existingValue = this.baseObject.field(currentPath);\n const resultAtPath = isMapValue(existingValue)\n ? // If there is already data at the current path, base our\n // modifications on top of the existing data.\n { ...existingValue.mapValue.fields }\n : {};\n\n currentOverlays.forEach((value, pathSegment) => {\n if (value instanceof Map) {\n const nested = this.applyOverlay(currentPath.child(pathSegment), value);\n if (nested != null) {\n resultAtPath[pathSegment] = nested;\n modified = true;\n }\n } else if (value !== null) {\n resultAtPath[pathSegment] = value;\n modified = true;\n } else if (resultAtPath.hasOwnProperty(pathSegment)) {\n delete resultAtPath[pathSegment];\n modified = true;\n }\n });\n\n return modified ? { mapValue: { fields: resultAtPath } } : null;\n }\n}\n\n/**\n * Returns a FieldMask built from all fields in a MapValue.\n */\nexport function extractFieldMask(value: api.MapValue): FieldMask {\n const fields: FieldPath[] = [];\n forEach(value!.fields || {}, (key, value) => {\n const currentPath = new FieldPath([key]);\n if (isMapValue(value)) {\n const nestedMask = extractFieldMask(value.mapValue!);\n const nestedFields = nestedMask.fields;\n if (nestedFields.length === 0) {\n // Preserve the empty map by adding it to the FieldMask.\n fields.push(currentPath);\n } else {\n // For nested and non-empty ObjectValues, add the FieldPath of the\n // leaf nodes.\n for (const nestedPath of nestedFields) {\n fields.push(currentPath.child(nestedPath));\n }\n }\n } else {\n // For nested and non-empty ObjectValues, add the FieldPath of the leaf\n // nodes.\n fields.push(currentPath);\n }\n });\n return new FieldMask(fields);\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { fail } from '../util/assert';\n\nimport { DocumentKey } from './document_key';\nimport { ObjectValue } from './object_value';\nimport { FieldPath } from './path';\nimport { valueCompare } from './values';\n\nexport interface DocumentOptions {\n hasLocalMutations?: boolean;\n hasCommittedMutations?: boolean;\n}\n\n/**\n * The result of a lookup for a given path may be an existing document or a\n * marker that this document does not exist at a given version.\n */\nexport abstract class MaybeDocument {\n constructor(readonly key: DocumentKey, readonly version: SnapshotVersion) {}\n\n /**\n * Whether this document had a local mutation applied that has not yet been\n * acknowledged by Watch.\n */\n abstract get hasPendingWrites(): boolean;\n\n abstract isEqual(other: MaybeDocument | null | undefined): boolean;\n\n abstract toString(): string;\n}\n\n/**\n * Represents a document in Firestore with a key, version, data and whether the\n * data has local mutations applied to it.\n */\nexport class Document extends MaybeDocument {\n readonly hasLocalMutations: boolean;\n readonly hasCommittedMutations: boolean;\n\n constructor(\n key: DocumentKey,\n version: SnapshotVersion,\n private readonly objectValue: ObjectValue,\n options: DocumentOptions\n ) {\n super(key, version);\n this.hasLocalMutations = !!options.hasLocalMutations;\n this.hasCommittedMutations = !!options.hasCommittedMutations;\n }\n\n field(path: FieldPath): api.Value | null {\n return this.objectValue.field(path);\n }\n\n data(): ObjectValue {\n return this.objectValue;\n }\n\n toProto(): { mapValue: api.MapValue } {\n return this.objectValue.proto;\n }\n\n isEqual(other: MaybeDocument | null | undefined): boolean {\n return (\n other instanceof Document &&\n this.key.isEqual(other.key) &&\n this.version.isEqual(other.version) &&\n this.hasLocalMutations === other.hasLocalMutations &&\n this.hasCommittedMutations === other.hasCommittedMutations &&\n this.objectValue.isEqual(other.objectValue)\n );\n }\n\n toString(): string {\n return (\n `Document(${this.key}, ${\n this.version\n }, ${this.objectValue.toString()}, ` +\n `{hasLocalMutations: ${this.hasLocalMutations}}), ` +\n `{hasCommittedMutations: ${this.hasCommittedMutations}})`\n );\n }\n\n get hasPendingWrites(): boolean {\n return this.hasLocalMutations || this.hasCommittedMutations;\n }\n}\n\n/**\n * Compares the value for field `field` in the provided documents. Throws if\n * the field does not exist in both documents.\n */\nexport function compareDocumentsByField(\n field: FieldPath,\n d1: Document,\n d2: Document\n): number {\n const v1 = d1.field(field);\n const v2 = d2.field(field);\n if (v1 !== null && v2 !== null) {\n return valueCompare(v1, v2);\n } else {\n return fail(\"Trying to compare documents on fields that don't exist\");\n }\n}\n\n/**\n * A class representing a deleted document.\n * Version is set to 0 if we don't point to any specific time, otherwise it\n * denotes time we know it didn't exist at.\n */\nexport class NoDocument extends MaybeDocument {\n readonly hasCommittedMutations: boolean;\n\n constructor(\n key: DocumentKey,\n version: SnapshotVersion,\n options?: DocumentOptions\n ) {\n super(key, version);\n this.hasCommittedMutations = !!(options && options.hasCommittedMutations);\n }\n\n toString(): string {\n return `NoDocument(${this.key}, ${this.version})`;\n }\n\n get hasPendingWrites(): boolean {\n return this.hasCommittedMutations;\n }\n\n isEqual(other: MaybeDocument | null | undefined): boolean {\n return (\n other instanceof NoDocument &&\n other.hasCommittedMutations === this.hasCommittedMutations &&\n other.version.isEqual(this.version) &&\n other.key.isEqual(this.key)\n );\n }\n}\n\n/**\n * A class representing an existing document whose data is unknown (e.g. a\n * document that was updated without a known base document).\n */\nexport class UnknownDocument extends MaybeDocument {\n toString(): string {\n return `UnknownDocument(${this.key}, ${this.version})`;\n }\n\n get hasPendingWrites(): boolean {\n return true;\n }\n\n isEqual(other: MaybeDocument | null | undefined): boolean {\n return (\n other instanceof UnknownDocument &&\n other.version.isEqual(this.version) &&\n other.key.isEqual(this.key)\n );\n }\n}\n","/**\n * @license\n * Copyright 2019 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { DocumentKey } from '../model/document_key';\nimport { ResourcePath } from '../model/path';\nimport { isNullOrUndefined } from '../util/types';\nimport { Bound, Filter, OrderBy } from './query';\n\n/**\n * A Target represents the WatchTarget representation of a Query, which is used\n * by the LocalStore and the RemoteStore to keep track of and to execute\n * backend queries. While a Query can represent multiple Targets, each Targets\n * maps to a single WatchTarget in RemoteStore and a single TargetData entry\n * in persistence.\n */\nexport class Target {\n private memoizedCanonicalId: string | null = null;\n\n /**\n * Initializes a Target with a path and optional additional query constraints.\n * Path must currently be empty if this is a collection group query.\n *\n * NOTE: you should always construct `Target` from `Query.toTarget` instead of\n * using this constructor, because `Query` provides an implicit `orderBy`\n * property.\n */\n constructor(\n readonly path: ResourcePath,\n readonly collectionGroup: string | null = null,\n readonly orderBy: OrderBy[] = [],\n readonly filters: Filter[] = [],\n readonly limit: number | null = null,\n readonly startAt: Bound | null = null,\n readonly endAt: Bound | null = null\n ) {}\n\n canonicalId(): string {\n if (this.memoizedCanonicalId === null) {\n let canonicalId = this.path.canonicalString();\n if (this.collectionGroup !== null) {\n canonicalId += '|cg:' + this.collectionGroup;\n }\n canonicalId += '|f:';\n canonicalId += this.filters.map(f => f.canonicalId()).join(',');\n canonicalId += '|ob:';\n canonicalId += this.orderBy.map(o => o.canonicalId()).join(',');\n\n if (!isNullOrUndefined(this.limit)) {\n canonicalId += '|l:';\n canonicalId += this.limit!;\n }\n if (this.startAt) {\n canonicalId += '|lb:';\n canonicalId += this.startAt.canonicalId();\n }\n if (this.endAt) {\n canonicalId += '|ub:';\n canonicalId += this.endAt.canonicalId();\n }\n this.memoizedCanonicalId = canonicalId;\n }\n return this.memoizedCanonicalId;\n }\n\n toString(): string {\n let str = this.path.canonicalString();\n if (this.collectionGroup !== null) {\n str += ' collectionGroup=' + this.collectionGroup;\n }\n if (this.filters.length > 0) {\n str += `, filters: [${this.filters.join(', ')}]`;\n }\n if (!isNullOrUndefined(this.limit)) {\n str += ', limit: ' + this.limit;\n }\n if (this.orderBy.length > 0) {\n str += `, orderBy: [${this.orderBy.join(', ')}]`;\n }\n if (this.startAt) {\n str += ', startAt: ' + this.startAt.canonicalId();\n }\n if (this.endAt) {\n str += ', endAt: ' + this.endAt.canonicalId();\n }\n return `Target(${str})`;\n }\n\n isEqual(other: Target): boolean {\n if (this.limit !== other.limit) {\n return false;\n }\n\n if (this.orderBy.length !== other.orderBy.length) {\n return false;\n }\n\n for (let i = 0; i < this.orderBy.length; i++) {\n if (!this.orderBy[i].isEqual(other.orderBy[i])) {\n return false;\n }\n }\n\n if (this.filters.length !== other.filters.length) {\n return false;\n }\n\n for (let i = 0; i < this.filters.length; i++) {\n if (!this.filters[i].isEqual(other.filters[i])) {\n return false;\n }\n }\n\n if (this.collectionGroup !== other.collectionGroup) {\n return false;\n }\n\n if (!this.path.isEqual(other.path)) {\n return false;\n }\n\n if (\n this.startAt !== null\n ? !this.startAt.isEqual(other.startAt)\n : other.startAt !== null\n ) {\n return false;\n }\n\n return this.endAt !== null\n ? this.endAt.isEqual(other.endAt)\n : other.endAt === null;\n }\n\n isDocumentQuery(): boolean {\n return (\n DocumentKey.isDocumentKey(this.path) &&\n this.collectionGroup === null &&\n this.filters.length === 0\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { compareDocumentsByField, Document } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport {\n canonicalId,\n valueCompare,\n arrayValueContains,\n valueEquals,\n isArray,\n isNanValue,\n isNullValue,\n isReferenceValue,\n typeOrder\n} from '../model/values';\nimport { FieldPath, ResourcePath } from '../model/path';\nimport { debugAssert, fail } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { isNullOrUndefined } from '../util/types';\nimport { Target } from './target';\n\nexport const enum LimitType {\n First = 'F',\n Last = 'L'\n}\n\n/**\n * Query encapsulates all the query attributes we support in the SDK. It can\n * be run against the LocalStore, as well as be converted to a `Target` to\n * query the RemoteStore results.\n */\nexport class Query {\n static atPath(path: ResourcePath): Query {\n return new Query(path);\n }\n\n private memoizedOrderBy: OrderBy[] | null = null;\n\n // The corresponding `Target` of this `Query` instance.\n private memoizedTarget: Target | null = null;\n\n /**\n * Initializes a Query with a path and optional additional query constraints.\n * Path must currently be empty if this is a collection group query.\n */\n constructor(\n readonly path: ResourcePath,\n readonly collectionGroup: string | null = null,\n readonly explicitOrderBy: OrderBy[] = [],\n readonly filters: Filter[] = [],\n readonly limit: number | null = null,\n readonly limitType: LimitType = LimitType.First,\n readonly startAt: Bound | null = null,\n readonly endAt: Bound | null = null\n ) {\n if (this.startAt) {\n this.assertValidBound(this.startAt);\n }\n if (this.endAt) {\n this.assertValidBound(this.endAt);\n }\n }\n\n get orderBy(): OrderBy[] {\n if (this.memoizedOrderBy === null) {\n this.memoizedOrderBy = [];\n\n const inequalityField = this.getInequalityFilterField();\n const firstOrderByField = this.getFirstOrderByField();\n if (inequalityField !== null && firstOrderByField === null) {\n // In order to implicitly add key ordering, we must also add the\n // inequality filter field for it to be a valid query.\n // Note that the default inequality field and key ordering is ascending.\n if (!inequalityField.isKeyField()) {\n this.memoizedOrderBy.push(new OrderBy(inequalityField));\n }\n this.memoizedOrderBy.push(\n new OrderBy(FieldPath.keyField(), Direction.ASCENDING)\n );\n } else {\n debugAssert(\n inequalityField === null ||\n (firstOrderByField !== null &&\n inequalityField.isEqual(firstOrderByField)),\n 'First orderBy should match inequality field.'\n );\n let foundKeyOrdering = false;\n for (const orderBy of this.explicitOrderBy) {\n this.memoizedOrderBy.push(orderBy);\n if (orderBy.field.isKeyField()) {\n foundKeyOrdering = true;\n }\n }\n if (!foundKeyOrdering) {\n // The order of the implicit key ordering always matches the last\n // explicit order by\n const lastDirection =\n this.explicitOrderBy.length > 0\n ? this.explicitOrderBy[this.explicitOrderBy.length - 1].dir\n : Direction.ASCENDING;\n this.memoizedOrderBy.push(\n new OrderBy(FieldPath.keyField(), lastDirection)\n );\n }\n }\n }\n return this.memoizedOrderBy;\n }\n\n addFilter(filter: Filter): Query {\n debugAssert(\n this.getInequalityFilterField() == null ||\n !(filter instanceof FieldFilter) ||\n !filter.isInequality() ||\n filter.field.isEqual(this.getInequalityFilterField()!),\n 'Query must only have one inequality field.'\n );\n\n debugAssert(\n !this.isDocumentQuery(),\n 'No filtering allowed for document query'\n );\n\n const newFilters = this.filters.concat([filter]);\n return new Query(\n this.path,\n this.collectionGroup,\n this.explicitOrderBy.slice(),\n newFilters,\n this.limit,\n this.limitType,\n this.startAt,\n this.endAt\n );\n }\n\n addOrderBy(orderBy: OrderBy): Query {\n debugAssert(\n !this.startAt && !this.endAt,\n 'Bounds must be set after orderBy'\n );\n // TODO(dimond): validate that orderBy does not list the same key twice.\n const newOrderBy = this.explicitOrderBy.concat([orderBy]);\n return new Query(\n this.path,\n this.collectionGroup,\n newOrderBy,\n this.filters.slice(),\n this.limit,\n this.limitType,\n this.startAt,\n this.endAt\n );\n }\n\n withLimitToFirst(limit: number | null): Query {\n return new Query(\n this.path,\n this.collectionGroup,\n this.explicitOrderBy.slice(),\n this.filters.slice(),\n limit,\n LimitType.First,\n this.startAt,\n this.endAt\n );\n }\n\n withLimitToLast(limit: number | null): Query {\n return new Query(\n this.path,\n this.collectionGroup,\n this.explicitOrderBy.slice(),\n this.filters.slice(),\n limit,\n LimitType.Last,\n this.startAt,\n this.endAt\n );\n }\n\n withStartAt(bound: Bound): Query {\n return new Query(\n this.path,\n this.collectionGroup,\n this.explicitOrderBy.slice(),\n this.filters.slice(),\n this.limit,\n this.limitType,\n bound,\n this.endAt\n );\n }\n\n withEndAt(bound: Bound): Query {\n return new Query(\n this.path,\n this.collectionGroup,\n this.explicitOrderBy.slice(),\n this.filters.slice(),\n this.limit,\n this.limitType,\n this.startAt,\n bound\n );\n }\n\n /**\n * Helper to convert a collection group query into a collection query at a\n * specific path. This is used when executing collection group queries, since\n * we have to split the query into a set of collection queries at multiple\n * paths.\n */\n asCollectionQueryAtPath(path: ResourcePath): Query {\n return new Query(\n path,\n /*collectionGroup=*/ null,\n this.explicitOrderBy.slice(),\n this.filters.slice(),\n this.limit,\n this.limitType,\n this.startAt,\n this.endAt\n );\n }\n\n /**\n * Returns true if this query does not specify any query constraints that\n * could remove results.\n */\n matchesAllDocuments(): boolean {\n return (\n this.filters.length === 0 &&\n this.limit === null &&\n this.startAt == null &&\n this.endAt == null &&\n (this.explicitOrderBy.length === 0 ||\n (this.explicitOrderBy.length === 1 &&\n this.explicitOrderBy[0].field.isKeyField()))\n );\n }\n\n // TODO(b/29183165): This is used to get a unique string from a query to, for\n // example, use as a dictionary key, but the implementation is subject to\n // collisions. Make it collision-free.\n canonicalId(): string {\n return `${this.toTarget().canonicalId()}|lt:${this.limitType}`;\n }\n\n toString(): string {\n return `Query(target=${this.toTarget().toString()}; limitType=${\n this.limitType\n })`;\n }\n\n isEqual(other: Query): boolean {\n return (\n this.toTarget().isEqual(other.toTarget()) &&\n this.limitType === other.limitType\n );\n }\n\n docComparator(d1: Document, d2: Document): number {\n let comparedOnKeyField = false;\n for (const orderBy of this.orderBy) {\n const comp = orderBy.compare(d1, d2);\n if (comp !== 0) {\n return comp;\n }\n comparedOnKeyField = comparedOnKeyField || orderBy.field.isKeyField();\n }\n // Assert that we actually compared by key\n debugAssert(\n comparedOnKeyField,\n \"orderBy used that doesn't compare on key field\"\n );\n return 0;\n }\n\n matches(doc: Document): boolean {\n return (\n this.matchesPathAndCollectionGroup(doc) &&\n this.matchesOrderBy(doc) &&\n this.matchesFilters(doc) &&\n this.matchesBounds(doc)\n );\n }\n\n hasLimitToFirst(): boolean {\n return !isNullOrUndefined(this.limit) && this.limitType === LimitType.First;\n }\n\n hasLimitToLast(): boolean {\n return !isNullOrUndefined(this.limit) && this.limitType === LimitType.Last;\n }\n\n getFirstOrderByField(): FieldPath | null {\n return this.explicitOrderBy.length > 0\n ? this.explicitOrderBy[0].field\n : null;\n }\n\n getInequalityFilterField(): FieldPath | null {\n for (const filter of this.filters) {\n if (filter instanceof FieldFilter && filter.isInequality()) {\n return filter.field;\n }\n }\n return null;\n }\n\n // Checks if any of the provided Operators are included in the query and\n // returns the first one that is, or null if none are.\n findFilterOperator(operators: Operator[]): Operator | null {\n for (const filter of this.filters) {\n if (filter instanceof FieldFilter) {\n if (operators.indexOf(filter.op) >= 0) {\n return filter.op;\n }\n }\n }\n return null;\n }\n\n isDocumentQuery(): boolean {\n return this.toTarget().isDocumentQuery();\n }\n\n isCollectionGroupQuery(): boolean {\n return this.collectionGroup !== null;\n }\n\n /**\n * Converts this `Query` instance to it's corresponding `Target`\n * representation.\n */\n toTarget(): Target {\n if (!this.memoizedTarget) {\n if (this.limitType === LimitType.First) {\n this.memoizedTarget = new Target(\n this.path,\n this.collectionGroup,\n this.orderBy,\n this.filters,\n this.limit,\n this.startAt,\n this.endAt\n );\n } else {\n // Flip the orderBy directions since we want the last results\n const orderBys = [] as OrderBy[];\n for (const orderBy of this.orderBy) {\n const dir =\n orderBy.dir === Direction.DESCENDING\n ? Direction.ASCENDING\n : Direction.DESCENDING;\n orderBys.push(new OrderBy(orderBy.field, dir));\n }\n\n // We need to swap the cursors to match the now-flipped query ordering.\n const startAt = this.endAt\n ? new Bound(this.endAt.position, !this.endAt.before)\n : null;\n const endAt = this.startAt\n ? new Bound(this.startAt.position, !this.startAt.before)\n : null;\n\n // Now return as a LimitType.First query.\n this.memoizedTarget = new Target(\n this.path,\n this.collectionGroup,\n orderBys,\n this.filters,\n this.limit,\n startAt,\n endAt\n );\n }\n }\n return this.memoizedTarget!;\n }\n\n private matchesPathAndCollectionGroup(doc: Document): boolean {\n const docPath = doc.key.path;\n if (this.collectionGroup !== null) {\n // NOTE: this.path is currently always empty since we don't expose Collection\n // Group queries rooted at a document path yet.\n return (\n doc.key.hasCollectionId(this.collectionGroup) &&\n this.path.isPrefixOf(docPath)\n );\n } else if (DocumentKey.isDocumentKey(this.path)) {\n // exact match for document queries\n return this.path.isEqual(docPath);\n } else {\n // shallow ancestor queries by default\n return this.path.isImmediateParentOf(docPath);\n }\n }\n\n /**\n * A document must have a value for every ordering clause in order to show up\n * in the results.\n */\n private matchesOrderBy(doc: Document): boolean {\n for (const orderBy of this.explicitOrderBy) {\n // order by key always matches\n if (!orderBy.field.isKeyField() && doc.field(orderBy.field) === null) {\n return false;\n }\n }\n return true;\n }\n\n private matchesFilters(doc: Document): boolean {\n for (const filter of this.filters) {\n if (!filter.matches(doc)) {\n return false;\n }\n }\n return true;\n }\n\n /**\n * Makes sure a document is within the bounds, if provided.\n */\n private matchesBounds(doc: Document): boolean {\n if (this.startAt && !this.startAt.sortsBeforeDocument(this.orderBy, doc)) {\n return false;\n }\n if (this.endAt && this.endAt.sortsBeforeDocument(this.orderBy, doc)) {\n return false;\n }\n return true;\n }\n\n private assertValidBound(bound: Bound): void {\n debugAssert(\n bound.position.length <= this.orderBy.length,\n 'Bound is longer than orderBy'\n );\n }\n}\n\nexport abstract class Filter {\n abstract matches(doc: Document): boolean;\n abstract canonicalId(): string;\n abstract isEqual(filter: Filter): boolean;\n}\n\nexport const enum Operator {\n LESS_THAN = '<',\n LESS_THAN_OR_EQUAL = '<=',\n EQUAL = '==',\n GREATER_THAN = '>',\n GREATER_THAN_OR_EQUAL = '>=',\n ARRAY_CONTAINS = 'array-contains',\n IN = 'in',\n ARRAY_CONTAINS_ANY = 'array-contains-any'\n}\n\nexport class FieldFilter extends Filter {\n protected constructor(\n public field: FieldPath,\n public op: Operator,\n public value: api.Value\n ) {\n super();\n }\n\n /**\n * Creates a filter based on the provided arguments.\n */\n static create(field: FieldPath, op: Operator, value: api.Value): FieldFilter {\n if (field.isKeyField()) {\n if (op === Operator.IN) {\n debugAssert(\n isArray(value),\n 'Comparing on key with IN, but filter value not an ArrayValue'\n );\n debugAssert(\n (value.arrayValue.values || []).every(elem => isReferenceValue(elem)),\n 'Comparing on key with IN, but an array value was not a RefValue'\n );\n return new KeyFieldInFilter(field, value);\n } else {\n debugAssert(\n isReferenceValue(value),\n 'Comparing on key, but filter value not a RefValue'\n );\n debugAssert(\n op !== Operator.ARRAY_CONTAINS && op !== Operator.ARRAY_CONTAINS_ANY,\n `'${op.toString()}' queries don't make sense on document keys.`\n );\n return new KeyFieldFilter(field, op, value);\n }\n } else if (isNullValue(value)) {\n if (op !== Operator.EQUAL) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. Null supports only equality comparisons.'\n );\n }\n return new FieldFilter(field, op, value);\n } else if (isNanValue(value)) {\n if (op !== Operator.EQUAL) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. NaN supports only equality comparisons.'\n );\n }\n return new FieldFilter(field, op, value);\n } else if (op === Operator.ARRAY_CONTAINS) {\n return new ArrayContainsFilter(field, value);\n } else if (op === Operator.IN) {\n debugAssert(\n isArray(value),\n 'IN filter has invalid value: ' + value.toString()\n );\n return new InFilter(field, value);\n } else if (op === Operator.ARRAY_CONTAINS_ANY) {\n debugAssert(\n isArray(value),\n 'ARRAY_CONTAINS_ANY filter has invalid value: ' + value.toString()\n );\n return new ArrayContainsAnyFilter(field, value);\n } else {\n return new FieldFilter(field, op, value);\n }\n }\n\n matches(doc: Document): boolean {\n const other = doc.field(this.field);\n\n // Only compare types with matching backend order (such as double and int).\n return (\n other !== null &&\n typeOrder(this.value) === typeOrder(other) &&\n this.matchesComparison(valueCompare(other, this.value))\n );\n }\n\n protected matchesComparison(comparison: number): boolean {\n switch (this.op) {\n case Operator.LESS_THAN:\n return comparison < 0;\n case Operator.LESS_THAN_OR_EQUAL:\n return comparison <= 0;\n case Operator.EQUAL:\n return comparison === 0;\n case Operator.GREATER_THAN:\n return comparison > 0;\n case Operator.GREATER_THAN_OR_EQUAL:\n return comparison >= 0;\n default:\n return fail('Unknown FieldFilter operator: ' + this.op);\n }\n }\n\n isInequality(): boolean {\n return (\n [\n Operator.LESS_THAN,\n Operator.LESS_THAN_OR_EQUAL,\n Operator.GREATER_THAN,\n Operator.GREATER_THAN_OR_EQUAL\n ].indexOf(this.op) >= 0\n );\n }\n\n canonicalId(): string {\n // TODO(b/29183165): Technically, this won't be unique if two values have\n // the same description, such as the int 3 and the string \"3\". So we should\n // add the types in here somehow, too.\n return (\n this.field.canonicalString() +\n this.op.toString() +\n canonicalId(this.value)\n );\n }\n\n isEqual(other: Filter): boolean {\n if (other instanceof FieldFilter) {\n return (\n this.op === other.op &&\n this.field.isEqual(other.field) &&\n valueEquals(this.value, other.value)\n );\n } else {\n return false;\n }\n }\n\n toString(): string {\n return `${this.field.canonicalString()} ${this.op} ${canonicalId(\n this.value\n )}`;\n }\n}\n\n/** Filter that matches on key fields (i.e. '__name__'). */\nexport class KeyFieldFilter extends FieldFilter {\n private readonly key: DocumentKey;\n\n constructor(field: FieldPath, op: Operator, value: api.Value) {\n super(field, op, value);\n debugAssert(\n isReferenceValue(value),\n 'KeyFieldFilter expects a ReferenceValue'\n );\n this.key = DocumentKey.fromName(value.referenceValue);\n }\n\n matches(doc: Document): boolean {\n const comparison = DocumentKey.comparator(doc.key, this.key);\n return this.matchesComparison(comparison);\n }\n}\n\n/** Filter that matches on key fields within an array. */\nexport class KeyFieldInFilter extends FieldFilter {\n private readonly keys: DocumentKey[];\n\n constructor(field: FieldPath, value: api.Value) {\n super(field, Operator.IN, value);\n debugAssert(isArray(value), 'KeyFieldInFilter expects an ArrayValue');\n this.keys = (value.arrayValue.values || []).map(v => {\n debugAssert(\n isReferenceValue(v),\n 'Comparing on key with IN, but an array value was not a ReferenceValue'\n );\n return DocumentKey.fromName(v.referenceValue);\n });\n }\n\n matches(doc: Document): boolean {\n return this.keys.some(key => key.isEqual(doc.key));\n }\n}\n\n/** A Filter that implements the array-contains operator. */\nexport class ArrayContainsFilter extends FieldFilter {\n constructor(field: FieldPath, value: api.Value) {\n super(field, Operator.ARRAY_CONTAINS, value);\n }\n\n matches(doc: Document): boolean {\n const other = doc.field(this.field);\n return isArray(other) && arrayValueContains(other.arrayValue, this.value);\n }\n}\n\n/** A Filter that implements the IN operator. */\nexport class InFilter extends FieldFilter {\n constructor(field: FieldPath, value: api.Value) {\n super(field, Operator.IN, value);\n debugAssert(isArray(value), 'InFilter expects an ArrayValue');\n }\n\n matches(doc: Document): boolean {\n const other = doc.field(this.field);\n return other !== null && arrayValueContains(this.value.arrayValue!, other);\n }\n}\n\n/** A Filter that implements the array-contains-any operator. */\nexport class ArrayContainsAnyFilter extends FieldFilter {\n constructor(field: FieldPath, value: api.Value) {\n super(field, Operator.ARRAY_CONTAINS_ANY, value);\n debugAssert(isArray(value), 'ArrayContainsAnyFilter expects an ArrayValue');\n }\n\n matches(doc: Document): boolean {\n const other = doc.field(this.field);\n if (!isArray(other) || !other.arrayValue.values) {\n return false;\n }\n return other.arrayValue.values.some(val =>\n arrayValueContains(this.value.arrayValue!, val)\n );\n }\n}\n\n/**\n * The direction of sorting in an order by.\n */\nexport const enum Direction {\n ASCENDING = 'asc',\n DESCENDING = 'desc'\n}\n\n/**\n * Represents a bound of a query.\n *\n * The bound is specified with the given components representing a position and\n * whether it's just before or just after the position (relative to whatever the\n * query order is).\n *\n * The position represents a logical index position for a query. It's a prefix\n * of values for the (potentially implicit) order by clauses of a query.\n *\n * Bound provides a function to determine whether a document comes before or\n * after a bound. This is influenced by whether the position is just before or\n * just after the provided values.\n */\nexport class Bound {\n constructor(readonly position: api.Value[], readonly before: boolean) {}\n\n canonicalId(): string {\n // TODO(b/29183165): Make this collision robust.\n return `${this.before ? 'b' : 'a'}:${this.position\n .map(p => canonicalId(p))\n .join(',')}`;\n }\n\n /**\n * Returns true if a document sorts before a bound using the provided sort\n * order.\n */\n sortsBeforeDocument(orderBy: OrderBy[], doc: Document): boolean {\n debugAssert(\n this.position.length <= orderBy.length,\n \"Bound has more components than query's orderBy\"\n );\n let comparison = 0;\n for (let i = 0; i < this.position.length; i++) {\n const orderByComponent = orderBy[i];\n const component = this.position[i];\n if (orderByComponent.field.isKeyField()) {\n debugAssert(\n isReferenceValue(component),\n 'Bound has a non-key value where the key path is being used.'\n );\n comparison = DocumentKey.comparator(\n DocumentKey.fromName(component.referenceValue),\n doc.key\n );\n } else {\n const docValue = doc.field(orderByComponent.field);\n debugAssert(\n docValue !== null,\n 'Field should exist since document matched the orderBy already.'\n );\n comparison = valueCompare(component, docValue);\n }\n if (orderByComponent.dir === Direction.DESCENDING) {\n comparison = comparison * -1;\n }\n if (comparison !== 0) {\n break;\n }\n }\n return this.before ? comparison <= 0 : comparison < 0;\n }\n\n isEqual(other: Bound | null): boolean {\n if (other === null) {\n return false;\n }\n if (\n this.before !== other.before ||\n this.position.length !== other.position.length\n ) {\n return false;\n }\n for (let i = 0; i < this.position.length; i++) {\n const thisPosition = this.position[i];\n const otherPosition = other.position[i];\n if (!valueEquals(thisPosition, otherPosition)) {\n return false;\n }\n }\n return true;\n }\n}\n\n/**\n * An ordering on a field, in some Direction. Direction defaults to ASCENDING.\n */\nexport class OrderBy {\n readonly dir: Direction;\n private readonly isKeyOrderBy: boolean;\n\n constructor(readonly field: FieldPath, dir?: Direction) {\n if (dir === undefined) {\n dir = Direction.ASCENDING;\n }\n this.dir = dir;\n this.isKeyOrderBy = field.isKeyField();\n }\n\n compare(d1: Document, d2: Document): number {\n const comparison = this.isKeyOrderBy\n ? DocumentKey.comparator(d1.key, d2.key)\n : compareDocumentsByField(this.field, d1, d2);\n switch (this.dir) {\n case Direction.ASCENDING:\n return comparison;\n case Direction.DESCENDING:\n return -1 * comparison;\n default:\n return fail('Unknown direction: ' + this.dir);\n }\n }\n\n canonicalId(): string {\n // TODO(b/29183165): Make this collision robust.\n return this.field.canonicalString() + this.dir.toString();\n }\n\n toString(): string {\n return `${this.field.canonicalString()} (${this.dir})`;\n }\n\n isEqual(other: OrderBy): boolean {\n return this.dir === other.dir && this.field.isEqual(other.field);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { Target } from '../core/target';\nimport { ListenSequenceNumber, TargetId } from '../core/types';\nimport { ByteString } from '../util/byte_string';\n\n/** An enumeration of the different purposes we have for targets. */\nexport const enum TargetPurpose {\n /** A regular, normal query target. */\n Listen,\n\n /**\n * The query target was used to refill a query after an existence filter mismatch.\n */\n ExistenceFilterMismatch,\n\n /** The query target was used to resolve a limbo document. */\n LimboResolution\n}\n\n/**\n * An immutable set of metadata that the local store tracks for each target.\n */\nexport class TargetData {\n constructor(\n /** The target being listened to. */\n readonly target: Target,\n /**\n * The target ID to which the target corresponds; Assigned by the\n * LocalStore for user listens and by the SyncEngine for limbo watches.\n */\n readonly targetId: TargetId,\n /** The purpose of the target. */\n readonly purpose: TargetPurpose,\n /**\n * The sequence number of the last transaction during which this target data\n * was modified.\n */\n readonly sequenceNumber: ListenSequenceNumber,\n /** The latest snapshot version seen for this target. */\n readonly snapshotVersion: SnapshotVersion = SnapshotVersion.min(),\n /**\n * The maximum snapshot version at which the associated view\n * contained no limbo documents.\n */\n readonly lastLimboFreeSnapshotVersion: SnapshotVersion = SnapshotVersion.min(),\n /**\n * An opaque, server-assigned token that allows watching a target to be\n * resumed after disconnecting without retransmitting all the data that\n * matches the target. The resume token essentially identifies a point in\n * time from which the server should resume sending results.\n */\n readonly resumeToken: ByteString = ByteString.EMPTY_BYTE_STRING\n ) {}\n\n /** Creates a new target data instance with an updated sequence number. */\n withSequenceNumber(sequenceNumber: number): TargetData {\n return new TargetData(\n this.target,\n this.targetId,\n this.purpose,\n sequenceNumber,\n this.snapshotVersion,\n this.lastLimboFreeSnapshotVersion,\n this.resumeToken\n );\n }\n\n /**\n * Creates a new target data instance with an updated resume token and\n * snapshot version.\n */\n withResumeToken(\n resumeToken: ByteString,\n snapshotVersion: SnapshotVersion\n ): TargetData {\n return new TargetData(\n this.target,\n this.targetId,\n this.purpose,\n this.sequenceNumber,\n snapshotVersion,\n this.lastLimboFreeSnapshotVersion,\n resumeToken\n );\n }\n\n /**\n * Creates a new target data instance with an updated last limbo free\n * snapshot version number.\n */\n withLastLimboFreeSnapshotVersion(\n lastLimboFreeSnapshotVersion: SnapshotVersion\n ): TargetData {\n return new TargetData(\n this.target,\n this.targetId,\n this.purpose,\n this.sequenceNumber,\n this.snapshotVersion,\n lastLimboFreeSnapshotVersion,\n this.resumeToken\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nexport class ExistenceFilter {\n // TODO(b/33078163): just use simplest form of existence filter for now\n constructor(public count: number) {}\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { fail } from '../util/assert';\nimport { Code } from '../util/error';\nimport { logError } from '../util/log';\n\n/**\n * Error Codes describing the different ways GRPC can fail. These are copied\n * directly from GRPC's sources here:\n *\n * https://github.com/grpc/grpc/blob/bceec94ea4fc5f0085d81235d8e1c06798dc341a/include/grpc%2B%2B/impl/codegen/status_code_enum.h\n *\n * Important! The names of these identifiers matter because the string forms\n * are used for reverse lookups from the webchannel stream. Do NOT change the\n * names of these identifiers or change this into a const enum.\n */\nenum RpcCode {\n OK = 0,\n CANCELLED = 1,\n UNKNOWN = 2,\n INVALID_ARGUMENT = 3,\n DEADLINE_EXCEEDED = 4,\n NOT_FOUND = 5,\n ALREADY_EXISTS = 6,\n PERMISSION_DENIED = 7,\n UNAUTHENTICATED = 16,\n RESOURCE_EXHAUSTED = 8,\n FAILED_PRECONDITION = 9,\n ABORTED = 10,\n OUT_OF_RANGE = 11,\n UNIMPLEMENTED = 12,\n INTERNAL = 13,\n UNAVAILABLE = 14,\n DATA_LOSS = 15\n}\n\n/**\n * Determines whether an error code represents a permanent error when received\n * in response to a non-write operation.\n *\n * See isPermanentWriteError for classifying write errors.\n */\nexport function isPermanentError(code: Code): boolean {\n switch (code) {\n case Code.OK:\n return fail('Treated status OK as error');\n case Code.CANCELLED:\n case Code.UNKNOWN:\n case Code.DEADLINE_EXCEEDED:\n case Code.RESOURCE_EXHAUSTED:\n case Code.INTERNAL:\n case Code.UNAVAILABLE:\n // Unauthenticated means something went wrong with our token and we need\n // to retry with new credentials which will happen automatically.\n case Code.UNAUTHENTICATED:\n return false;\n case Code.INVALID_ARGUMENT:\n case Code.NOT_FOUND:\n case Code.ALREADY_EXISTS:\n case Code.PERMISSION_DENIED:\n case Code.FAILED_PRECONDITION:\n // Aborted might be retried in some scenarios, but that is dependant on\n // the context and should handled individually by the calling code.\n // See https://cloud.google.com/apis/design/errors.\n case Code.ABORTED:\n case Code.OUT_OF_RANGE:\n case Code.UNIMPLEMENTED:\n case Code.DATA_LOSS:\n return true;\n default:\n return fail('Unknown status code: ' + code);\n }\n}\n\n/**\n * Determines whether an error code represents a permanent error when received\n * in response to a write operation.\n *\n * Write operations must be handled specially because as of b/119437764, ABORTED\n * errors on the write stream should be retried too (even though ABORTED errors\n * are not generally retryable).\n *\n * Note that during the initial handshake on the write stream an ABORTED error\n * signals that we should discard our stream token (i.e. it is permanent). This\n * means a handshake error should be classified with isPermanentError, above.\n */\nexport function isPermanentWriteError(code: Code): boolean {\n return isPermanentError(code) && code !== Code.ABORTED;\n}\n\n/**\n * Maps an error Code from a GRPC status identifier like 'NOT_FOUND'.\n *\n * @returns The Code equivalent to the given status string or undefined if\n * there is no match.\n */\nexport function mapCodeFromRpcStatus(status: string): Code | undefined {\n // lookup by string\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n const code: RpcCode = RpcCode[status as any] as any;\n if (code === undefined) {\n return undefined;\n }\n\n return mapCodeFromRpcCode(code);\n}\n\n/**\n * Maps an error Code from GRPC status code number, like 0, 1, or 14. These\n * are not the same as HTTP status codes.\n *\n * @returns The Code equivalent to the given GRPC status code. Fails if there\n * is no match.\n */\nexport function mapCodeFromRpcCode(code: number | undefined): Code {\n if (code === undefined) {\n // This shouldn't normally happen, but in certain error cases (like trying\n // to send invalid proto messages) we may get an error with no GRPC code.\n logError('GRPC error has no .code');\n return Code.UNKNOWN;\n }\n\n switch (code) {\n case RpcCode.OK:\n return Code.OK;\n case RpcCode.CANCELLED:\n return Code.CANCELLED;\n case RpcCode.UNKNOWN:\n return Code.UNKNOWN;\n case RpcCode.DEADLINE_EXCEEDED:\n return Code.DEADLINE_EXCEEDED;\n case RpcCode.RESOURCE_EXHAUSTED:\n return Code.RESOURCE_EXHAUSTED;\n case RpcCode.INTERNAL:\n return Code.INTERNAL;\n case RpcCode.UNAVAILABLE:\n return Code.UNAVAILABLE;\n case RpcCode.UNAUTHENTICATED:\n return Code.UNAUTHENTICATED;\n case RpcCode.INVALID_ARGUMENT:\n return Code.INVALID_ARGUMENT;\n case RpcCode.NOT_FOUND:\n return Code.NOT_FOUND;\n case RpcCode.ALREADY_EXISTS:\n return Code.ALREADY_EXISTS;\n case RpcCode.PERMISSION_DENIED:\n return Code.PERMISSION_DENIED;\n case RpcCode.FAILED_PRECONDITION:\n return Code.FAILED_PRECONDITION;\n case RpcCode.ABORTED:\n return Code.ABORTED;\n case RpcCode.OUT_OF_RANGE:\n return Code.OUT_OF_RANGE;\n case RpcCode.UNIMPLEMENTED:\n return Code.UNIMPLEMENTED;\n case RpcCode.DATA_LOSS:\n return Code.DATA_LOSS;\n default:\n return fail('Unknown status code: ' + code);\n }\n}\n\n/**\n * Maps an RPC code from a Code. This is the reverse operation from\n * mapCodeFromRpcCode and should really only be used in tests.\n */\nexport function mapRpcCodeFromCode(code: Code | undefined): number {\n if (code === undefined) {\n return RpcCode.OK;\n }\n\n switch (code) {\n case Code.OK:\n return RpcCode.OK;\n case Code.CANCELLED:\n return RpcCode.CANCELLED;\n case Code.UNKNOWN:\n return RpcCode.UNKNOWN;\n case Code.DEADLINE_EXCEEDED:\n return RpcCode.DEADLINE_EXCEEDED;\n case Code.RESOURCE_EXHAUSTED:\n return RpcCode.RESOURCE_EXHAUSTED;\n case Code.INTERNAL:\n return RpcCode.INTERNAL;\n case Code.UNAVAILABLE:\n return RpcCode.UNAVAILABLE;\n case Code.UNAUTHENTICATED:\n return RpcCode.UNAUTHENTICATED;\n case Code.INVALID_ARGUMENT:\n return RpcCode.INVALID_ARGUMENT;\n case Code.NOT_FOUND:\n return RpcCode.NOT_FOUND;\n case Code.ALREADY_EXISTS:\n return RpcCode.ALREADY_EXISTS;\n case Code.PERMISSION_DENIED:\n return RpcCode.PERMISSION_DENIED;\n case Code.FAILED_PRECONDITION:\n return RpcCode.FAILED_PRECONDITION;\n case Code.ABORTED:\n return RpcCode.ABORTED;\n case Code.OUT_OF_RANGE:\n return RpcCode.OUT_OF_RANGE;\n case Code.UNIMPLEMENTED:\n return RpcCode.UNIMPLEMENTED;\n case Code.DATA_LOSS:\n return RpcCode.DATA_LOSS;\n default:\n return fail('Unknown status code: ' + code);\n }\n}\n\n/**\n * Converts an HTTP Status Code to the equivalent error code.\n *\n * @param status An HTTP Status Code, like 200, 404, 503, etc.\n * @returns The equivalent Code. Unknown status codes are mapped to\n * Code.UNKNOWN.\n */\nexport function mapCodeFromHttpStatus(status: number): Code {\n // The canonical error codes for Google APIs [1] specify mapping onto HTTP\n // status codes but the mapping is not bijective. In each case of ambiguity\n // this function chooses a primary error.\n //\n // [1]\n // https://github.com/googleapis/googleapis/blob/master/google/rpc/code.proto\n switch (status) {\n case 200: // OK\n return Code.OK;\n\n case 400: // Bad Request\n return Code.INVALID_ARGUMENT;\n // Other possibilities based on the forward mapping\n // return Code.FAILED_PRECONDITION;\n // return Code.OUT_OF_RANGE;\n\n case 401: // Unauthorized\n return Code.UNAUTHENTICATED;\n\n case 403: // Forbidden\n return Code.PERMISSION_DENIED;\n\n case 404: // Not Found\n return Code.NOT_FOUND;\n\n case 409: // Conflict\n return Code.ABORTED;\n // Other possibilities:\n // return Code.ALREADY_EXISTS;\n\n case 416: // Range Not Satisfiable\n return Code.OUT_OF_RANGE;\n\n case 429: // Too Many Requests\n return Code.RESOURCE_EXHAUSTED;\n\n case 499: // Client Closed Request\n return Code.CANCELLED;\n\n case 500: // Internal Server Error\n return Code.UNKNOWN;\n // Other possibilities:\n // return Code.INTERNAL;\n // return Code.DATA_LOSS;\n\n case 501: // Unimplemented\n return Code.UNIMPLEMENTED;\n\n case 503: // Service Unavailable\n return Code.UNAVAILABLE;\n\n case 504: // Gateway Timeout\n return Code.DEADLINE_EXCEEDED;\n\n default:\n if (status >= 200 && status < 300) {\n return Code.OK;\n }\n if (status >= 400 && status < 500) {\n return Code.FAILED_PRECONDITION;\n }\n if (status >= 500 && status < 600) {\n return Code.INTERNAL;\n }\n return Code.UNKNOWN;\n }\n}\n\n/**\n * Converts an HTTP response's error status to the equivalent error code.\n *\n * @param status An HTTP error response status (\"FAILED_PRECONDITION\",\n * \"UNKNOWN\", etc.)\n * @returns The equivalent Code. Non-matching responses are mapped to\n * Code.UNKNOWN.\n */\nexport function mapCodeFromHttpResponseErrorStatus(status: string): Code {\n const serverError = status.toLowerCase().replace('_', '-');\n return Object.values(Code).indexOf(serverError as Code) >= 0\n ? (serverError as Code)\n : Code.UNKNOWN;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert, fail } from './assert';\n\n/*\n * Implementation of an immutable SortedMap using a Left-leaning\n * Red-Black Tree, adapted from the implementation in Mugs\n * (http://mads379.github.com/mugs/) by Mads Hartmann Jensen\n * (mads379@gmail.com).\n *\n * Original paper on Left-leaning Red-Black Trees:\n * http://www.cs.princeton.edu/~rs/talks/LLRB/LLRB.pdf\n *\n * Invariant 1: No red node has a red child\n * Invariant 2: Every leaf path has the same number of black nodes\n * Invariant 3: Only the left child can be red (left leaning)\n */\n\nexport type Comparator<K> = (key1: K, key2: K) => number;\n\nexport interface Entry<K, V> {\n key: K;\n value: V;\n}\n\n// An immutable sorted map implementation, based on a Left-leaning Red-Black\n// tree.\nexport class SortedMap<K, V> {\n // visible for testing\n root: LLRBNode<K, V> | LLRBEmptyNode<K, V>;\n\n constructor(\n public comparator: Comparator<K>,\n root?: LLRBNode<K, V> | LLRBEmptyNode<K, V>\n ) {\n this.root = root ? root : LLRBNode.EMPTY;\n }\n\n // Returns a copy of the map, with the specified key/value added or replaced.\n insert(key: K, value: V): SortedMap<K, V> {\n return new SortedMap<K, V>(\n this.comparator,\n this.root\n .insert(key, value, this.comparator)\n .copy(null, null, LLRBNode.BLACK, null, null)\n );\n }\n\n // Returns a copy of the map, with the specified key removed.\n remove(key: K): SortedMap<K, V> {\n return new SortedMap<K, V>(\n this.comparator,\n this.root\n .remove(key, this.comparator)\n .copy(null, null, LLRBNode.BLACK, null, null)\n );\n }\n\n // Returns the value of the node with the given key, or null.\n get(key: K): V | null {\n let node = this.root;\n while (!node.isEmpty()) {\n const cmp = this.comparator(key, node.key);\n if (cmp === 0) {\n return node.value;\n } else if (cmp < 0) {\n node = node.left;\n } else if (cmp > 0) {\n node = node.right;\n }\n }\n return null;\n }\n\n // Returns the index of the element in this sorted map, or -1 if it doesn't\n // exist.\n indexOf(key: K): number {\n // Number of nodes that were pruned when descending right\n let prunedNodes = 0;\n let node = this.root;\n while (!node.isEmpty()) {\n const cmp = this.comparator(key, node.key);\n if (cmp === 0) {\n return prunedNodes + node.left.size;\n } else if (cmp < 0) {\n node = node.left;\n } else {\n // Count all nodes left of the node plus the node itself\n prunedNodes += node.left.size + 1;\n node = node.right;\n }\n }\n // Node not found\n return -1;\n }\n\n isEmpty(): boolean {\n return this.root.isEmpty();\n }\n\n // Returns the total number of nodes in the map.\n get size(): number {\n return this.root.size;\n }\n\n // Returns the minimum key in the map.\n minKey(): K | null {\n return this.root.minKey();\n }\n\n // Returns the maximum key in the map.\n maxKey(): K | null {\n return this.root.maxKey();\n }\n\n // Traverses the map in key order and calls the specified action function\n // for each key/value pair. If action returns true, traversal is aborted.\n // Returns the first truthy value returned by action, or the last falsey\n // value returned by action.\n inorderTraversal<T>(action: (k: K, v: V) => T): T {\n return (this.root as LLRBNode<K, V>).inorderTraversal(action);\n }\n\n forEach(fn: (k: K, v: V) => void): void {\n this.inorderTraversal((k, v) => {\n fn(k, v);\n return false;\n });\n }\n\n toString(): string {\n const descriptions: string[] = [];\n this.inorderTraversal((k, v) => {\n descriptions.push(`${k}:${v}`);\n return false;\n });\n return `{${descriptions.join(', ')}}`;\n }\n\n // Traverses the map in reverse key order and calls the specified action\n // function for each key/value pair. If action returns true, traversal is\n // aborted.\n // Returns the first truthy value returned by action, or the last falsey\n // value returned by action.\n reverseTraversal<T>(action: (k: K, v: V) => T): T {\n return (this.root as LLRBNode<K, V>).reverseTraversal(action);\n }\n\n // Returns an iterator over the SortedMap.\n getIterator(): SortedMapIterator<K, V> {\n return new SortedMapIterator<K, V>(this.root, null, this.comparator, false);\n }\n\n getIteratorFrom(key: K): SortedMapIterator<K, V> {\n return new SortedMapIterator<K, V>(this.root, key, this.comparator, false);\n }\n\n getReverseIterator(): SortedMapIterator<K, V> {\n return new SortedMapIterator<K, V>(this.root, null, this.comparator, true);\n }\n\n getReverseIteratorFrom(key: K): SortedMapIterator<K, V> {\n return new SortedMapIterator<K, V>(this.root, key, this.comparator, true);\n }\n} // end SortedMap\n\n// An iterator over an LLRBNode.\nexport class SortedMapIterator<K, V> {\n private isReverse: boolean;\n private nodeStack: Array<LLRBNode<K, V> | LLRBEmptyNode<K, V>>;\n\n constructor(\n node: LLRBNode<K, V> | LLRBEmptyNode<K, V>,\n startKey: K | null,\n comparator: Comparator<K>,\n isReverse: boolean\n ) {\n this.isReverse = isReverse;\n this.nodeStack = [];\n\n let cmp = 1;\n while (!node.isEmpty()) {\n cmp = startKey ? comparator(node.key, startKey) : 1;\n // flip the comparison if we're going in reverse\n if (isReverse) {\n cmp *= -1;\n }\n\n if (cmp < 0) {\n // This node is less than our start key. ignore it\n if (this.isReverse) {\n node = node.left;\n } else {\n node = node.right;\n }\n } else if (cmp === 0) {\n // This node is exactly equal to our start key. Push it on the stack,\n // but stop iterating;\n this.nodeStack.push(node);\n break;\n } else {\n // This node is greater than our start key, add it to the stack and move\n // to the next one\n this.nodeStack.push(node);\n if (this.isReverse) {\n node = node.right;\n } else {\n node = node.left;\n }\n }\n }\n }\n\n getNext(): Entry<K, V> {\n debugAssert(\n this.nodeStack.length > 0,\n 'getNext() called on iterator when hasNext() is false.'\n );\n\n let node = this.nodeStack.pop()!;\n const result = { key: node.key, value: node.value };\n\n if (this.isReverse) {\n node = node.left;\n while (!node.isEmpty()) {\n this.nodeStack.push(node);\n node = node.right;\n }\n } else {\n node = node.right;\n while (!node.isEmpty()) {\n this.nodeStack.push(node);\n node = node.left;\n }\n }\n\n return result;\n }\n\n hasNext(): boolean {\n return this.nodeStack.length > 0;\n }\n\n peek(): Entry<K, V> | null {\n if (this.nodeStack.length === 0) {\n return null;\n }\n\n const node = this.nodeStack[this.nodeStack.length - 1];\n return { key: node.key, value: node.value };\n }\n} // end SortedMapIterator\n\n// Represents a node in a Left-leaning Red-Black tree.\nexport class LLRBNode<K, V> {\n readonly color: boolean;\n readonly left: LLRBNode<K, V> | LLRBEmptyNode<K, V>;\n readonly right: LLRBNode<K, V> | LLRBEmptyNode<K, V>;\n readonly size: number;\n\n // Empty node is shared between all LLRB trees.\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n static EMPTY: LLRBEmptyNode<any, any> = null as any;\n\n static RED = true;\n static BLACK = false;\n\n constructor(\n public key: K,\n public value: V,\n color?: boolean,\n left?: LLRBNode<K, V> | LLRBEmptyNode<K, V>,\n right?: LLRBNode<K, V> | LLRBEmptyNode<K, V>\n ) {\n this.color = color != null ? color : LLRBNode.RED;\n this.left = left != null ? left : LLRBNode.EMPTY;\n this.right = right != null ? right : LLRBNode.EMPTY;\n this.size = this.left.size + 1 + this.right.size;\n }\n\n // Returns a copy of the current node, optionally replacing pieces of it.\n copy(\n key: K | null,\n value: V | null,\n color: boolean | null,\n left: LLRBNode<K, V> | LLRBEmptyNode<K, V> | null,\n right: LLRBNode<K, V> | LLRBEmptyNode<K, V> | null\n ): LLRBNode<K, V> {\n return new LLRBNode<K, V>(\n key != null ? key : this.key,\n value != null ? value : this.value,\n color != null ? color : this.color,\n left != null ? left : this.left,\n right != null ? right : this.right\n );\n }\n\n isEmpty(): boolean {\n return false;\n }\n\n // Traverses the tree in key order and calls the specified action function\n // for each node. If action returns true, traversal is aborted.\n // Returns the first truthy value returned by action, or the last falsey\n // value returned by action.\n inorderTraversal<T>(action: (k: K, v: V) => T): T {\n return (\n (this.left as LLRBNode<K, V>).inorderTraversal(action) ||\n action(this.key, this.value) ||\n (this.right as LLRBNode<K, V>).inorderTraversal(action)\n );\n }\n\n // Traverses the tree in reverse key order and calls the specified action\n // function for each node. If action returns true, traversal is aborted.\n // Returns the first truthy value returned by action, or the last falsey\n // value returned by action.\n reverseTraversal<T>(action: (k: K, v: V) => T): T {\n return (\n (this.right as LLRBNode<K, V>).reverseTraversal(action) ||\n action(this.key, this.value) ||\n (this.left as LLRBNode<K, V>).reverseTraversal(action)\n );\n }\n\n // Returns the minimum node in the tree.\n private min(): LLRBNode<K, V> {\n if (this.left.isEmpty()) {\n return this;\n } else {\n return (this.left as LLRBNode<K, V>).min();\n }\n }\n\n // Returns the maximum key in the tree.\n minKey(): K | null {\n return this.min().key;\n }\n\n // Returns the maximum key in the tree.\n maxKey(): K | null {\n if (this.right.isEmpty()) {\n return this.key;\n } else {\n return this.right.maxKey();\n }\n }\n\n // Returns new tree, with the key/value added.\n insert(key: K, value: V, comparator: Comparator<K>): LLRBNode<K, V> {\n let n: LLRBNode<K, V> = this;\n const cmp = comparator(key, n.key);\n if (cmp < 0) {\n n = n.copy(null, null, null, n.left.insert(key, value, comparator), null);\n } else if (cmp === 0) {\n n = n.copy(null, value, null, null, null);\n } else {\n n = n.copy(\n null,\n null,\n null,\n null,\n n.right.insert(key, value, comparator)\n );\n }\n return n.fixUp();\n }\n\n private removeMin(): LLRBNode<K, V> | LLRBEmptyNode<K, V> {\n if (this.left.isEmpty()) {\n return LLRBNode.EMPTY;\n }\n let n: LLRBNode<K, V> = this;\n if (!n.left.isRed() && !n.left.left.isRed()) {\n n = n.moveRedLeft();\n }\n n = n.copy(null, null, null, (n.left as LLRBNode<K, V>).removeMin(), null);\n return n.fixUp();\n }\n\n // Returns new tree, with the specified item removed.\n remove(\n key: K,\n comparator: Comparator<K>\n ): LLRBNode<K, V> | LLRBEmptyNode<K, V> {\n let smallest: LLRBNode<K, V>;\n let n: LLRBNode<K, V> = this;\n if (comparator(key, n.key) < 0) {\n if (!n.left.isEmpty() && !n.left.isRed() && !n.left.left.isRed()) {\n n = n.moveRedLeft();\n }\n n = n.copy(null, null, null, n.left.remove(key, comparator), null);\n } else {\n if (n.left.isRed()) {\n n = n.rotateRight();\n }\n if (!n.right.isEmpty() && !n.right.isRed() && !n.right.left.isRed()) {\n n = n.moveRedRight();\n }\n if (comparator(key, n.key) === 0) {\n if (n.right.isEmpty()) {\n return LLRBNode.EMPTY;\n } else {\n smallest = (n.right as LLRBNode<K, V>).min();\n n = n.copy(\n smallest.key,\n smallest.value,\n null,\n null,\n (n.right as LLRBNode<K, V>).removeMin()\n );\n }\n }\n n = n.copy(null, null, null, null, n.right.remove(key, comparator));\n }\n return n.fixUp();\n }\n\n isRed(): boolean {\n return this.color;\n }\n\n // Returns new tree after performing any needed rotations.\n private fixUp(): LLRBNode<K, V> {\n let n: LLRBNode<K, V> = this;\n if (n.right.isRed() && !n.left.isRed()) {\n n = n.rotateLeft();\n }\n if (n.left.isRed() && n.left.left.isRed()) {\n n = n.rotateRight();\n }\n if (n.left.isRed() && n.right.isRed()) {\n n = n.colorFlip();\n }\n return n;\n }\n\n private moveRedLeft(): LLRBNode<K, V> {\n let n = this.colorFlip();\n if (n.right.left.isRed()) {\n n = n.copy(\n null,\n null,\n null,\n null,\n (n.right as LLRBNode<K, V>).rotateRight()\n );\n n = n.rotateLeft();\n n = n.colorFlip();\n }\n return n;\n }\n\n private moveRedRight(): LLRBNode<K, V> {\n let n = this.colorFlip();\n if (n.left.left.isRed()) {\n n = n.rotateRight();\n n = n.colorFlip();\n }\n return n;\n }\n\n private rotateLeft(): LLRBNode<K, V> {\n const nl = this.copy(null, null, LLRBNode.RED, null, this.right.left);\n return (this.right as LLRBNode<K, V>).copy(\n null,\n null,\n this.color,\n nl,\n null\n );\n }\n\n private rotateRight(): LLRBNode<K, V> {\n const nr = this.copy(null, null, LLRBNode.RED, this.left.right, null);\n return (this.left as LLRBNode<K, V>).copy(null, null, this.color, null, nr);\n }\n\n private colorFlip(): LLRBNode<K, V> {\n const left = this.left.copy(null, null, !this.left.color, null, null);\n const right = this.right.copy(null, null, !this.right.color, null, null);\n return this.copy(null, null, !this.color, left, right);\n }\n\n // For testing.\n checkMaxDepth(): boolean {\n const blackDepth = this.check();\n if (Math.pow(2.0, blackDepth) <= this.size + 1) {\n return true;\n } else {\n return false;\n }\n }\n\n // In a balanced RB tree, the black-depth (number of black nodes) from root to\n // leaves is equal on both sides. This function verifies that or asserts.\n protected check(): number {\n if (this.isRed() && this.left.isRed()) {\n throw fail('Red node has red child(' + this.key + ',' + this.value + ')');\n }\n if (this.right.isRed()) {\n throw fail('Right child of (' + this.key + ',' + this.value + ') is red');\n }\n const blackDepth = (this.left as LLRBNode<K, V>).check();\n if (blackDepth !== (this.right as LLRBNode<K, V>).check()) {\n throw fail('Black depths differ');\n } else {\n return blackDepth + (this.isRed() ? 0 : 1);\n }\n }\n} // end LLRBNode\n\n// Represents an empty node (a leaf node in the Red-Black Tree).\nexport class LLRBEmptyNode<K, V> {\n get key(): never {\n throw fail('LLRBEmptyNode has no key.');\n }\n get value(): never {\n throw fail('LLRBEmptyNode has no value.');\n }\n get color(): never {\n throw fail('LLRBEmptyNode has no color.');\n }\n get left(): never {\n throw fail('LLRBEmptyNode has no left child.');\n }\n get right(): never {\n throw fail('LLRBEmptyNode has no right child.');\n }\n size = 0;\n\n // Returns a copy of the current node.\n copy(\n key: K | null,\n value: V | null,\n color: boolean | null,\n left: LLRBNode<K, V> | LLRBEmptyNode<K, V> | null,\n right: LLRBNode<K, V> | LLRBEmptyNode<K, V> | null\n ): LLRBEmptyNode<K, V> {\n return this;\n }\n\n // Returns a copy of the tree, with the specified key/value added.\n insert(key: K, value: V, comparator: Comparator<K>): LLRBNode<K, V> {\n return new LLRBNode<K, V>(key, value);\n }\n\n // Returns a copy of the tree, with the specified key removed.\n remove(key: K, comparator: Comparator<K>): LLRBEmptyNode<K, V> {\n return this;\n }\n\n isEmpty(): boolean {\n return true;\n }\n\n inorderTraversal(action: (k: K, v: V) => boolean): boolean {\n return false;\n }\n\n reverseTraversal(action: (k: K, v: V) => boolean): boolean {\n return false;\n }\n\n minKey(): K | null {\n return null;\n }\n\n maxKey(): K | null {\n return null;\n }\n\n isRed(): boolean {\n return false;\n }\n\n // For testing.\n checkMaxDepth(): boolean {\n return true;\n }\n\n protected check(): 0 {\n return 0;\n }\n} // end LLRBEmptyNode\n\nLLRBNode.EMPTY = new LLRBEmptyNode<unknown, unknown>();\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SortedMap, SortedMapIterator } from './sorted_map';\n\n/**\n * SortedSet is an immutable (copy-on-write) collection that holds elements\n * in order specified by the provided comparator.\n *\n * NOTE: if provided comparator returns 0 for two elements, we consider them to\n * be equal!\n */\nexport class SortedSet<T> {\n private data: SortedMap<T, boolean>;\n\n constructor(private comparator: (left: T, right: T) => number) {\n this.data = new SortedMap<T, boolean>(this.comparator);\n }\n\n has(elem: T): boolean {\n return this.data.get(elem) !== null;\n }\n\n first(): T | null {\n return this.data.minKey();\n }\n\n last(): T | null {\n return this.data.maxKey();\n }\n\n get size(): number {\n return this.data.size;\n }\n\n indexOf(elem: T): number {\n return this.data.indexOf(elem);\n }\n\n /** Iterates elements in order defined by \"comparator\" */\n forEach(cb: (elem: T) => void): void {\n this.data.inorderTraversal((k: T, v: boolean) => {\n cb(k);\n return false;\n });\n }\n\n /** Iterates over `elem`s such that: range[0] <= elem < range[1]. */\n forEachInRange(range: [T, T], cb: (elem: T) => void): void {\n const iter = this.data.getIteratorFrom(range[0]);\n while (iter.hasNext()) {\n const elem = iter.getNext();\n if (this.comparator(elem.key, range[1]) >= 0) {\n return;\n }\n cb(elem.key);\n }\n }\n\n /**\n * Iterates over `elem`s such that: start <= elem until false is returned.\n */\n forEachWhile(cb: (elem: T) => boolean, start?: T): void {\n let iter: SortedMapIterator<T, boolean>;\n if (start !== undefined) {\n iter = this.data.getIteratorFrom(start);\n } else {\n iter = this.data.getIterator();\n }\n while (iter.hasNext()) {\n const elem = iter.getNext();\n const result = cb(elem.key);\n if (!result) {\n return;\n }\n }\n }\n\n /** Finds the least element greater than or equal to `elem`. */\n firstAfterOrEqual(elem: T): T | null {\n const iter = this.data.getIteratorFrom(elem);\n return iter.hasNext() ? iter.getNext().key : null;\n }\n\n getIterator(): SortedSetIterator<T> {\n return new SortedSetIterator<T>(this.data.getIterator());\n }\n\n getIteratorFrom(key: T): SortedSetIterator<T> {\n return new SortedSetIterator<T>(this.data.getIteratorFrom(key));\n }\n\n /** Inserts or updates an element */\n add(elem: T): SortedSet<T> {\n return this.copy(this.data.remove(elem).insert(elem, true));\n }\n\n /** Deletes an element */\n delete(elem: T): SortedSet<T> {\n if (!this.has(elem)) {\n return this;\n }\n return this.copy(this.data.remove(elem));\n }\n\n isEmpty(): boolean {\n return this.data.isEmpty();\n }\n\n unionWith(other: SortedSet<T>): SortedSet<T> {\n let result: SortedSet<T> = this;\n\n // Make sure `result` always refers to the larger one of the two sets.\n if (result.size < other.size) {\n result = other;\n other = this;\n }\n\n other.forEach(elem => {\n result = result.add(elem);\n });\n return result;\n }\n\n isEqual(other: SortedSet<T>): boolean {\n if (!(other instanceof SortedSet)) {\n return false;\n }\n if (this.size !== other.size) {\n return false;\n }\n\n const thisIt = this.data.getIterator();\n const otherIt = other.data.getIterator();\n while (thisIt.hasNext()) {\n const thisElem = thisIt.getNext().key;\n const otherElem = otherIt.getNext().key;\n if (this.comparator(thisElem, otherElem) !== 0) {\n return false;\n }\n }\n return true;\n }\n\n toArray(): T[] {\n const res: T[] = [];\n this.forEach(targetId => {\n res.push(targetId);\n });\n return res;\n }\n\n toString(): string {\n const result: T[] = [];\n this.forEach(elem => result.push(elem));\n return 'SortedSet(' + result.toString() + ')';\n }\n\n private copy(data: SortedMap<T, boolean>): SortedSet<T> {\n const result = new SortedSet(this.comparator);\n result.data = data;\n return result;\n }\n}\n\nexport class SortedSetIterator<T> {\n constructor(private iter: SortedMapIterator<T, boolean>) {}\n\n getNext(): T {\n return this.iter.getNext().key;\n }\n\n hasNext(): boolean {\n return this.iter.hasNext();\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { SortedMap } from '../util/sorted_map';\nimport { SortedSet } from '../util/sorted_set';\n\nimport { TargetId } from '../core/types';\nimport { primitiveComparator } from '../util/misc';\nimport { Document, MaybeDocument } from './document';\nimport { DocumentKey } from './document_key';\n\n/** Miscellaneous collection types / constants. */\nexport interface DocumentSizeEntry {\n maybeDocument: MaybeDocument;\n size: number;\n}\n\nexport type MaybeDocumentMap = SortedMap<DocumentKey, MaybeDocument>;\nconst EMPTY_MAYBE_DOCUMENT_MAP = new SortedMap<DocumentKey, MaybeDocument>(\n DocumentKey.comparator\n);\nexport function maybeDocumentMap(): MaybeDocumentMap {\n return EMPTY_MAYBE_DOCUMENT_MAP;\n}\n\nexport type NullableMaybeDocumentMap = SortedMap<\n DocumentKey,\n MaybeDocument | null\n>;\n\nexport function nullableMaybeDocumentMap(): NullableMaybeDocumentMap {\n return maybeDocumentMap();\n}\n\nexport interface DocumentSizeEntries {\n maybeDocuments: NullableMaybeDocumentMap;\n sizeMap: SortedMap<DocumentKey, number>;\n}\n\nexport type DocumentMap = SortedMap<DocumentKey, Document>;\nconst EMPTY_DOCUMENT_MAP = new SortedMap<DocumentKey, Document>(\n DocumentKey.comparator\n);\nexport function documentMap(): DocumentMap {\n return EMPTY_DOCUMENT_MAP;\n}\n\nexport type DocumentVersionMap = SortedMap<DocumentKey, SnapshotVersion>;\nconst EMPTY_DOCUMENT_VERSION_MAP = new SortedMap<DocumentKey, SnapshotVersion>(\n DocumentKey.comparator\n);\nexport function documentVersionMap(): DocumentVersionMap {\n return EMPTY_DOCUMENT_VERSION_MAP;\n}\n\nexport type DocumentKeySet = SortedSet<DocumentKey>;\nconst EMPTY_DOCUMENT_KEY_SET = new SortedSet(DocumentKey.comparator);\nexport function documentKeySet(...keys: DocumentKey[]): DocumentKeySet {\n let set = EMPTY_DOCUMENT_KEY_SET;\n for (const key of keys) {\n set = set.add(key);\n }\n return set;\n}\n\nexport type TargetIdSet = SortedSet<TargetId>;\nconst EMPTY_TARGET_ID_SET = new SortedSet<TargetId>(primitiveComparator);\nexport function targetIdSet(): SortedSet<TargetId> {\n return EMPTY_TARGET_ID_SET;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SortedMap } from '../util/sorted_map';\n\nimport { documentMap } from './collections';\nimport { Document } from './document';\nimport { DocumentComparator } from './document_comparator';\nimport { DocumentKey } from './document_key';\n\n/**\n * DocumentSet is an immutable (copy-on-write) collection that holds documents\n * in order specified by the provided comparator. We always add a document key\n * comparator on top of what is provided to guarantee document equality based on\n * the key.\n */\n\nexport class DocumentSet {\n /**\n * Returns an empty copy of the existing DocumentSet, using the same\n * comparator.\n */\n static emptySet(oldSet: DocumentSet): DocumentSet {\n return new DocumentSet(oldSet.comparator);\n }\n\n private comparator: DocumentComparator;\n private keyedMap: SortedMap<DocumentKey, Document>;\n private sortedSet: SortedMap<Document, null>;\n\n /** The default ordering is by key if the comparator is omitted */\n constructor(comp?: DocumentComparator) {\n // We are adding document key comparator to the end as it's the only\n // guaranteed unique property of a document.\n if (comp) {\n this.comparator = (d1: Document, d2: Document) =>\n comp(d1, d2) || DocumentKey.comparator(d1.key, d2.key);\n } else {\n this.comparator = (d1: Document, d2: Document) =>\n DocumentKey.comparator(d1.key, d2.key);\n }\n\n this.keyedMap = documentMap();\n this.sortedSet = new SortedMap<Document, null>(this.comparator);\n }\n\n has(key: DocumentKey): boolean {\n return this.keyedMap.get(key) != null;\n }\n\n get(key: DocumentKey): Document | null {\n return this.keyedMap.get(key);\n }\n\n first(): Document | null {\n return this.sortedSet.minKey();\n }\n\n last(): Document | null {\n return this.sortedSet.maxKey();\n }\n\n isEmpty(): boolean {\n return this.sortedSet.isEmpty();\n }\n\n /**\n * Returns the index of the provided key in the document set, or -1 if the\n * document key is not present in the set;\n */\n indexOf(key: DocumentKey): number {\n const doc = this.keyedMap.get(key);\n return doc ? this.sortedSet.indexOf(doc) : -1;\n }\n\n get size(): number {\n return this.sortedSet.size;\n }\n\n /** Iterates documents in order defined by \"comparator\" */\n forEach(cb: (doc: Document) => void): void {\n this.sortedSet.inorderTraversal((k, v) => {\n cb(k);\n return false;\n });\n }\n\n /** Inserts or updates a document with the same key */\n add(doc: Document): DocumentSet {\n // First remove the element if we have it.\n const set = this.delete(doc.key);\n return set.copy(\n set.keyedMap.insert(doc.key, doc),\n set.sortedSet.insert(doc, null)\n );\n }\n\n /** Deletes a document with a given key */\n delete(key: DocumentKey): DocumentSet {\n const doc = this.get(key);\n if (!doc) {\n return this;\n }\n\n return this.copy(this.keyedMap.remove(key), this.sortedSet.remove(doc));\n }\n\n isEqual(other: DocumentSet | null | undefined): boolean {\n if (!(other instanceof DocumentSet)) {\n return false;\n }\n if (this.size !== other.size) {\n return false;\n }\n\n const thisIt = this.sortedSet.getIterator();\n const otherIt = other.sortedSet.getIterator();\n while (thisIt.hasNext()) {\n const thisDoc = thisIt.getNext().key;\n const otherDoc = otherIt.getNext().key;\n if (!thisDoc.isEqual(otherDoc)) {\n return false;\n }\n }\n return true;\n }\n\n toString(): string {\n const docStrings: string[] = [];\n this.forEach(doc => {\n docStrings.push(doc.toString());\n });\n if (docStrings.length === 0) {\n return 'DocumentSet ()';\n } else {\n return 'DocumentSet (\\n ' + docStrings.join(' \\n') + '\\n)';\n }\n }\n\n private copy(\n keyedMap: SortedMap<DocumentKey, Document>,\n sortedSet: SortedMap<Document, null>\n ): DocumentSet {\n const newSet = new DocumentSet();\n newSet.comparator = this.comparator;\n newSet.keyedMap = keyedMap;\n newSet.sortedSet = sortedSet;\n return newSet;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Document } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { DocumentSet } from '../model/document_set';\nimport { fail } from '../util/assert';\nimport { SortedMap } from '../util/sorted_map';\n\nimport { DocumentKeySet } from '../model/collections';\nimport { Query } from './query';\n\nexport const enum ChangeType {\n Added,\n Removed,\n Modified,\n Metadata\n}\n\nexport interface DocumentViewChange {\n type: ChangeType;\n doc: Document;\n}\n\nexport const enum SyncState {\n Local,\n Synced\n}\n\n/**\n * DocumentChangeSet keeps track of a set of changes to docs in a query, merging\n * duplicate events for the same doc.\n */\nexport class DocumentChangeSet {\n private changeMap = new SortedMap<DocumentKey, DocumentViewChange>(\n DocumentKey.comparator\n );\n\n track(change: DocumentViewChange): void {\n const key = change.doc.key;\n const oldChange = this.changeMap.get(key);\n if (!oldChange) {\n this.changeMap = this.changeMap.insert(key, change);\n return;\n }\n\n // Merge the new change with the existing change.\n if (\n change.type !== ChangeType.Added &&\n oldChange.type === ChangeType.Metadata\n ) {\n this.changeMap = this.changeMap.insert(key, change);\n } else if (\n change.type === ChangeType.Metadata &&\n oldChange.type !== ChangeType.Removed\n ) {\n this.changeMap = this.changeMap.insert(key, {\n type: oldChange.type,\n doc: change.doc\n });\n } else if (\n change.type === ChangeType.Modified &&\n oldChange.type === ChangeType.Modified\n ) {\n this.changeMap = this.changeMap.insert(key, {\n type: ChangeType.Modified,\n doc: change.doc\n });\n } else if (\n change.type === ChangeType.Modified &&\n oldChange.type === ChangeType.Added\n ) {\n this.changeMap = this.changeMap.insert(key, {\n type: ChangeType.Added,\n doc: change.doc\n });\n } else if (\n change.type === ChangeType.Removed &&\n oldChange.type === ChangeType.Added\n ) {\n this.changeMap = this.changeMap.remove(key);\n } else if (\n change.type === ChangeType.Removed &&\n oldChange.type === ChangeType.Modified\n ) {\n this.changeMap = this.changeMap.insert(key, {\n type: ChangeType.Removed,\n doc: oldChange.doc\n });\n } else if (\n change.type === ChangeType.Added &&\n oldChange.type === ChangeType.Removed\n ) {\n this.changeMap = this.changeMap.insert(key, {\n type: ChangeType.Modified,\n doc: change.doc\n });\n } else {\n // This includes these cases, which don't make sense:\n // Added->Added\n // Removed->Removed\n // Modified->Added\n // Removed->Modified\n // Metadata->Added\n // Removed->Metadata\n fail(\n 'unsupported combination of changes: ' +\n JSON.stringify(change) +\n ' after ' +\n JSON.stringify(oldChange)\n );\n }\n }\n\n getChanges(): DocumentViewChange[] {\n const changes: DocumentViewChange[] = [];\n this.changeMap.inorderTraversal(\n (key: DocumentKey, change: DocumentViewChange) => {\n changes.push(change);\n }\n );\n return changes;\n }\n}\n\nexport class ViewSnapshot {\n constructor(\n readonly query: Query,\n readonly docs: DocumentSet,\n readonly oldDocs: DocumentSet,\n readonly docChanges: DocumentViewChange[],\n readonly mutatedKeys: DocumentKeySet,\n readonly fromCache: boolean,\n readonly syncStateChanged: boolean,\n readonly excludesMetadataChanges: boolean\n ) {}\n\n /** Returns a view snapshot as if all documents in the snapshot were added. */\n static fromInitialDocuments(\n query: Query,\n documents: DocumentSet,\n mutatedKeys: DocumentKeySet,\n fromCache: boolean\n ): ViewSnapshot {\n const changes: DocumentViewChange[] = [];\n documents.forEach(doc => {\n changes.push({ type: ChangeType.Added, doc });\n });\n\n return new ViewSnapshot(\n query,\n documents,\n DocumentSet.emptySet(documents),\n changes,\n mutatedKeys,\n fromCache,\n /* syncStateChanged= */ true,\n /* excludesMetadataChanges= */ false\n );\n }\n\n get hasPendingWrites(): boolean {\n return !this.mutatedKeys.isEmpty();\n }\n\n isEqual(other: ViewSnapshot): boolean {\n if (\n this.fromCache !== other.fromCache ||\n this.syncStateChanged !== other.syncStateChanged ||\n !this.mutatedKeys.isEqual(other.mutatedKeys) ||\n !this.query.isEqual(other.query) ||\n !this.docs.isEqual(other.docs) ||\n !this.oldDocs.isEqual(other.oldDocs)\n ) {\n return false;\n }\n const changes: DocumentViewChange[] = this.docChanges;\n const otherChanges: DocumentViewChange[] = other.docChanges;\n if (changes.length !== otherChanges.length) {\n return false;\n }\n for (let i = 0; i < changes.length; i++) {\n if (\n changes[i].type !== otherChanges[i].type ||\n !changes[i].doc.isEqual(otherChanges[i].doc)\n ) {\n return false;\n }\n }\n return true;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { TargetId } from '../core/types';\nimport {\n documentKeySet,\n DocumentKeySet,\n maybeDocumentMap,\n MaybeDocumentMap,\n targetIdSet\n} from '../model/collections';\nimport { SortedSet } from '../util/sorted_set';\nimport { ByteString } from '../util/byte_string';\n\n/**\n * An event from the RemoteStore. It is split into targetChanges (changes to the\n * state or the set of documents in our watched targets) and documentUpdates\n * (changes to the actual documents).\n */\nexport class RemoteEvent {\n constructor(\n /**\n * The snapshot version this event brings us up to, or MIN if not set.\n */\n readonly snapshotVersion: SnapshotVersion,\n /**\n * A map from target to changes to the target. See TargetChange.\n */\n readonly targetChanges: Map<TargetId, TargetChange>,\n /**\n * A set of targets that is known to be inconsistent. Listens for these\n * targets should be re-established without resume tokens.\n */\n readonly targetMismatches: SortedSet<TargetId>,\n /**\n * A set of which documents have changed or been deleted, along with the\n * doc's new values (if not deleted).\n */\n readonly documentUpdates: MaybeDocumentMap,\n /**\n * A set of which document updates are due only to limbo resolution targets.\n */\n readonly resolvedLimboDocuments: DocumentKeySet\n ) {}\n\n /**\n * HACK: Views require RemoteEvents in order to determine whether the view is\n * CURRENT, but secondary tabs don't receive remote events. So this method is\n * used to create a synthesized RemoteEvent that can be used to apply a\n * CURRENT status change to a View, for queries executed in a different tab.\n */\n // PORTING NOTE: Multi-tab only\n static createSynthesizedRemoteEventForCurrentChange(\n targetId: TargetId,\n current: boolean\n ): RemoteEvent {\n const targetChanges = new Map<TargetId, TargetChange>();\n targetChanges.set(\n targetId,\n TargetChange.createSynthesizedTargetChangeForCurrentChange(\n targetId,\n current\n )\n );\n return new RemoteEvent(\n SnapshotVersion.min(),\n targetChanges,\n targetIdSet(),\n maybeDocumentMap(),\n documentKeySet()\n );\n }\n}\n\n/**\n * A TargetChange specifies the set of changes for a specific target as part of\n * a RemoteEvent. These changes track which documents are added, modified or\n * removed, as well as the target's resume token and whether the target is\n * marked CURRENT.\n * The actual changes *to* documents are not part of the TargetChange since\n * documents may be part of multiple targets.\n */\nexport class TargetChange {\n constructor(\n /**\n * An opaque, server-assigned token that allows watching a query to be resumed\n * after disconnecting without retransmitting all the data that matches the\n * query. The resume token essentially identifies a point in time from which\n * the server should resume sending results.\n */\n readonly resumeToken: ByteString,\n /**\n * The \"current\" (synced) status of this target. Note that \"current\"\n * has special meaning in the RPC protocol that implies that a target is\n * both up-to-date and consistent with the rest of the watch stream.\n */\n readonly current: boolean,\n /**\n * The set of documents that were newly assigned to this target as part of\n * this remote event.\n */\n readonly addedDocuments: DocumentKeySet,\n /**\n * The set of documents that were already assigned to this target but received\n * an update during this remote event.\n */\n readonly modifiedDocuments: DocumentKeySet,\n /**\n * The set of documents that were removed from this target as part of this\n * remote event.\n */\n readonly removedDocuments: DocumentKeySet\n ) {}\n\n /**\n * This method is used to create a synthesized TargetChanges that can be used to\n * apply a CURRENT status change to a View (for queries executed in a different\n * tab) or for new queries (to raise snapshots with correct CURRENT status).\n */\n static createSynthesizedTargetChangeForCurrentChange(\n targetId: TargetId,\n current: boolean\n ): TargetChange {\n return new TargetChange(\n ByteString.EMPTY_BYTE_STRING,\n current,\n documentKeySet(),\n documentKeySet(),\n documentKeySet()\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { TargetId } from '../core/types';\nimport { ChangeType } from '../core/view_snapshot';\nimport { TargetData, TargetPurpose } from '../local/target_data';\nimport {\n documentKeySet,\n DocumentKeySet,\n maybeDocumentMap\n} from '../model/collections';\nimport { Document, MaybeDocument, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { debugAssert, fail, hardAssert } from '../util/assert';\nimport { FirestoreError } from '../util/error';\nimport { logDebug } from '../util/log';\nimport { primitiveComparator } from '../util/misc';\nimport { SortedMap } from '../util/sorted_map';\nimport { SortedSet } from '../util/sorted_set';\nimport { ExistenceFilter } from './existence_filter';\nimport { RemoteEvent, TargetChange } from './remote_event';\nimport { ByteString } from '../util/byte_string';\n\n/**\n * Internal representation of the watcher API protocol buffers.\n */\nexport type WatchChange =\n | DocumentWatchChange\n | WatchTargetChange\n | ExistenceFilterChange;\n\n/**\n * Represents a changed document and a list of target ids to which this change\n * applies.\n *\n * If document has been deleted NoDocument will be provided.\n */\nexport class DocumentWatchChange {\n constructor(\n /** The new document applies to all of these targets. */\n public updatedTargetIds: TargetId[],\n /** The new document is removed from all of these targets. */\n public removedTargetIds: TargetId[],\n /** The key of the document for this change. */\n public key: DocumentKey,\n /**\n * The new document or NoDocument if it was deleted. Is null if the\n * document went out of view without the server sending a new document.\n */\n public newDoc: MaybeDocument | null\n ) {}\n}\n\nexport class ExistenceFilterChange {\n constructor(\n public targetId: TargetId,\n public existenceFilter: ExistenceFilter\n ) {}\n}\n\nexport const enum WatchTargetChangeState {\n NoChange,\n Added,\n Removed,\n Current,\n Reset\n}\n\nexport class WatchTargetChange {\n constructor(\n /** What kind of change occurred to the watch target. */\n public state: WatchTargetChangeState,\n /** The target IDs that were added/removed/set. */\n public targetIds: TargetId[],\n /**\n * An opaque, server-assigned token that allows watching a target to be\n * resumed after disconnecting without retransmitting all the data that\n * matches the target. The resume token essentially identifies a point in\n * time from which the server should resume sending results.\n */\n public resumeToken: ByteString = ByteString.EMPTY_BYTE_STRING,\n /** An RPC error indicating why the watch failed. */\n public cause: FirestoreError | null = null\n ) {}\n}\n\n/** Tracks the internal state of a Watch target. */\nclass TargetState {\n /**\n * The number of pending responses (adds or removes) that we are waiting on.\n * We only consider targets active that have no pending responses.\n */\n private pendingResponses = 0;\n\n /**\n * Keeps track of the document changes since the last raised snapshot.\n *\n * These changes are continuously updated as we receive document updates and\n * always reflect the current set of changes against the last issued snapshot.\n */\n private documentChanges: SortedMap<\n DocumentKey,\n ChangeType\n > = snapshotChangesMap();\n\n /** See public getters for explanations of these fields. */\n private _resumeToken: ByteString = ByteString.EMPTY_BYTE_STRING;\n private _current = false;\n\n /**\n * Whether this target state should be included in the next snapshot. We\n * initialize to true so that newly-added targets are included in the next\n * RemoteEvent.\n */\n private _hasPendingChanges = true;\n\n /**\n * Whether this target has been marked 'current'.\n *\n * 'Current' has special meaning in the RPC protocol: It implies that the\n * Watch backend has sent us all changes up to the point at which the target\n * was added and that the target is consistent with the rest of the watch\n * stream.\n */\n get current(): boolean {\n return this._current;\n }\n\n /** The last resume token sent to us for this target. */\n get resumeToken(): ByteString {\n return this._resumeToken;\n }\n\n /** Whether this target has pending target adds or target removes. */\n get isPending(): boolean {\n return this.pendingResponses !== 0;\n }\n\n /** Whether we have modified any state that should trigger a snapshot. */\n get hasPendingChanges(): boolean {\n return this._hasPendingChanges;\n }\n\n /**\n * Applies the resume token to the TargetChange, but only when it has a new\n * value. Empty resumeTokens are discarded.\n */\n updateResumeToken(resumeToken: ByteString): void {\n if (resumeToken.approximateByteSize() > 0) {\n this._hasPendingChanges = true;\n this._resumeToken = resumeToken;\n }\n }\n\n /**\n * Creates a target change from the current set of changes.\n *\n * To reset the document changes after raising this snapshot, call\n * `clearPendingChanges()`.\n */\n toTargetChange(): TargetChange {\n let addedDocuments = documentKeySet();\n let modifiedDocuments = documentKeySet();\n let removedDocuments = documentKeySet();\n\n this.documentChanges.forEach((key, changeType) => {\n switch (changeType) {\n case ChangeType.Added:\n addedDocuments = addedDocuments.add(key);\n break;\n case ChangeType.Modified:\n modifiedDocuments = modifiedDocuments.add(key);\n break;\n case ChangeType.Removed:\n removedDocuments = removedDocuments.add(key);\n break;\n default:\n fail('Encountered invalid change type: ' + changeType);\n }\n });\n\n return new TargetChange(\n this._resumeToken,\n this._current,\n addedDocuments,\n modifiedDocuments,\n removedDocuments\n );\n }\n\n /**\n * Resets the document changes and sets `hasPendingChanges` to false.\n */\n clearPendingChanges(): void {\n this._hasPendingChanges = false;\n this.documentChanges = snapshotChangesMap();\n }\n\n addDocumentChange(key: DocumentKey, changeType: ChangeType): void {\n this._hasPendingChanges = true;\n this.documentChanges = this.documentChanges.insert(key, changeType);\n }\n\n removeDocumentChange(key: DocumentKey): void {\n this._hasPendingChanges = true;\n this.documentChanges = this.documentChanges.remove(key);\n }\n\n recordPendingTargetRequest(): void {\n this.pendingResponses += 1;\n }\n\n recordTargetResponse(): void {\n this.pendingResponses -= 1;\n }\n\n markCurrent(): void {\n this._hasPendingChanges = true;\n this._current = true;\n }\n}\n\n/**\n * Interface implemented by RemoteStore to expose target metadata to the\n * WatchChangeAggregator.\n */\nexport interface TargetMetadataProvider {\n /**\n * Returns the set of remote document keys for the given target ID as of the\n * last raised snapshot.\n */\n getRemoteKeysForTarget(targetId: TargetId): DocumentKeySet;\n\n /**\n * Returns the TargetData for an active target ID or 'null' if this target\n * has become inactive\n */\n getTargetDataForTarget(targetId: TargetId): TargetData | null;\n}\n\nconst LOG_TAG = 'WatchChangeAggregator';\n\n/**\n * A helper class to accumulate watch changes into a RemoteEvent.\n */\nexport class WatchChangeAggregator {\n constructor(private metadataProvider: TargetMetadataProvider) {}\n\n /** The internal state of all tracked targets. */\n private targetStates = new Map<TargetId, TargetState>();\n\n /** Keeps track of the documents to update since the last raised snapshot. */\n private pendingDocumentUpdates = maybeDocumentMap();\n\n /** A mapping of document keys to their set of target IDs. */\n private pendingDocumentTargetMapping = documentTargetMap();\n\n /**\n * A list of targets with existence filter mismatches. These targets are\n * known to be inconsistent and their listens needs to be re-established by\n * RemoteStore.\n */\n private pendingTargetResets = new SortedSet<TargetId>(primitiveComparator);\n\n /**\n * Processes and adds the DocumentWatchChange to the current set of changes.\n */\n handleDocumentChange(docChange: DocumentWatchChange): void {\n for (const targetId of docChange.updatedTargetIds) {\n if (docChange.newDoc instanceof Document) {\n this.addDocumentToTarget(targetId, docChange.newDoc);\n } else if (docChange.newDoc instanceof NoDocument) {\n this.removeDocumentFromTarget(\n targetId,\n docChange.key,\n docChange.newDoc\n );\n }\n }\n\n for (const targetId of docChange.removedTargetIds) {\n this.removeDocumentFromTarget(targetId, docChange.key, docChange.newDoc);\n }\n }\n\n /** Processes and adds the WatchTargetChange to the current set of changes. */\n handleTargetChange(targetChange: WatchTargetChange): void {\n this.forEachTarget(targetChange, targetId => {\n const targetState = this.ensureTargetState(targetId);\n switch (targetChange.state) {\n case WatchTargetChangeState.NoChange:\n if (this.isActiveTarget(targetId)) {\n targetState.updateResumeToken(targetChange.resumeToken);\n }\n break;\n case WatchTargetChangeState.Added:\n // We need to decrement the number of pending acks needed from watch\n // for this targetId.\n targetState.recordTargetResponse();\n if (!targetState.isPending) {\n // We have a freshly added target, so we need to reset any state\n // that we had previously. This can happen e.g. when remove and add\n // back a target for existence filter mismatches.\n targetState.clearPendingChanges();\n }\n targetState.updateResumeToken(targetChange.resumeToken);\n break;\n case WatchTargetChangeState.Removed:\n // We need to keep track of removed targets to we can post-filter and\n // remove any target changes.\n // We need to decrement the number of pending acks needed from watch\n // for this targetId.\n targetState.recordTargetResponse();\n if (!targetState.isPending) {\n this.removeTarget(targetId);\n }\n debugAssert(\n !targetChange.cause,\n 'WatchChangeAggregator does not handle errored targets'\n );\n break;\n case WatchTargetChangeState.Current:\n if (this.isActiveTarget(targetId)) {\n targetState.markCurrent();\n targetState.updateResumeToken(targetChange.resumeToken);\n }\n break;\n case WatchTargetChangeState.Reset:\n if (this.isActiveTarget(targetId)) {\n // Reset the target and synthesizes removes for all existing\n // documents. The backend will re-add any documents that still\n // match the target before it sends the next global snapshot.\n this.resetTarget(targetId);\n targetState.updateResumeToken(targetChange.resumeToken);\n }\n break;\n default:\n fail('Unknown target watch change state: ' + targetChange.state);\n }\n });\n }\n\n /**\n * Iterates over all targetIds that the watch change applies to: either the\n * targetIds explicitly listed in the change or the targetIds of all currently\n * active targets.\n */\n forEachTarget(\n targetChange: WatchTargetChange,\n fn: (targetId: TargetId) => void\n ): void {\n if (targetChange.targetIds.length > 0) {\n targetChange.targetIds.forEach(fn);\n } else {\n this.targetStates.forEach((_, targetId) => {\n if (this.isActiveTarget(targetId)) {\n fn(targetId);\n }\n });\n }\n }\n\n /**\n * Handles existence filters and synthesizes deletes for filter mismatches.\n * Targets that are invalidated by filter mismatches are added to\n * `pendingTargetResets`.\n */\n handleExistenceFilter(watchChange: ExistenceFilterChange): void {\n const targetId = watchChange.targetId;\n const expectedCount = watchChange.existenceFilter.count;\n\n const targetData = this.targetDataForActiveTarget(targetId);\n if (targetData) {\n const target = targetData.target;\n if (target.isDocumentQuery()) {\n if (expectedCount === 0) {\n // The existence filter told us the document does not exist. We deduce\n // that this document does not exist and apply a deleted document to\n // our updates. Without applying this deleted document there might be\n // another query that will raise this document as part of a snapshot\n // until it is resolved, essentially exposing inconsistency between\n // queries.\n const key = new DocumentKey(target.path);\n this.removeDocumentFromTarget(\n targetId,\n key,\n new NoDocument(key, SnapshotVersion.min())\n );\n } else {\n hardAssert(\n expectedCount === 1,\n 'Single document existence filter with count: ' + expectedCount\n );\n }\n } else {\n const currentSize = this.getCurrentDocumentCountForTarget(targetId);\n if (currentSize !== expectedCount) {\n // Existence filter mismatch: We reset the mapping and raise a new\n // snapshot with `isFromCache:true`.\n this.resetTarget(targetId);\n this.pendingTargetResets = this.pendingTargetResets.add(targetId);\n }\n }\n }\n }\n\n /**\n * Converts the currently accumulated state into a remote event at the\n * provided snapshot version. Resets the accumulated changes before returning.\n */\n createRemoteEvent(snapshotVersion: SnapshotVersion): RemoteEvent {\n const targetChanges = new Map<TargetId, TargetChange>();\n\n this.targetStates.forEach((targetState, targetId) => {\n const targetData = this.targetDataForActiveTarget(targetId);\n if (targetData) {\n if (targetState.current && targetData.target.isDocumentQuery()) {\n // Document queries for document that don't exist can produce an empty\n // result set. To update our local cache, we synthesize a document\n // delete if we have not previously received the document. This\n // resolves the limbo state of the document, removing it from\n // limboDocumentRefs.\n //\n // TODO(dimond): Ideally we would have an explicit lookup target\n // instead resulting in an explicit delete message and we could\n // remove this special logic.\n const key = new DocumentKey(targetData.target.path);\n if (\n this.pendingDocumentUpdates.get(key) === null &&\n !this.targetContainsDocument(targetId, key)\n ) {\n this.removeDocumentFromTarget(\n targetId,\n key,\n new NoDocument(key, snapshotVersion)\n );\n }\n }\n\n if (targetState.hasPendingChanges) {\n targetChanges.set(targetId, targetState.toTargetChange());\n targetState.clearPendingChanges();\n }\n }\n });\n\n let resolvedLimboDocuments = documentKeySet();\n\n // We extract the set of limbo-only document updates as the GC logic\n // special-cases documents that do not appear in the target cache.\n //\n // TODO(gsoltis): Expand on this comment once GC is available in the JS\n // client.\n this.pendingDocumentTargetMapping.forEach((key, targets) => {\n let isOnlyLimboTarget = true;\n\n targets.forEachWhile(targetId => {\n const targetData = this.targetDataForActiveTarget(targetId);\n if (\n targetData &&\n targetData.purpose !== TargetPurpose.LimboResolution\n ) {\n isOnlyLimboTarget = false;\n return false;\n }\n\n return true;\n });\n\n if (isOnlyLimboTarget) {\n resolvedLimboDocuments = resolvedLimboDocuments.add(key);\n }\n });\n\n const remoteEvent = new RemoteEvent(\n snapshotVersion,\n targetChanges,\n this.pendingTargetResets,\n this.pendingDocumentUpdates,\n resolvedLimboDocuments\n );\n\n this.pendingDocumentUpdates = maybeDocumentMap();\n this.pendingDocumentTargetMapping = documentTargetMap();\n this.pendingTargetResets = new SortedSet<TargetId>(primitiveComparator);\n\n return remoteEvent;\n }\n\n /**\n * Adds the provided document to the internal list of document updates and\n * its document key to the given target's mapping.\n */\n // Visible for testing.\n addDocumentToTarget(targetId: TargetId, document: MaybeDocument): void {\n if (!this.isActiveTarget(targetId)) {\n return;\n }\n\n const changeType = this.targetContainsDocument(targetId, document.key)\n ? ChangeType.Modified\n : ChangeType.Added;\n\n const targetState = this.ensureTargetState(targetId);\n targetState.addDocumentChange(document.key, changeType);\n\n this.pendingDocumentUpdates = this.pendingDocumentUpdates.insert(\n document.key,\n document\n );\n\n this.pendingDocumentTargetMapping = this.pendingDocumentTargetMapping.insert(\n document.key,\n this.ensureDocumentTargetMapping(document.key).add(targetId)\n );\n }\n\n /**\n * Removes the provided document from the target mapping. If the\n * document no longer matches the target, but the document's state is still\n * known (e.g. we know that the document was deleted or we received the change\n * that caused the filter mismatch), the new document can be provided\n * to update the remote document cache.\n */\n // Visible for testing.\n removeDocumentFromTarget(\n targetId: TargetId,\n key: DocumentKey,\n updatedDocument: MaybeDocument | null\n ): void {\n if (!this.isActiveTarget(targetId)) {\n return;\n }\n\n const targetState = this.ensureTargetState(targetId);\n if (this.targetContainsDocument(targetId, key)) {\n targetState.addDocumentChange(key, ChangeType.Removed);\n } else {\n // The document may have entered and left the target before we raised a\n // snapshot, so we can just ignore the change.\n targetState.removeDocumentChange(key);\n }\n\n this.pendingDocumentTargetMapping = this.pendingDocumentTargetMapping.insert(\n key,\n this.ensureDocumentTargetMapping(key).delete(targetId)\n );\n\n if (updatedDocument) {\n this.pendingDocumentUpdates = this.pendingDocumentUpdates.insert(\n key,\n updatedDocument\n );\n }\n }\n\n removeTarget(targetId: TargetId): void {\n this.targetStates.delete(targetId);\n }\n\n /**\n * Returns the current count of documents in the target. This includes both\n * the number of documents that the LocalStore considers to be part of the\n * target as well as any accumulated changes.\n */\n private getCurrentDocumentCountForTarget(targetId: TargetId): number {\n const targetState = this.ensureTargetState(targetId);\n const targetChange = targetState.toTargetChange();\n return (\n this.metadataProvider.getRemoteKeysForTarget(targetId).size +\n targetChange.addedDocuments.size -\n targetChange.removedDocuments.size\n );\n }\n\n /**\n * Increment the number of acks needed from watch before we can consider the\n * server to be 'in-sync' with the client's active targets.\n */\n recordPendingTargetRequest(targetId: TargetId): void {\n // For each request we get we need to record we need a response for it.\n const targetState = this.ensureTargetState(targetId);\n targetState.recordPendingTargetRequest();\n }\n\n private ensureTargetState(targetId: TargetId): TargetState {\n let result = this.targetStates.get(targetId);\n if (!result) {\n result = new TargetState();\n this.targetStates.set(targetId, result);\n }\n return result;\n }\n\n private ensureDocumentTargetMapping(key: DocumentKey): SortedSet<TargetId> {\n let targetMapping = this.pendingDocumentTargetMapping.get(key);\n\n if (!targetMapping) {\n targetMapping = new SortedSet<TargetId>(primitiveComparator);\n this.pendingDocumentTargetMapping = this.pendingDocumentTargetMapping.insert(\n key,\n targetMapping\n );\n }\n\n return targetMapping;\n }\n\n /**\n * Verifies that the user is still interested in this target (by calling\n * `getTargetDataForTarget()`) and that we are not waiting for pending ADDs\n * from watch.\n */\n protected isActiveTarget(targetId: TargetId): boolean {\n const targetActive = this.targetDataForActiveTarget(targetId) !== null;\n if (!targetActive) {\n logDebug(LOG_TAG, 'Detected inactive target', targetId);\n }\n return targetActive;\n }\n\n /**\n * Returns the TargetData for an active target (i.e. a target that the user\n * is still interested in that has no outstanding target change requests).\n */\n protected targetDataForActiveTarget(targetId: TargetId): TargetData | null {\n const targetState = this.targetStates.get(targetId);\n return targetState && targetState.isPending\n ? null\n : this.metadataProvider.getTargetDataForTarget(targetId);\n }\n\n /**\n * Resets the state of a Watch target to its initial state (e.g. sets\n * 'current' to false, clears the resume token and removes its target mapping\n * from all documents).\n */\n private resetTarget(targetId: TargetId): void {\n debugAssert(\n !this.targetStates.get(targetId)!.isPending,\n 'Should only reset active targets'\n );\n this.targetStates.set(targetId, new TargetState());\n\n // Trigger removal for any documents currently mapped to this target.\n // These removals will be part of the initial snapshot if Watch does not\n // resend these documents.\n const existingKeys = this.metadataProvider.getRemoteKeysForTarget(targetId);\n existingKeys.forEach(key => {\n this.removeDocumentFromTarget(targetId, key, /*updatedDocument=*/ null);\n });\n }\n /**\n * Returns whether the LocalStore considers the document to be part of the\n * specified target.\n */\n private targetContainsDocument(\n targetId: TargetId,\n key: DocumentKey\n ): boolean {\n const existingKeys = this.metadataProvider.getRemoteKeysForTarget(targetId);\n return existingKeys.has(key);\n }\n}\n\nfunction documentTargetMap(): SortedMap<DocumentKey, SortedSet<TargetId>> {\n return new SortedMap<DocumentKey, SortedSet<TargetId>>(\n DocumentKey.comparator\n );\n}\n\nfunction snapshotChangesMap(): SortedMap<DocumentKey, ChangeType> {\n return new SortedMap<DocumentKey, ChangeType>(DocumentKey.comparator);\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Blob } from '../api/blob';\nimport { Timestamp } from '../api/timestamp';\nimport { DatabaseId } from '../core/database_info';\nimport {\n Bound,\n Direction,\n FieldFilter,\n Filter,\n LimitType,\n Operator,\n OrderBy,\n Query\n} from '../core/query';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { Target } from '../core/target';\nimport { TargetId } from '../core/types';\nimport { TargetData, TargetPurpose } from '../local/target_data';\nimport { Document, MaybeDocument, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { ObjectValue } from '../model/object_value';\nimport {\n DeleteMutation,\n FieldMask,\n FieldTransform,\n Mutation,\n MutationResult,\n PatchMutation,\n Precondition,\n SetMutation,\n TransformMutation,\n VerifyMutation\n} from '../model/mutation';\nimport { FieldPath, ResourcePath } from '../model/path';\nimport * as api from '../protos/firestore_proto_api';\nimport { debugAssert, fail, hardAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { ByteString } from '../util/byte_string';\nimport {\n isNegativeZero,\n isNullOrUndefined,\n isSafeInteger\n} from '../util/types';\nimport {\n ArrayRemoveTransformOperation,\n ArrayUnionTransformOperation,\n NumericIncrementTransformOperation,\n ServerTimestampTransform,\n TransformOperation\n} from '../model/transform_operation';\nimport { ExistenceFilter } from './existence_filter';\nimport { mapCodeFromRpcCode } from './rpc_error';\nimport {\n DocumentWatchChange,\n ExistenceFilterChange,\n WatchChange,\n WatchTargetChange,\n WatchTargetChangeState\n} from './watch_change';\nimport { isNanValue, isNullValue, normalizeTimestamp } from '../model/values';\n\nconst DIRECTIONS = (() => {\n const dirs: { [dir: string]: api.OrderDirection } = {};\n dirs[Direction.ASCENDING] = 'ASCENDING';\n dirs[Direction.DESCENDING] = 'DESCENDING';\n return dirs;\n})();\n\nconst OPERATORS = (() => {\n const ops: { [op: string]: api.FieldFilterOp } = {};\n ops[Operator.LESS_THAN] = 'LESS_THAN';\n ops[Operator.LESS_THAN_OR_EQUAL] = 'LESS_THAN_OR_EQUAL';\n ops[Operator.GREATER_THAN] = 'GREATER_THAN';\n ops[Operator.GREATER_THAN_OR_EQUAL] = 'GREATER_THAN_OR_EQUAL';\n ops[Operator.EQUAL] = 'EQUAL';\n ops[Operator.ARRAY_CONTAINS] = 'ARRAY_CONTAINS';\n ops[Operator.IN] = 'IN';\n ops[Operator.ARRAY_CONTAINS_ANY] = 'ARRAY_CONTAINS_ANY';\n return ops;\n})();\n\nfunction assertPresent(value: unknown, description: string): asserts value {\n debugAssert(!isNullOrUndefined(value), description + ' is missing');\n}\n\nexport interface SerializerOptions {\n /**\n * The serializer supports both Protobuf.js and Proto3 JSON formats. By\n * setting this flag to true, the serializer will use the Proto3 JSON format.\n *\n * For a description of the Proto3 JSON format check\n * https://developers.google.com/protocol-buffers/docs/proto3#json\n */\n useProto3Json: boolean;\n}\n\n/**\n * Generates JsonObject values for the Datastore API suitable for sending to\n * either GRPC stub methods or via the JSON/HTTP REST API.\n * TODO(klimt): We can remove the databaseId argument if we keep the full\n * resource name in documents.\n */\nexport class JsonProtoSerializer {\n constructor(\n private databaseId: DatabaseId,\n private options: SerializerOptions\n ) {}\n\n fromRpcStatus(status: api.Status): FirestoreError {\n const code =\n status.code === undefined\n ? Code.UNKNOWN\n : mapCodeFromRpcCode(status.code);\n return new FirestoreError(code, status.message || '');\n }\n\n /**\n * Returns a value for a number (or null) that's appropriate to put into\n * a google.protobuf.Int32Value proto.\n * DO NOT USE THIS FOR ANYTHING ELSE.\n * This method cheats. It's typed as returning \"number\" because that's what\n * our generated proto interfaces say Int32Value must be. But GRPC actually\n * expects a { value: <number> } struct.\n */\n private toInt32Proto(val: number | null): number | { value: number } | null {\n if (this.options.useProto3Json || isNullOrUndefined(val)) {\n return val;\n } else {\n return { value: val };\n }\n }\n\n /**\n * Returns a number (or null) from a google.protobuf.Int32Value proto.\n */\n private fromInt32Proto(\n val: number | { value: number } | undefined\n ): number | null {\n let result;\n if (typeof val === 'object') {\n result = val.value;\n } else {\n result = val;\n }\n return isNullOrUndefined(result) ? null : result;\n }\n\n /**\n * Returns an IntegerValue for `value`.\n */\n toInteger(value: number): api.Value {\n return { integerValue: '' + value };\n }\n\n /**\n * Returns an DoubleValue for `value` that is encoded based the serializer's\n * `useProto3Json` setting.\n */\n toDouble(value: number): api.Value {\n if (this.options.useProto3Json) {\n if (isNaN(value)) {\n return { doubleValue: 'NaN' };\n } else if (value === Infinity) {\n return { doubleValue: 'Infinity' };\n } else if (value === -Infinity) {\n return { doubleValue: '-Infinity' };\n }\n }\n return { doubleValue: isNegativeZero(value) ? '-0' : value };\n }\n\n /**\n * Returns a value for a number that's appropriate to put into a proto.\n * The return value is an IntegerValue if it can safely represent the value,\n * otherwise a DoubleValue is returned.\n */\n toNumber(value: number): api.Value {\n return isSafeInteger(value) ? this.toInteger(value) : this.toDouble(value);\n }\n\n /**\n * Returns a value for a Date that's appropriate to put into a proto.\n */\n toTimestamp(timestamp: Timestamp): api.Timestamp {\n if (this.options.useProto3Json) {\n // Serialize to ISO-8601 date format, but with full nano resolution.\n // Since JS Date has only millis, let's only use it for the seconds and\n // then manually add the fractions to the end.\n const jsDateStr = new Date(timestamp.seconds * 1000).toISOString();\n // Remove .xxx frac part and Z in the end.\n const strUntilSeconds = jsDateStr.replace(/\\.\\d*/, '').replace('Z', '');\n // Pad the fraction out to 9 digits (nanos).\n const nanoStr = ('000000000' + timestamp.nanoseconds).slice(-9);\n\n return `${strUntilSeconds}.${nanoStr}Z`;\n } else {\n return {\n seconds: '' + timestamp.seconds,\n nanos: timestamp.nanoseconds\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n } as any;\n }\n }\n\n private fromTimestamp(date: api.Timestamp): Timestamp {\n const timestamp = normalizeTimestamp(date);\n return new Timestamp(timestamp.seconds, timestamp.nanos);\n }\n\n /**\n * Returns a value for bytes that's appropriate to put in a proto.\n *\n * Visible for testing.\n */\n toBytes(bytes: Blob | ByteString): string | Uint8Array {\n if (this.options.useProto3Json) {\n return bytes.toBase64();\n } else {\n return bytes.toUint8Array();\n }\n }\n\n /**\n * Returns a ByteString based on the proto string value.\n */\n fromBytes(value: string | Uint8Array | undefined): ByteString {\n if (this.options.useProto3Json) {\n hardAssert(\n value === undefined || typeof value === 'string',\n 'value must be undefined or a string when using proto3 Json'\n );\n return ByteString.fromBase64String(value ? value : '');\n } else {\n hardAssert(\n value === undefined || value instanceof Uint8Array,\n 'value must be undefined or Uint8Array'\n );\n return ByteString.fromUint8Array(value ? value : new Uint8Array());\n }\n }\n\n toVersion(version: SnapshotVersion): api.Timestamp {\n return this.toTimestamp(version.toTimestamp());\n }\n\n fromVersion(version: api.Timestamp): SnapshotVersion {\n hardAssert(!!version, \"Trying to deserialize version that isn't set\");\n return SnapshotVersion.fromTimestamp(this.fromTimestamp(version));\n }\n\n toResourceName(path: ResourcePath, databaseId?: DatabaseId): string {\n return this.fullyQualifiedPrefixPath(databaseId || this.databaseId)\n .child('documents')\n .child(path)\n .canonicalString();\n }\n\n fromResourceName(name: string): ResourcePath {\n const resource = ResourcePath.fromString(name);\n hardAssert(\n isValidResourceName(resource),\n 'Tried to deserialize invalid key ' + resource.toString()\n );\n return resource;\n }\n\n toName(key: DocumentKey): string {\n return this.toResourceName(key.path);\n }\n\n fromName(name: string): DocumentKey {\n const resource = this.fromResourceName(name);\n hardAssert(\n resource.get(1) === this.databaseId.projectId,\n 'Tried to deserialize key from different project: ' +\n resource.get(1) +\n ' vs ' +\n this.databaseId.projectId\n );\n hardAssert(\n (!resource.get(3) && !this.databaseId.database) ||\n resource.get(3) === this.databaseId.database,\n 'Tried to deserialize key from different database: ' +\n resource.get(3) +\n ' vs ' +\n this.databaseId.database\n );\n return new DocumentKey(this.extractLocalPathFromResourceName(resource));\n }\n\n toQueryPath(path: ResourcePath): string {\n return this.toResourceName(path);\n }\n\n fromQueryPath(name: string): ResourcePath {\n const resourceName = this.fromResourceName(name);\n // In v1beta1 queries for collections at the root did not have a trailing\n // \"/documents\". In v1 all resource paths contain \"/documents\". Preserve the\n // ability to read the v1beta1 form for compatibility with queries persisted\n // in the local target cache.\n if (resourceName.length === 4) {\n return ResourcePath.EMPTY_PATH;\n }\n return this.extractLocalPathFromResourceName(resourceName);\n }\n\n get encodedDatabaseId(): string {\n const path = new ResourcePath([\n 'projects',\n this.databaseId.projectId,\n 'databases',\n this.databaseId.database\n ]);\n return path.canonicalString();\n }\n\n private fullyQualifiedPrefixPath(databaseId: DatabaseId): ResourcePath {\n return new ResourcePath([\n 'projects',\n databaseId.projectId,\n 'databases',\n databaseId.database\n ]);\n }\n\n private extractLocalPathFromResourceName(\n resourceName: ResourcePath\n ): ResourcePath {\n hardAssert(\n resourceName.length > 4 && resourceName.get(4) === 'documents',\n 'tried to deserialize invalid key ' + resourceName.toString()\n );\n return resourceName.popFirst(5);\n }\n\n /** Creates an api.Document from key and fields (but no create/update time) */\n toMutationDocument(key: DocumentKey, fields: ObjectValue): api.Document {\n return {\n name: this.toName(key),\n fields: fields.proto.mapValue.fields\n };\n }\n\n toDocument(document: Document): api.Document {\n debugAssert(\n !document.hasLocalMutations,\n \"Can't serialize documents with mutations.\"\n );\n return {\n name: this.toName(document.key),\n fields: document.toProto().mapValue.fields,\n updateTime: this.toTimestamp(document.version.toTimestamp())\n };\n }\n\n fromDocument(\n document: api.Document,\n hasCommittedMutations?: boolean\n ): Document {\n const key = this.fromName(document.name!);\n const version = this.fromVersion(document.updateTime!);\n const data = new ObjectValue({ mapValue: { fields: document.fields } });\n return new Document(key, version, data, {\n hasCommittedMutations: !!hasCommittedMutations\n });\n }\n\n private fromFound(doc: api.BatchGetDocumentsResponse): Document {\n hardAssert(\n !!doc.found,\n 'Tried to deserialize a found document from a missing document.'\n );\n assertPresent(doc.found.name, 'doc.found.name');\n assertPresent(doc.found.updateTime, 'doc.found.updateTime');\n const key = this.fromName(doc.found.name);\n const version = this.fromVersion(doc.found.updateTime);\n const data = new ObjectValue({ mapValue: { fields: doc.found.fields } });\n return new Document(key, version, data, {});\n }\n\n private fromMissing(result: api.BatchGetDocumentsResponse): NoDocument {\n hardAssert(\n !!result.missing,\n 'Tried to deserialize a missing document from a found document.'\n );\n hardAssert(\n !!result.readTime,\n 'Tried to deserialize a missing document without a read time.'\n );\n const key = this.fromName(result.missing);\n const version = this.fromVersion(result.readTime);\n return new NoDocument(key, version);\n }\n\n fromMaybeDocument(result: api.BatchGetDocumentsResponse): MaybeDocument {\n if ('found' in result) {\n return this.fromFound(result);\n } else if ('missing' in result) {\n return this.fromMissing(result);\n }\n return fail('invalid batch get response: ' + JSON.stringify(result));\n }\n\n fromWatchChange(change: api.ListenResponse): WatchChange {\n let watchChange: WatchChange;\n if ('targetChange' in change) {\n assertPresent(change.targetChange, 'targetChange');\n // proto3 default value is unset in JSON (undefined), so use 'NO_CHANGE'\n // if unset\n const state = this.fromWatchTargetChangeState(\n change.targetChange.targetChangeType || 'NO_CHANGE'\n );\n const targetIds: TargetId[] = change.targetChange.targetIds || [];\n\n const resumeToken = this.fromBytes(change.targetChange.resumeToken);\n const causeProto = change.targetChange!.cause;\n const cause = causeProto && this.fromRpcStatus(causeProto);\n watchChange = new WatchTargetChange(\n state,\n targetIds,\n resumeToken,\n cause || null\n );\n } else if ('documentChange' in change) {\n assertPresent(change.documentChange, 'documentChange');\n const entityChange = change.documentChange;\n assertPresent(entityChange.document, 'documentChange.name');\n assertPresent(entityChange.document.name, 'documentChange.document.name');\n assertPresent(\n entityChange.document.updateTime,\n 'documentChange.document.updateTime'\n );\n const key = this.fromName(entityChange.document.name);\n const version = this.fromVersion(entityChange.document.updateTime);\n const data = new ObjectValue({\n mapValue: { fields: entityChange.document.fields }\n });\n const doc = new Document(key, version, data, {});\n const updatedTargetIds = entityChange.targetIds || [];\n const removedTargetIds = entityChange.removedTargetIds || [];\n watchChange = new DocumentWatchChange(\n updatedTargetIds,\n removedTargetIds,\n doc.key,\n doc\n );\n } else if ('documentDelete' in change) {\n assertPresent(change.documentDelete, 'documentDelete');\n const docDelete = change.documentDelete;\n assertPresent(docDelete.document, 'documentDelete.document');\n const key = this.fromName(docDelete.document);\n const version = docDelete.readTime\n ? this.fromVersion(docDelete.readTime)\n : SnapshotVersion.min();\n const doc = new NoDocument(key, version);\n const removedTargetIds = docDelete.removedTargetIds || [];\n watchChange = new DocumentWatchChange([], removedTargetIds, doc.key, doc);\n } else if ('documentRemove' in change) {\n assertPresent(change.documentRemove, 'documentRemove');\n const docRemove = change.documentRemove;\n assertPresent(docRemove.document, 'documentRemove');\n const key = this.fromName(docRemove.document);\n const removedTargetIds = docRemove.removedTargetIds || [];\n watchChange = new DocumentWatchChange([], removedTargetIds, key, null);\n } else if ('filter' in change) {\n // TODO(dimond): implement existence filter parsing with strategy.\n assertPresent(change.filter, 'filter');\n const filter = change.filter;\n assertPresent(filter.targetId, 'filter.targetId');\n const count = filter.count || 0;\n const existenceFilter = new ExistenceFilter(count);\n const targetId = filter.targetId;\n watchChange = new ExistenceFilterChange(targetId, existenceFilter);\n } else {\n return fail('Unknown change type ' + JSON.stringify(change));\n }\n return watchChange;\n }\n\n fromWatchTargetChangeState(\n state: api.TargetChangeTargetChangeType\n ): WatchTargetChangeState {\n if (state === 'NO_CHANGE') {\n return WatchTargetChangeState.NoChange;\n } else if (state === 'ADD') {\n return WatchTargetChangeState.Added;\n } else if (state === 'REMOVE') {\n return WatchTargetChangeState.Removed;\n } else if (state === 'CURRENT') {\n return WatchTargetChangeState.Current;\n } else if (state === 'RESET') {\n return WatchTargetChangeState.Reset;\n } else {\n return fail('Got unexpected TargetChange.state: ' + state);\n }\n }\n\n versionFromListenResponse(change: api.ListenResponse): SnapshotVersion {\n // We have only reached a consistent snapshot for the entire stream if there\n // is a read_time set and it applies to all targets (i.e. the list of\n // targets is empty). The backend is guaranteed to send such responses.\n if (!('targetChange' in change)) {\n return SnapshotVersion.min();\n }\n const targetChange = change.targetChange!;\n if (targetChange.targetIds && targetChange.targetIds.length) {\n return SnapshotVersion.min();\n }\n if (!targetChange.readTime) {\n return SnapshotVersion.min();\n }\n return this.fromVersion(targetChange.readTime);\n }\n\n toMutation(mutation: Mutation): api.Write {\n let result: api.Write;\n if (mutation instanceof SetMutation) {\n result = {\n update: this.toMutationDocument(mutation.key, mutation.value)\n };\n } else if (mutation instanceof DeleteMutation) {\n result = { delete: this.toName(mutation.key) };\n } else if (mutation instanceof PatchMutation) {\n result = {\n update: this.toMutationDocument(mutation.key, mutation.data),\n updateMask: this.toDocumentMask(mutation.fieldMask)\n };\n } else if (mutation instanceof TransformMutation) {\n result = {\n transform: {\n document: this.toName(mutation.key),\n fieldTransforms: mutation.fieldTransforms.map(transform =>\n this.toFieldTransform(transform)\n )\n }\n };\n } else if (mutation instanceof VerifyMutation) {\n result = {\n verify: this.toName(mutation.key)\n };\n } else {\n return fail('Unknown mutation type ' + mutation.type);\n }\n\n if (!mutation.precondition.isNone) {\n result.currentDocument = this.toPrecondition(mutation.precondition);\n }\n\n return result;\n }\n\n fromMutation(proto: api.Write): Mutation {\n const precondition = proto.currentDocument\n ? this.fromPrecondition(proto.currentDocument)\n : Precondition.none();\n\n if (proto.update) {\n assertPresent(proto.update.name, 'name');\n const key = this.fromName(proto.update.name);\n const value = new ObjectValue({\n mapValue: { fields: proto.update.fields }\n });\n if (proto.updateMask) {\n const fieldMask = this.fromDocumentMask(proto.updateMask);\n return new PatchMutation(key, value, fieldMask, precondition);\n } else {\n return new SetMutation(key, value, precondition);\n }\n } else if (proto.delete) {\n const key = this.fromName(proto.delete);\n return new DeleteMutation(key, precondition);\n } else if (proto.transform) {\n const key = this.fromName(proto.transform.document!);\n const fieldTransforms = proto.transform.fieldTransforms!.map(transform =>\n this.fromFieldTransform(transform)\n );\n hardAssert(\n precondition.exists === true,\n 'Transforms only support precondition \"exists == true\"'\n );\n return new TransformMutation(key, fieldTransforms);\n } else if (proto.verify) {\n const key = this.fromName(proto.verify);\n return new VerifyMutation(key, precondition);\n } else {\n return fail('unknown mutation proto: ' + JSON.stringify(proto));\n }\n }\n\n private toPrecondition(precondition: Precondition): api.Precondition {\n debugAssert(!precondition.isNone, \"Can't serialize an empty precondition\");\n if (precondition.updateTime !== undefined) {\n return {\n updateTime: this.toVersion(precondition.updateTime)\n };\n } else if (precondition.exists !== undefined) {\n return { exists: precondition.exists };\n } else {\n return fail('Unknown precondition');\n }\n }\n\n private fromPrecondition(precondition: api.Precondition): Precondition {\n if (precondition.updateTime !== undefined) {\n return Precondition.updateTime(this.fromVersion(precondition.updateTime));\n } else if (precondition.exists !== undefined) {\n return Precondition.exists(precondition.exists);\n } else {\n return Precondition.none();\n }\n }\n\n private fromWriteResult(\n proto: api.WriteResult,\n commitTime: api.Timestamp\n ): MutationResult {\n // NOTE: Deletes don't have an updateTime.\n let version = proto.updateTime\n ? this.fromVersion(proto.updateTime)\n : this.fromVersion(commitTime);\n\n if (version.isEqual(SnapshotVersion.min())) {\n // The Firestore Emulator currently returns an update time of 0 for\n // deletes of non-existing documents (rather than null). This breaks the\n // test \"get deleted doc while offline with source=cache\" as NoDocuments\n // with version 0 are filtered by IndexedDb's RemoteDocumentCache.\n // TODO(#2149): Remove this when Emulator is fixed\n version = this.fromVersion(commitTime);\n }\n\n let transformResults: api.Value[] | null = null;\n if (proto.transformResults && proto.transformResults.length > 0) {\n transformResults = proto.transformResults;\n }\n return new MutationResult(version, transformResults);\n }\n\n fromWriteResults(\n protos: api.WriteResult[] | undefined,\n commitTime?: api.Timestamp\n ): MutationResult[] {\n if (protos && protos.length > 0) {\n hardAssert(\n commitTime !== undefined,\n 'Received a write result without a commit time'\n );\n return protos.map(proto => this.fromWriteResult(proto, commitTime));\n } else {\n return [];\n }\n }\n\n private toFieldTransform(fieldTransform: FieldTransform): api.FieldTransform {\n const transform = fieldTransform.transform;\n if (transform instanceof ServerTimestampTransform) {\n return {\n fieldPath: fieldTransform.field.canonicalString(),\n setToServerValue: 'REQUEST_TIME'\n };\n } else if (transform instanceof ArrayUnionTransformOperation) {\n return {\n fieldPath: fieldTransform.field.canonicalString(),\n appendMissingElements: {\n values: transform.elements\n }\n };\n } else if (transform instanceof ArrayRemoveTransformOperation) {\n return {\n fieldPath: fieldTransform.field.canonicalString(),\n removeAllFromArray: {\n values: transform.elements\n }\n };\n } else if (transform instanceof NumericIncrementTransformOperation) {\n return {\n fieldPath: fieldTransform.field.canonicalString(),\n increment: transform.operand\n };\n } else {\n throw fail('Unknown transform: ' + fieldTransform.transform);\n }\n }\n\n private fromFieldTransform(proto: api.FieldTransform): FieldTransform {\n let transform: TransformOperation | null = null;\n if ('setToServerValue' in proto) {\n hardAssert(\n proto.setToServerValue === 'REQUEST_TIME',\n 'Unknown server value transform proto: ' + JSON.stringify(proto)\n );\n transform = ServerTimestampTransform.instance;\n } else if ('appendMissingElements' in proto) {\n const values = proto.appendMissingElements!.values || [];\n transform = new ArrayUnionTransformOperation(values);\n } else if ('removeAllFromArray' in proto) {\n const values = proto.removeAllFromArray!.values || [];\n transform = new ArrayRemoveTransformOperation(values);\n } else if ('increment' in proto) {\n transform = new NumericIncrementTransformOperation(\n this,\n proto.increment!\n );\n } else {\n fail('Unknown transform proto: ' + JSON.stringify(proto));\n }\n const fieldPath = FieldPath.fromServerFormat(proto.fieldPath!);\n return new FieldTransform(fieldPath, transform!);\n }\n\n toDocumentsTarget(target: Target): api.DocumentsTarget {\n return { documents: [this.toQueryPath(target.path)] };\n }\n\n fromDocumentsTarget(documentsTarget: api.DocumentsTarget): Target {\n const count = documentsTarget.documents!.length;\n hardAssert(\n count === 1,\n 'DocumentsTarget contained other than 1 document: ' + count\n );\n const name = documentsTarget.documents![0];\n return Query.atPath(this.fromQueryPath(name)).toTarget();\n }\n\n toQueryTarget(target: Target): api.QueryTarget {\n // Dissect the path into parent, collectionId, and optional key filter.\n const result: api.QueryTarget = { structuredQuery: {} };\n const path = target.path;\n if (target.collectionGroup !== null) {\n debugAssert(\n path.length % 2 === 0,\n 'Collection Group queries should be within a document path or root.'\n );\n result.parent = this.toQueryPath(path);\n result.structuredQuery!.from = [\n {\n collectionId: target.collectionGroup,\n allDescendants: true\n }\n ];\n } else {\n debugAssert(\n path.length % 2 !== 0,\n 'Document queries with filters are not supported.'\n );\n result.parent = this.toQueryPath(path.popLast());\n result.structuredQuery!.from = [{ collectionId: path.lastSegment() }];\n }\n\n const where = this.toFilter(target.filters);\n if (where) {\n result.structuredQuery!.where = where;\n }\n\n const orderBy = this.toOrder(target.orderBy);\n if (orderBy) {\n result.structuredQuery!.orderBy = orderBy;\n }\n\n const limit = this.toInt32Proto(target.limit);\n if (limit !== null) {\n result.structuredQuery!.limit = limit;\n }\n\n if (target.startAt) {\n result.structuredQuery!.startAt = this.toCursor(target.startAt);\n }\n if (target.endAt) {\n result.structuredQuery!.endAt = this.toCursor(target.endAt);\n }\n\n return result;\n }\n\n fromQueryTarget(target: api.QueryTarget): Target {\n let path = this.fromQueryPath(target.parent!);\n\n const query = target.structuredQuery!;\n const fromCount = query.from ? query.from.length : 0;\n let collectionGroup: string | null = null;\n if (fromCount > 0) {\n hardAssert(\n fromCount === 1,\n 'StructuredQuery.from with more than one collection is not supported.'\n );\n const from = query.from![0];\n if (from.allDescendants) {\n collectionGroup = from.collectionId!;\n } else {\n path = path.child(from.collectionId!);\n }\n }\n\n let filterBy: Filter[] = [];\n if (query.where) {\n filterBy = this.fromFilter(query.where);\n }\n\n let orderBy: OrderBy[] = [];\n if (query.orderBy) {\n orderBy = this.fromOrder(query.orderBy);\n }\n\n let limit: number | null = null;\n if (query.limit) {\n limit = this.fromInt32Proto(query.limit);\n }\n\n let startAt: Bound | null = null;\n if (query.startAt) {\n startAt = this.fromCursor(query.startAt);\n }\n\n let endAt: Bound | null = null;\n if (query.endAt) {\n endAt = this.fromCursor(query.endAt);\n }\n\n return new Query(\n path,\n collectionGroup,\n orderBy,\n filterBy,\n limit,\n LimitType.First,\n startAt,\n endAt\n ).toTarget();\n }\n\n toListenRequestLabels(\n targetData: TargetData\n ): api.ApiClientObjectMap<string> | null {\n const value = this.toLabel(targetData.purpose);\n if (value == null) {\n return null;\n } else {\n return {\n 'goog-listen-tags': value\n };\n }\n }\n\n private toLabel(purpose: TargetPurpose): string | null {\n switch (purpose) {\n case TargetPurpose.Listen:\n return null;\n case TargetPurpose.ExistenceFilterMismatch:\n return 'existence-filter-mismatch';\n case TargetPurpose.LimboResolution:\n return 'limbo-document';\n default:\n return fail('Unrecognized query purpose: ' + purpose);\n }\n }\n\n toTarget(targetData: TargetData): api.Target {\n let result: api.Target;\n const target = targetData.target;\n\n if (target.isDocumentQuery()) {\n result = { documents: this.toDocumentsTarget(target) };\n } else {\n result = { query: this.toQueryTarget(target) };\n }\n\n result.targetId = targetData.targetId;\n\n if (targetData.resumeToken.approximateByteSize() > 0) {\n result.resumeToken = this.toBytes(targetData.resumeToken);\n }\n\n return result;\n }\n\n private toFilter(filters: Filter[]): api.Filter | undefined {\n if (filters.length === 0) {\n return;\n }\n const protos = filters.map(filter => {\n if (filter instanceof FieldFilter) {\n return this.toUnaryOrFieldFilter(filter);\n } else {\n return fail('Unrecognized filter: ' + JSON.stringify(filter));\n }\n });\n if (protos.length === 1) {\n return protos[0];\n }\n return { compositeFilter: { op: 'AND', filters: protos } };\n }\n\n private fromFilter(filter: api.Filter | undefined): Filter[] {\n if (!filter) {\n return [];\n } else if (filter.unaryFilter !== undefined) {\n return [this.fromUnaryFilter(filter)];\n } else if (filter.fieldFilter !== undefined) {\n return [this.fromFieldFilter(filter)];\n } else if (filter.compositeFilter !== undefined) {\n return filter.compositeFilter\n .filters!.map(f => this.fromFilter(f))\n .reduce((accum, current) => accum.concat(current));\n } else {\n return fail('Unknown filter: ' + JSON.stringify(filter));\n }\n }\n\n private toOrder(orderBys: OrderBy[]): api.Order[] | undefined {\n if (orderBys.length === 0) {\n return;\n }\n return orderBys.map(order => this.toPropertyOrder(order));\n }\n\n private fromOrder(orderBys: api.Order[]): OrderBy[] {\n return orderBys.map(order => this.fromPropertyOrder(order));\n }\n\n private toCursor(cursor: Bound): api.Cursor {\n return {\n before: cursor.before,\n values: cursor.position\n };\n }\n\n private fromCursor(cursor: api.Cursor): Bound {\n const before = !!cursor.before;\n const position = cursor.values || [];\n return new Bound(position, before);\n }\n\n // visible for testing\n toDirection(dir: Direction): api.OrderDirection {\n return DIRECTIONS[dir];\n }\n\n // visible for testing\n fromDirection(dir: api.OrderDirection | undefined): Direction | undefined {\n switch (dir) {\n case 'ASCENDING':\n return Direction.ASCENDING;\n case 'DESCENDING':\n return Direction.DESCENDING;\n default:\n return undefined;\n }\n }\n\n // visible for testing\n toOperatorName(op: Operator): api.FieldFilterOp {\n return OPERATORS[op];\n }\n\n fromOperatorName(op: api.FieldFilterOp): Operator {\n switch (op) {\n case 'EQUAL':\n return Operator.EQUAL;\n case 'GREATER_THAN':\n return Operator.GREATER_THAN;\n case 'GREATER_THAN_OR_EQUAL':\n return Operator.GREATER_THAN_OR_EQUAL;\n case 'LESS_THAN':\n return Operator.LESS_THAN;\n case 'LESS_THAN_OR_EQUAL':\n return Operator.LESS_THAN_OR_EQUAL;\n case 'ARRAY_CONTAINS':\n return Operator.ARRAY_CONTAINS;\n case 'IN':\n return Operator.IN;\n case 'ARRAY_CONTAINS_ANY':\n return Operator.ARRAY_CONTAINS_ANY;\n case 'OPERATOR_UNSPECIFIED':\n return fail('Unspecified operator');\n default:\n return fail('Unknown operator');\n }\n }\n\n toFieldPathReference(path: FieldPath): api.FieldReference {\n return { fieldPath: path.canonicalString() };\n }\n\n fromFieldPathReference(fieldReference: api.FieldReference): FieldPath {\n return FieldPath.fromServerFormat(fieldReference.fieldPath!);\n }\n\n // visible for testing\n toPropertyOrder(orderBy: OrderBy): api.Order {\n return {\n field: this.toFieldPathReference(orderBy.field),\n direction: this.toDirection(orderBy.dir)\n };\n }\n\n fromPropertyOrder(orderBy: api.Order): OrderBy {\n return new OrderBy(\n this.fromFieldPathReference(orderBy.field!),\n this.fromDirection(orderBy.direction)\n );\n }\n\n fromFieldFilter(filter: api.Filter): Filter {\n return FieldFilter.create(\n this.fromFieldPathReference(filter.fieldFilter!.field!),\n this.fromOperatorName(filter.fieldFilter!.op!),\n filter.fieldFilter!.value!\n );\n }\n\n // visible for testing\n toUnaryOrFieldFilter(filter: FieldFilter): api.Filter {\n if (filter.op === Operator.EQUAL) {\n if (isNanValue(filter.value)) {\n return {\n unaryFilter: {\n field: this.toFieldPathReference(filter.field),\n op: 'IS_NAN'\n }\n };\n } else if (isNullValue(filter.value)) {\n return {\n unaryFilter: {\n field: this.toFieldPathReference(filter.field),\n op: 'IS_NULL'\n }\n };\n }\n }\n return {\n fieldFilter: {\n field: this.toFieldPathReference(filter.field),\n op: this.toOperatorName(filter.op),\n value: filter.value\n }\n };\n }\n\n fromUnaryFilter(filter: api.Filter): Filter {\n switch (filter.unaryFilter!.op!) {\n case 'IS_NAN':\n const nanField = this.fromFieldPathReference(\n filter.unaryFilter!.field!\n );\n return FieldFilter.create(nanField, Operator.EQUAL, {\n doubleValue: NaN\n });\n case 'IS_NULL':\n const nullField = this.fromFieldPathReference(\n filter.unaryFilter!.field!\n );\n return FieldFilter.create(nullField, Operator.EQUAL, {\n nullValue: 'NULL_VALUE'\n });\n case 'OPERATOR_UNSPECIFIED':\n return fail('Unspecified filter');\n default:\n return fail('Unknown filter');\n }\n }\n\n toDocumentMask(fieldMask: FieldMask): api.DocumentMask {\n const canonicalFields: string[] = [];\n fieldMask.fields.forEach(field =>\n canonicalFields.push(field.canonicalString())\n );\n return {\n fieldPaths: canonicalFields\n };\n }\n\n fromDocumentMask(proto: api.DocumentMask): FieldMask {\n const paths = proto.fieldPaths || [];\n return new FieldMask(paths.map(path => FieldPath.fromServerFormat(path)));\n }\n}\n\nexport function isValidResourceName(path: ResourcePath): boolean {\n // Resource names have at least 4 components (project ID, database ID)\n return (\n path.length >= 4 &&\n path.get(0) === 'projects' &&\n path.get(2) === 'databases'\n );\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { DatabaseId, DatabaseInfo } from '../core/database_info';\nimport { Connection } from '../remote/connection';\nimport { JsonProtoSerializer } from '../remote/serializer';\nimport { fail } from '../util/assert';\nimport { ConnectivityMonitor } from './../remote/connectivity_monitor';\n\n/**\n * Provides a common interface to load anything platform dependent, e.g.\n * the connection implementation.\n *\n * An implementation of this must be provided at compile time for the platform.\n */\n// TODO: Consider only exposing the APIs of 'document' and 'window' that we\n// use in our client.\nexport interface Platform {\n loadConnection(databaseInfo: DatabaseInfo): Promise<Connection>;\n newConnectivityMonitor(): ConnectivityMonitor;\n newSerializer(databaseId: DatabaseId): JsonProtoSerializer;\n\n /** Formats an object as a JSON string, suitable for logging. */\n formatJSON(value: unknown): string;\n\n /** Converts a Base64 encoded string to a binary string. */\n atob(encoded: string): string;\n\n /** Converts a binary string to a Base64 encoded string. */\n btoa(raw: string): string;\n\n /**\n * Generates `nBytes` of random bytes.\n *\n * If `nBytes < 0` , an error will be thrown.\n */\n randomBytes(nBytes: number): Uint8Array;\n\n /** The Platform's 'window' implementation or null if not available. */\n readonly window: Window | null;\n\n /** The Platform's 'document' implementation or null if not available. */\n readonly document: Document | null;\n\n /** True if and only if the Base64 conversion functions are available. */\n readonly base64Available: boolean;\n}\n\n/**\n * Provides singleton helpers where setup code can inject a platform at runtime.\n * setPlatform needs to be set before Firestore is used and must be set exactly\n * once.\n */\nexport class PlatformSupport {\n private static platform: Platform;\n static setPlatform(platform: Platform): void {\n if (PlatformSupport.platform) {\n fail('Platform already defined');\n }\n PlatformSupport.platform = platform;\n }\n\n static getPlatform(): Platform {\n if (!PlatformSupport.platform) {\n fail('Platform not set');\n }\n return PlatformSupport.platform;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Logger, LogLevel } from '@firebase/logger';\nimport { SDK_VERSION } from '../core/version';\nimport { PlatformSupport } from '../platform/platform';\n\nexport { LogLevel };\n\nconst logClient = new Logger('@firebase/firestore');\n\n// Helper methods are needed because variables can't be exported as read/write\nexport function getLogLevel(): LogLevel {\n return logClient.logLevel;\n}\n\nexport function setLogLevel(newLevel: LogLevel): void {\n logClient.logLevel = newLevel;\n}\n\nexport function logDebug(msg: string, ...obj: unknown[]): void {\n if (logClient.logLevel <= LogLevel.DEBUG) {\n const args = obj.map(argToString);\n logClient.debug(`Firestore (${SDK_VERSION}): ${msg}`, ...args);\n }\n}\n\nexport function logError(msg: string, ...obj: unknown[]): void {\n if (logClient.logLevel <= LogLevel.ERROR) {\n const args = obj.map(argToString);\n logClient.error(`Firestore (${SDK_VERSION}): ${msg}`, ...args);\n }\n}\n\n/**\n * Converts an additional log parameter to a string representation.\n */\nfunction argToString(obj: unknown): string | unknown {\n if (typeof obj === 'string') {\n return obj;\n } else {\n const platform = PlatformSupport.getPlatform();\n try {\n return platform.formatJSON(obj);\n } catch (e) {\n // Converting to JSON failed, just log the object directly\n return obj;\n }\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SDK_VERSION } from '../core/version';\nimport { logError } from './log';\n\n/**\n * Unconditionally fails, throwing an Error with the given message.\n * Messages are stripped in production builds.\n *\n * Returns `never` and can be used in expressions:\n * @example\n * let futureVar = fail('not implemented yet');\n */\nexport function fail(failure: string = 'Unexpected state'): never {\n // Log the failure in addition to throw an exception, just in case the\n // exception is swallowed.\n const message =\n `FIRESTORE (${SDK_VERSION}) INTERNAL ASSERTION FAILED: ` + failure;\n logError(message);\n\n // NOTE: We don't use FirestoreError here because these are internal failures\n // that cannot be handled by the user. (Also it would create a circular\n // dependency between the error and assert modules which doesn't work.)\n throw new Error(message);\n}\n\n/**\n * Fails if the given assertion condition is false, throwing an Error with the\n * given message if it did.\n *\n * Messages are stripped in production builds.\n */\nexport function hardAssert(\n assertion: boolean,\n message?: string\n): asserts assertion {\n if (!assertion) {\n fail(message);\n }\n}\n\n/**\n * Fails if the given assertion condition is false, throwing an Error with the\n * given message if it did.\n *\n * The code of callsites invoking this function are stripped out in production\n * builds. Any side-effects of code within the debugAssert() invocation will not\n * happen in this case.\n */\nexport function debugAssert(\n assertion: boolean,\n message: string\n): asserts assertion {\n if (!assertion) {\n fail(message);\n }\n}\n\n/**\n * Casts `obj` to `T`. In non-production builds, verifies that `obj` is an\n * instance of `T` before casting.\n */\nexport function debugCast<T>(\n obj: object,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n constructor: { new (...args: any[]): T }\n): T {\n debugAssert(\n obj instanceof constructor,\n `Expected type '${constructor.name}', but was '${obj.constructor.name}'`\n );\n return obj as T;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert } from './assert';\nimport { PlatformSupport } from '../platform/platform';\n\nexport type EventHandler<E> = (value: E) => void;\nexport interface Indexable {\n [k: string]: unknown;\n}\n\nexport class AutoId {\n static newId(): string {\n // Alphanumeric characters\n const chars =\n 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789';\n // The largest byte value that is a multiple of `char.length`.\n const maxMultiple = Math.floor(256 / chars.length) * chars.length;\n debugAssert(\n 0 < maxMultiple && maxMultiple < 256,\n `Expect maxMultiple to be (0, 256), but got ${maxMultiple}`\n );\n\n let autoId = '';\n const targetLength = 20;\n while (autoId.length < targetLength) {\n const bytes = PlatformSupport.getPlatform().randomBytes(40);\n for (let i = 0; i < bytes.length; ++i) {\n // Only accept values that are [0, maxMultiple), this ensures they can\n // be evenly mapped to indices of `chars` via a modulo operation.\n if (autoId.length < targetLength && bytes[i] < maxMultiple) {\n autoId += chars.charAt(bytes[i] % chars.length);\n }\n }\n }\n debugAssert(autoId.length === targetLength, 'Invalid auto ID: ' + autoId);\n\n return autoId;\n }\n}\n\nexport function primitiveComparator<T>(left: T, right: T): number {\n if (left < right) {\n return -1;\n }\n if (left > right) {\n return 1;\n }\n return 0;\n}\n\nexport interface Equatable<T> {\n isEqual(other: T): boolean;\n}\n\n/** Helper to compare arrays using isEqual(). */\nexport function arrayEquals<T>(\n left: T[],\n right: T[],\n comparator: (l: T, r: T) => boolean\n): boolean {\n if (left.length !== right.length) {\n return false;\n }\n return left.every((value, index) => comparator(value, right[index]));\n}\n/**\n * Returns the immediate lexicographically-following string. This is useful to\n * construct an inclusive range for indexeddb iterators.\n */\nexport function immediateSuccessor(s: string): string {\n // Return the input string, with an additional NUL byte appended.\n return s + '\\0';\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { primitiveComparator } from '../util/misc';\n\nexport class DatabaseInfo {\n /**\n * Constructs a DatabaseInfo using the provided host, databaseId and\n * persistenceKey.\n *\n * @param databaseId The database to use.\n * @param persistenceKey A unique identifier for this Firestore's local\n * storage (used in conjunction with the databaseId).\n * @param host The Firestore backend host to connect to.\n * @param ssl Whether to use SSL when connecting.\n * @param forceLongPolling Whether to use the forceLongPolling option\n * when using WebChannel as the network transport.\n */\n constructor(\n readonly databaseId: DatabaseId,\n readonly persistenceKey: string,\n readonly host: string,\n readonly ssl: boolean,\n readonly forceLongPolling: boolean\n ) {}\n}\n\n/** The default database name for a project. */\nconst DEFAULT_DATABASE_NAME = '(default)';\n\n/** Represents the database ID a Firestore client is associated with. */\nexport class DatabaseId {\n readonly database: string;\n constructor(readonly projectId: string, database?: string) {\n this.database = database ? database : DEFAULT_DATABASE_NAME;\n }\n\n get isDefaultDatabase(): boolean {\n return this.database === DEFAULT_DATABASE_NAME;\n }\n\n isEqual(other: {}): boolean {\n return (\n other instanceof DatabaseId &&\n other.projectId === this.projectId &&\n other.database === this.database\n );\n }\n\n compareTo(other: DatabaseId): number {\n return (\n primitiveComparator(this.projectId, other.projectId) ||\n primitiveComparator(this.database, other.database)\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Equatable } from './misc';\nimport { forEach, isEmpty } from './obj';\n\ntype Entry<K, V> = [K, V];\n\n/**\n * A map implementation that uses objects as keys. Objects must implement the\n * Equatable interface and must be immutable. Entries in the map are stored\n * together with the key being produced from the mapKeyFn. This map\n * automatically handles collisions of keys.\n */\nexport class ObjectMap<KeyType extends Equatable<KeyType>, ValueType> {\n /**\n * The inner map for a key -> value pair. Due to the possibility of\n * collisions we keep a list of entries that we do a linear search through\n * to find an actual match. Note that collisions should be rare, so we still\n * expect near constant time lookups in practice.\n */\n private inner: {\n [canonicalId: string]: Array<Entry<KeyType, ValueType>>;\n } = {};\n\n constructor(private mapKeyFn: (key: KeyType) => string) {}\n\n /** Get a value for this key, or undefined if it does not exist. */\n get(key: KeyType): ValueType | undefined {\n const id = this.mapKeyFn(key);\n const matches = this.inner[id];\n if (matches === undefined) {\n return undefined;\n }\n for (const [otherKey, value] of matches) {\n if (otherKey.isEqual(key)) {\n return value;\n }\n }\n return undefined;\n }\n\n has(key: KeyType): boolean {\n return this.get(key) !== undefined;\n }\n\n /** Put this key and value in the map. */\n set(key: KeyType, value: ValueType): void {\n const id = this.mapKeyFn(key);\n const matches = this.inner[id];\n if (matches === undefined) {\n this.inner[id] = [[key, value]];\n return;\n }\n for (let i = 0; i < matches.length; i++) {\n if (matches[i][0].isEqual(key)) {\n matches[i] = [key, value];\n return;\n }\n }\n matches.push([key, value]);\n }\n\n /**\n * Remove this key from the map. Returns a boolean if anything was deleted.\n */\n delete(key: KeyType): boolean {\n const id = this.mapKeyFn(key);\n const matches = this.inner[id];\n if (matches === undefined) {\n return false;\n }\n for (let i = 0; i < matches.length; i++) {\n if (matches[i][0].isEqual(key)) {\n if (matches.length === 1) {\n delete this.inner[id];\n } else {\n matches.splice(i, 1);\n }\n return true;\n }\n }\n return false;\n }\n\n forEach(fn: (key: KeyType, val: ValueType) => void): void {\n forEach(this.inner, (_, entries) => {\n for (const [k, v] of entries) {\n fn(k, v);\n }\n });\n }\n\n isEmpty(): boolean {\n return isEmpty(this.inner);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Timestamp } from '../api/timestamp';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { BatchId } from '../core/types';\nimport { hardAssert, debugAssert } from '../util/assert';\nimport { arrayEquals } from '../util/misc';\nimport { ByteString } from '../util/byte_string';\nimport {\n documentKeySet,\n DocumentKeySet,\n DocumentVersionMap,\n documentVersionMap,\n MaybeDocumentMap\n} from './collections';\nimport { MaybeDocument } from './document';\nimport { DocumentKey } from './document_key';\nimport { Mutation, MutationResult } from './mutation';\n\nexport const BATCHID_UNKNOWN = -1;\n\n/**\n * A batch of mutations that will be sent as one unit to the backend.\n */\nexport class MutationBatch {\n /**\n * @param batchId The unique ID of this mutation batch.\n * @param localWriteTime The original write time of this mutation.\n * @param baseMutations Mutations that are used to populate the base\n * values when this mutation is applied locally. This can be used to locally\n * overwrite values that are persisted in the remote document cache. Base\n * mutations are never sent to the backend.\n * @param mutations The user-provided mutations in this mutation batch.\n * User-provided mutations are applied both locally and remotely on the\n * backend.\n */\n constructor(\n public batchId: BatchId,\n public localWriteTime: Timestamp,\n public baseMutations: Mutation[],\n public mutations: Mutation[]\n ) {\n debugAssert(mutations.length > 0, 'Cannot create an empty mutation batch');\n }\n\n /**\n * Applies all the mutations in this MutationBatch to the specified document\n * to create a new remote document\n *\n * @param docKey The key of the document to apply mutations to.\n * @param maybeDoc The document to apply mutations to.\n * @param batchResult The result of applying the MutationBatch to the\n * backend.\n */\n applyToRemoteDocument(\n docKey: DocumentKey,\n maybeDoc: MaybeDocument | null,\n batchResult: MutationBatchResult\n ): MaybeDocument | null {\n if (maybeDoc) {\n debugAssert(\n maybeDoc.key.isEqual(docKey),\n `applyToRemoteDocument: key ${docKey} should match maybeDoc key\n ${maybeDoc.key}`\n );\n }\n\n const mutationResults = batchResult.mutationResults;\n debugAssert(\n mutationResults.length === this.mutations.length,\n `Mismatch between mutations length\n (${this.mutations.length}) and mutation results length\n (${mutationResults.length}).`\n );\n\n for (let i = 0; i < this.mutations.length; i++) {\n const mutation = this.mutations[i];\n if (mutation.key.isEqual(docKey)) {\n const mutationResult = mutationResults[i];\n maybeDoc = mutation.applyToRemoteDocument(maybeDoc, mutationResult);\n }\n }\n return maybeDoc;\n }\n\n /**\n * Computes the local view of a document given all the mutations in this\n * batch.\n *\n * @param docKey The key of the document to apply mutations to.\n * @param maybeDoc The document to apply mutations to.\n */\n applyToLocalView(\n docKey: DocumentKey,\n maybeDoc: MaybeDocument | null\n ): MaybeDocument | null {\n if (maybeDoc) {\n debugAssert(\n maybeDoc.key.isEqual(docKey),\n `applyToLocalDocument: key ${docKey} should match maybeDoc key\n ${maybeDoc.key}`\n );\n }\n\n // First, apply the base state. This allows us to apply non-idempotent\n // transform against a consistent set of values.\n for (const mutation of this.baseMutations) {\n if (mutation.key.isEqual(docKey)) {\n maybeDoc = mutation.applyToLocalView(\n maybeDoc,\n maybeDoc,\n this.localWriteTime\n );\n }\n }\n\n const baseDoc = maybeDoc;\n\n // Second, apply all user-provided mutations.\n for (const mutation of this.mutations) {\n if (mutation.key.isEqual(docKey)) {\n maybeDoc = mutation.applyToLocalView(\n maybeDoc,\n baseDoc,\n this.localWriteTime\n );\n }\n }\n return maybeDoc;\n }\n\n /**\n * Computes the local view for all provided documents given the mutations in\n * this batch.\n */\n applyToLocalDocumentSet(maybeDocs: MaybeDocumentMap): MaybeDocumentMap {\n // TODO(mrschmidt): This implementation is O(n^2). If we apply the mutations\n // directly (as done in `applyToLocalView()`), we can reduce the complexity\n // to O(n).\n let mutatedDocuments = maybeDocs;\n this.mutations.forEach(m => {\n const mutatedDocument = this.applyToLocalView(\n m.key,\n maybeDocs.get(m.key)\n );\n if (mutatedDocument) {\n mutatedDocuments = mutatedDocuments.insert(m.key, mutatedDocument);\n }\n });\n return mutatedDocuments;\n }\n\n keys(): DocumentKeySet {\n return this.mutations.reduce(\n (keys, m) => keys.add(m.key),\n documentKeySet()\n );\n }\n\n isEqual(other: MutationBatch): boolean {\n return (\n this.batchId === other.batchId &&\n arrayEquals(this.mutations, other.mutations, (l, r) => l.isEqual(r)) &&\n arrayEquals(this.baseMutations, other.baseMutations, (l, r) =>\n l.isEqual(r)\n )\n );\n }\n}\n\n/** The result of applying a mutation batch to the backend. */\nexport class MutationBatchResult {\n private constructor(\n readonly batch: MutationBatch,\n readonly commitVersion: SnapshotVersion,\n readonly mutationResults: MutationResult[],\n readonly streamToken: ByteString,\n /**\n * A pre-computed mapping from each mutated document to the resulting\n * version.\n */\n readonly docVersions: DocumentVersionMap\n ) {}\n\n /**\n * Creates a new MutationBatchResult for the given batch and results. There\n * must be one result for each mutation in the batch. This static factory\n * caches a document=>version mapping (docVersions).\n */\n static from(\n batch: MutationBatch,\n commitVersion: SnapshotVersion,\n results: MutationResult[],\n streamToken: ByteString\n ): MutationBatchResult {\n hardAssert(\n batch.mutations.length === results.length,\n 'Mutations sent ' +\n batch.mutations.length +\n ' must equal results received ' +\n results.length\n );\n\n let versionMap = documentVersionMap();\n const mutations = batch.mutations;\n for (let i = 0; i < mutations.length; i++) {\n versionMap = versionMap.insert(mutations[i].key, results[i].version);\n }\n\n return new MutationBatchResult(\n batch,\n commitVersion,\n results,\n streamToken,\n versionMap\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { fail } from '../util/assert';\n\nexport type FulfilledHandler<T, R> =\n | ((result: T) => R | PersistencePromise<R>)\n | null;\nexport type RejectedHandler<R> =\n | ((reason: Error) => R | PersistencePromise<R>)\n | null;\nexport type Resolver<T> = (value?: T) => void;\nexport type Rejector = (error: Error) => void;\n\n/**\n * PersistencePromise<> is essentially a re-implementation of Promise<> except\n * it has a .next() method instead of .then() and .next() and .catch() callbacks\n * are executed synchronously when a PersistencePromise resolves rather than\n * asynchronously (Promise<> implementations use setImmediate() or similar).\n *\n * This is necessary to interoperate with IndexedDB which will automatically\n * commit transactions if control is returned to the event loop without\n * synchronously initiating another operation on the transaction.\n *\n * NOTE: .then() and .catch() only allow a single consumer, unlike normal\n * Promises.\n */\nexport class PersistencePromise<T> {\n // NOTE: next/catchCallback will always point to our own wrapper functions,\n // not the user's raw next() or catch() callbacks.\n private nextCallback: FulfilledHandler<T, unknown> = null;\n private catchCallback: RejectedHandler<unknown> = null;\n\n // When the operation resolves, we'll set result or error and mark isDone.\n private result: T | undefined = undefined;\n private error: Error | undefined = undefined;\n private isDone = false;\n\n // Set to true when .then() or .catch() are called and prevents additional\n // chaining.\n private callbackAttached = false;\n\n constructor(callback: (resolve: Resolver<T>, reject: Rejector) => void) {\n callback(\n value => {\n this.isDone = true;\n this.result = value;\n if (this.nextCallback) {\n // value should be defined unless T is Void, but we can't express\n // that in the type system.\n this.nextCallback(value!);\n }\n },\n error => {\n this.isDone = true;\n this.error = error;\n if (this.catchCallback) {\n this.catchCallback(error);\n }\n }\n );\n }\n\n catch<R>(\n fn: (error: Error) => R | PersistencePromise<R>\n ): PersistencePromise<R> {\n return this.next(undefined, fn);\n }\n\n next<R>(\n nextFn?: FulfilledHandler<T, R>,\n catchFn?: RejectedHandler<R>\n ): PersistencePromise<R> {\n if (this.callbackAttached) {\n fail('Called next() or catch() twice for PersistencePromise');\n }\n this.callbackAttached = true;\n if (this.isDone) {\n if (!this.error) {\n return this.wrapSuccess(nextFn, this.result!);\n } else {\n return this.wrapFailure(catchFn, this.error);\n }\n } else {\n return new PersistencePromise<R>((resolve, reject) => {\n this.nextCallback = (value: T) => {\n this.wrapSuccess(nextFn, value).next(resolve, reject);\n };\n this.catchCallback = (error: Error) => {\n this.wrapFailure(catchFn, error).next(resolve, reject);\n };\n });\n }\n }\n\n toPromise(): Promise<T> {\n return new Promise((resolve, reject) => {\n this.next(resolve, reject);\n });\n }\n\n private wrapUserFunction<R>(\n fn: () => R | PersistencePromise<R>\n ): PersistencePromise<R> {\n try {\n const result = fn();\n if (result instanceof PersistencePromise) {\n return result;\n } else {\n return PersistencePromise.resolve(result);\n }\n } catch (e) {\n return PersistencePromise.reject<R>(e);\n }\n }\n\n private wrapSuccess<R>(\n nextFn: FulfilledHandler<T, R> | undefined,\n value: T\n ): PersistencePromise<R> {\n if (nextFn) {\n return this.wrapUserFunction(() => nextFn(value));\n } else {\n // If there's no nextFn, then R must be the same as T\n return PersistencePromise.resolve<R>((value as unknown) as R);\n }\n }\n\n private wrapFailure<R>(\n catchFn: RejectedHandler<R> | undefined,\n error: Error\n ): PersistencePromise<R> {\n if (catchFn) {\n return this.wrapUserFunction(() => catchFn(error));\n } else {\n return PersistencePromise.reject<R>(error);\n }\n }\n\n static resolve(): PersistencePromise<void>;\n static resolve<R>(result: R): PersistencePromise<R>;\n static resolve<R>(result?: R): PersistencePromise<R | void> {\n return new PersistencePromise<R | void>((resolve, reject) => {\n resolve(result);\n });\n }\n\n static reject<R>(error: Error): PersistencePromise<R> {\n return new PersistencePromise<R>((resolve, reject) => {\n reject(error);\n });\n }\n\n static waitFor(\n // Accept all Promise types in waitFor().\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n all: { forEach: (cb: (el: PersistencePromise<any>) => void) => void }\n ): PersistencePromise<void> {\n return new PersistencePromise<void>((resolve, reject) => {\n let expectedCount = 0;\n let resolvedCount = 0;\n let done = false;\n\n all.forEach(element => {\n ++expectedCount;\n element.next(\n () => {\n ++resolvedCount;\n if (done && resolvedCount === expectedCount) {\n resolve();\n }\n },\n err => reject(err)\n );\n });\n\n done = true;\n if (resolvedCount === expectedCount) {\n resolve();\n }\n });\n }\n\n /**\n * Given an array of predicate functions that asynchronously evaluate to a\n * boolean, implements a short-circuiting `or` between the results. Predicates\n * will be evaluated until one of them returns `true`, then stop. The final\n * result will be whether any of them returned `true`.\n */\n static or(\n predicates: Array<() => PersistencePromise<boolean>>\n ): PersistencePromise<boolean> {\n let p: PersistencePromise<boolean> = PersistencePromise.resolve<boolean>(\n false\n );\n for (const predicate of predicates) {\n p = p.next(isTrue => {\n if (isTrue) {\n return PersistencePromise.resolve<boolean>(isTrue);\n } else {\n return predicate();\n }\n });\n }\n return p;\n }\n\n /**\n * Given an iterable, call the given function on each element in the\n * collection and wait for all of the resulting concurrent PersistencePromises\n * to resolve.\n */\n static forEach<R, S>(\n collection: { forEach: (cb: (r: R, s: S) => void) => void },\n f:\n | ((r: R, s: S) => PersistencePromise<void>)\n | ((r: R) => PersistencePromise<void>)\n ): PersistencePromise<void>;\n static forEach<R>(\n collection: { forEach: (cb: (r: R) => void) => void },\n f: (r: R) => PersistencePromise<void>\n ): PersistencePromise<void>;\n static forEach<R, S>(\n collection: { forEach: (cb: (r: R, s?: S) => void) => void },\n f: (r: R, s?: S) => PersistencePromise<void>\n ): PersistencePromise<void> {\n const promises: Array<PersistencePromise<void>> = [];\n collection.forEach((r, s) => {\n promises.push(f.call(this, r, s));\n });\n return this.waitFor(promises);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Query } from '../core/query';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport {\n DocumentKeySet,\n documentKeySet,\n DocumentMap,\n documentMap,\n MaybeDocumentMap,\n maybeDocumentMap,\n NullableMaybeDocumentMap,\n nullableMaybeDocumentMap\n} from '../model/collections';\nimport { Document, MaybeDocument, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { MutationBatch } from '../model/mutation_batch';\nimport { ResourcePath } from '../model/path';\n\nimport { debugAssert } from '../util/assert';\nimport { IndexManager } from './index_manager';\nimport { MutationQueue } from './mutation_queue';\nimport { PatchMutation } from '../model/mutation';\nimport { PersistenceTransaction } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { RemoteDocumentCache } from './remote_document_cache';\n\n/**\n * A readonly view of the local state of all documents we're tracking (i.e. we\n * have a cached version in remoteDocumentCache or local mutations for the\n * document). The view is computed by applying the mutations in the\n * MutationQueue to the RemoteDocumentCache.\n */\nexport class LocalDocumentsView {\n constructor(\n readonly remoteDocumentCache: RemoteDocumentCache,\n readonly mutationQueue: MutationQueue,\n readonly indexManager: IndexManager\n ) {}\n\n /**\n * Get the local view of the document identified by `key`.\n *\n * @return Local view of the document or null if we don't have any cached\n * state for it.\n */\n getDocument(\n transaction: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<MaybeDocument | null> {\n return this.mutationQueue\n .getAllMutationBatchesAffectingDocumentKey(transaction, key)\n .next(batches => this.getDocumentInternal(transaction, key, batches));\n }\n\n /** Internal version of `getDocument` that allows reusing batches. */\n private getDocumentInternal(\n transaction: PersistenceTransaction,\n key: DocumentKey,\n inBatches: MutationBatch[]\n ): PersistencePromise<MaybeDocument | null> {\n return this.remoteDocumentCache.getEntry(transaction, key).next(doc => {\n for (const batch of inBatches) {\n doc = batch.applyToLocalView(key, doc);\n }\n return doc;\n });\n }\n\n // Returns the view of the given `docs` as they would appear after applying\n // all mutations in the given `batches`.\n private applyLocalMutationsToDocuments(\n transaction: PersistenceTransaction,\n docs: NullableMaybeDocumentMap,\n batches: MutationBatch[]\n ): NullableMaybeDocumentMap {\n let results = nullableMaybeDocumentMap();\n docs.forEach((key, localView) => {\n for (const batch of batches) {\n localView = batch.applyToLocalView(key, localView);\n }\n results = results.insert(key, localView);\n });\n return results;\n }\n\n /**\n * Gets the local view of the documents identified by `keys`.\n *\n * If we don't have cached state for a document in `keys`, a NoDocument will\n * be stored for that key in the resulting set.\n */\n getDocuments(\n transaction: PersistenceTransaction,\n keys: DocumentKeySet\n ): PersistencePromise<MaybeDocumentMap> {\n return this.remoteDocumentCache\n .getEntries(transaction, keys)\n .next(docs => this.getLocalViewOfDocuments(transaction, docs));\n }\n\n /**\n * Similar to `getDocuments`, but creates the local view from the given\n * `baseDocs` without retrieving documents from the local store.\n */\n getLocalViewOfDocuments(\n transaction: PersistenceTransaction,\n baseDocs: NullableMaybeDocumentMap\n ): PersistencePromise<MaybeDocumentMap> {\n return this.mutationQueue\n .getAllMutationBatchesAffectingDocumentKeys(transaction, baseDocs)\n .next(batches => {\n const docs = this.applyLocalMutationsToDocuments(\n transaction,\n baseDocs,\n batches\n );\n let results = maybeDocumentMap();\n docs.forEach((key, maybeDoc) => {\n // TODO(http://b/32275378): Don't conflate missing / deleted.\n if (!maybeDoc) {\n maybeDoc = new NoDocument(key, SnapshotVersion.min());\n }\n results = results.insert(key, maybeDoc);\n });\n\n return results;\n });\n }\n\n /**\n * Performs a query against the local view of all documents.\n *\n * @param transaction The persistence transaction.\n * @param query The query to match documents against.\n * @param sinceReadTime If not set to SnapshotVersion.min(), return only\n * documents that have been read since this snapshot version (exclusive).\n */\n getDocumentsMatchingQuery(\n transaction: PersistenceTransaction,\n query: Query,\n sinceReadTime: SnapshotVersion\n ): PersistencePromise<DocumentMap> {\n if (query.isDocumentQuery()) {\n return this.getDocumentsMatchingDocumentQuery(transaction, query.path);\n } else if (query.isCollectionGroupQuery()) {\n return this.getDocumentsMatchingCollectionGroupQuery(\n transaction,\n query,\n sinceReadTime\n );\n } else {\n return this.getDocumentsMatchingCollectionQuery(\n transaction,\n query,\n sinceReadTime\n );\n }\n }\n\n private getDocumentsMatchingDocumentQuery(\n transaction: PersistenceTransaction,\n docPath: ResourcePath\n ): PersistencePromise<DocumentMap> {\n // Just do a simple document lookup.\n return this.getDocument(transaction, new DocumentKey(docPath)).next(\n maybeDoc => {\n let result = documentMap();\n if (maybeDoc instanceof Document) {\n result = result.insert(maybeDoc.key, maybeDoc);\n }\n return result;\n }\n );\n }\n\n private getDocumentsMatchingCollectionGroupQuery(\n transaction: PersistenceTransaction,\n query: Query,\n sinceReadTime: SnapshotVersion\n ): PersistencePromise<DocumentMap> {\n debugAssert(\n query.path.isEmpty(),\n 'Currently we only support collection group queries at the root.'\n );\n const collectionId = query.collectionGroup!;\n let results = documentMap();\n return this.indexManager\n .getCollectionParents(transaction, collectionId)\n .next(parents => {\n // Perform a collection query against each parent that contains the\n // collectionId and aggregate the results.\n return PersistencePromise.forEach(parents, (parent: ResourcePath) => {\n const collectionQuery = query.asCollectionQueryAtPath(\n parent.child(collectionId)\n );\n return this.getDocumentsMatchingCollectionQuery(\n transaction,\n collectionQuery,\n sinceReadTime\n ).next(r => {\n r.forEach((key, doc) => {\n results = results.insert(key, doc);\n });\n });\n }).next(() => results);\n });\n }\n\n private getDocumentsMatchingCollectionQuery(\n transaction: PersistenceTransaction,\n query: Query,\n sinceReadTime: SnapshotVersion\n ): PersistencePromise<DocumentMap> {\n // Query the remote documents and overlay mutations.\n let results: DocumentMap;\n let mutationBatches: MutationBatch[];\n return this.remoteDocumentCache\n .getDocumentsMatchingQuery(transaction, query, sinceReadTime)\n .next(queryResults => {\n results = queryResults;\n return this.mutationQueue.getAllMutationBatchesAffectingQuery(\n transaction,\n query\n );\n })\n .next(matchingMutationBatches => {\n mutationBatches = matchingMutationBatches;\n // It is possible that a PatchMutation can make a document match a query, even if\n // the version in the RemoteDocumentCache is not a match yet (waiting for server\n // to ack). To handle this, we find all document keys affected by the PatchMutations\n // that are not in `result` yet, and back fill them via `remoteDocumentCache.getEntries`,\n // otherwise those `PatchMutations` will be ignored because no base document can be found,\n // and lead to missing result for the query.\n return this.addMissingBaseDocuments(\n transaction,\n mutationBatches,\n results\n ).next(mergedDocuments => {\n results = mergedDocuments;\n\n for (const batch of mutationBatches) {\n for (const mutation of batch.mutations) {\n const key = mutation.key;\n const baseDoc = results.get(key);\n const mutatedDoc = mutation.applyToLocalView(\n baseDoc,\n baseDoc,\n batch.localWriteTime\n );\n if (mutatedDoc instanceof Document) {\n results = results.insert(key, mutatedDoc);\n } else {\n results = results.remove(key);\n }\n }\n }\n });\n })\n .next(() => {\n // Finally, filter out any documents that don't actually match\n // the query.\n results.forEach((key, doc) => {\n if (!query.matches(doc)) {\n results = results.remove(key);\n }\n });\n\n return results;\n });\n }\n\n private addMissingBaseDocuments(\n transaction: PersistenceTransaction,\n matchingMutationBatches: MutationBatch[],\n existingDocuments: DocumentMap\n ): PersistencePromise<DocumentMap> {\n let missingBaseDocEntriesForPatching = documentKeySet();\n for (const batch of matchingMutationBatches) {\n for (const mutation of batch.mutations) {\n if (\n mutation instanceof PatchMutation &&\n existingDocuments.get(mutation.key) === null\n ) {\n missingBaseDocEntriesForPatching = missingBaseDocEntriesForPatching.add(\n mutation.key\n );\n }\n }\n }\n\n let mergedDocuments = existingDocuments;\n return this.remoteDocumentCache\n .getEntries(transaction, missingBaseDocEntriesForPatching)\n .next(missingBaseDocs => {\n missingBaseDocs.forEach((key, doc) => {\n if (doc !== null && doc instanceof Document) {\n mergedDocuments = mergedDocuments.insert(key, doc);\n }\n });\n return mergedDocuments;\n });\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { TargetId } from '../core/types';\nimport { ChangeType, ViewSnapshot } from '../core/view_snapshot';\nimport { documentKeySet, DocumentKeySet } from '../model/collections';\n\n/**\n * A set of changes to what documents are currently in view and out of view for\n * a given query. These changes are sent to the LocalStore by the View (via\n * the SyncEngine) and are used to pin / unpin documents as appropriate.\n */\nexport class LocalViewChanges {\n constructor(\n readonly targetId: TargetId,\n readonly fromCache: boolean,\n readonly addedKeys: DocumentKeySet,\n readonly removedKeys: DocumentKeySet\n ) {}\n\n static fromSnapshot(\n targetId: TargetId,\n viewSnapshot: ViewSnapshot\n ): LocalViewChanges {\n let addedKeys = documentKeySet();\n let removedKeys = documentKeySet();\n\n for (const docChange of viewSnapshot.docChanges) {\n switch (docChange.type) {\n case ChangeType.Added:\n addedKeys = addedKeys.add(docChange.doc.key);\n break;\n case ChangeType.Removed:\n removedKeys = removedKeys.add(docChange.doc.key);\n break;\n default:\n // do nothing\n }\n }\n\n return new LocalViewChanges(\n targetId,\n viewSnapshot.fromCache,\n addedKeys,\n removedKeys\n );\n }\n}\n","/**\n * @license\n * Copyright 2018 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { ListenSequenceNumber } from './types';\n\n/**\n * `SequenceNumberSyncer` defines the methods required to keep multiple instances of a\n * `ListenSequence` in sync.\n */\nexport interface SequenceNumberSyncer {\n // Notify the syncer that a new sequence number has been used.\n writeSequenceNumber(sequenceNumber: ListenSequenceNumber): void;\n // Setting this property allows the syncer to notify when a sequence number has been used, and\n // and lets the ListenSequence adjust its internal previous value accordingly.\n sequenceNumberHandler:\n | ((sequenceNumber: ListenSequenceNumber) => void)\n | null;\n}\n\n/**\n * `ListenSequence` is a monotonic sequence. It is initialized with a minimum value to\n * exceed. All subsequent calls to next will return increasing values. If provided with a\n * `SequenceNumberSyncer`, it will additionally bump its next value when told of a new value, as\n * well as write out sequence numbers that it produces via `next()`.\n */\nexport class ListenSequence {\n static readonly INVALID: ListenSequenceNumber = -1;\n\n private writeNewSequenceNumber?: (\n newSequenceNumber: ListenSequenceNumber\n ) => void;\n\n constructor(\n private previousValue: ListenSequenceNumber,\n sequenceNumberSyncer?: SequenceNumberSyncer\n ) {\n if (sequenceNumberSyncer) {\n sequenceNumberSyncer.sequenceNumberHandler = sequenceNumber =>\n this.setPreviousValue(sequenceNumber);\n this.writeNewSequenceNumber = sequenceNumber =>\n sequenceNumberSyncer.writeSequenceNumber(sequenceNumber);\n }\n }\n\n private setPreviousValue(\n externalPreviousValue: ListenSequenceNumber\n ): ListenSequenceNumber {\n this.previousValue = Math.max(externalPreviousValue, this.previousValue);\n return this.previousValue;\n }\n\n next(): ListenSequenceNumber {\n const nextValue = ++this.previousValue;\n if (this.writeNewSequenceNumber) {\n this.writeNewSequenceNumber(nextValue);\n }\n return nextValue;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nexport interface Resolver<R> {\n (value?: R | Promise<R>): void;\n}\n\nexport interface Rejecter {\n (reason?: Error): void;\n}\n\nexport class Deferred<R> {\n promise: Promise<R>;\n // Assigned synchronously in constructor by Promise constructor callback.\n resolve!: Resolver<R>;\n reject!: Rejecter;\n\n constructor() {\n this.promise = new Promise((resolve: Resolver<R>, reject: Rejecter) => {\n this.resolve = resolve;\n this.reject = reject;\n });\n }\n}\n\n/**\n * Takes an array of values and a function from a value to a Promise. The function is run on each\n * value sequentially, waiting for the previous promise to resolve before starting the next one.\n * The returned promise resolves once the function has been run on all values.\n */\nexport function sequence<T>(\n values: T[],\n fn: (value: T) => Promise<void>\n): Promise<void> {\n let p = Promise.resolve();\n for (const value of values) {\n p = p.then(() => fn(value));\n }\n return p;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { AsyncQueue, DelayedOperation, TimerId } from '../util/async_queue';\nimport { logDebug } from '../util/log';\n\nconst LOG_TAG = 'ExponentialBackoff';\n\n/**\n * Initial backoff time in milliseconds after an error.\n * Set to 1s according to https://cloud.google.com/apis/design/errors.\n */\nconst DEFAULT_BACKOFF_INITIAL_DELAY_MS = 1000;\n\nconst DEFAULT_BACKOFF_FACTOR = 1.5;\n\n/** Maximum backoff time in milliseconds */\nconst DEFAULT_BACKOFF_MAX_DELAY_MS = 60 * 1000;\n\n/**\n * A helper for running delayed tasks following an exponential backoff curve\n * between attempts.\n *\n * Each delay is made up of a \"base\" delay which follows the exponential\n * backoff curve, and a +/- 50% \"jitter\" that is calculated and added to the\n * base delay. This prevents clients from accidentally synchronizing their\n * delays causing spikes of load to the backend.\n */\nexport class ExponentialBackoff {\n private currentBaseMs: number = 0;\n private timerPromise: DelayedOperation<void> | null = null;\n /** The last backoff attempt, as epoch milliseconds. */\n private lastAttemptTime = Date.now();\n\n constructor(\n /**\n * The AsyncQueue to run backoff operations on.\n */\n private readonly queue: AsyncQueue,\n /**\n * The ID to use when scheduling backoff operations on the AsyncQueue.\n */\n private readonly timerId: TimerId,\n /**\n * The initial delay (used as the base delay on the first retry attempt).\n * Note that jitter will still be applied, so the actual delay could be as\n * little as 0.5*initialDelayMs.\n */\n private readonly initialDelayMs: number = DEFAULT_BACKOFF_INITIAL_DELAY_MS,\n /**\n * The multiplier to use to determine the extended base delay after each\n * attempt.\n */\n private readonly backoffFactor: number = DEFAULT_BACKOFF_FACTOR,\n /**\n * The maximum base delay after which no further backoff is performed.\n * Note that jitter will still be applied, so the actual delay could be as\n * much as 1.5*maxDelayMs.\n */\n private readonly maxDelayMs: number = DEFAULT_BACKOFF_MAX_DELAY_MS\n ) {\n this.reset();\n }\n\n /**\n * Resets the backoff delay.\n *\n * The very next backoffAndWait() will have no delay. If it is called again\n * (i.e. due to an error), initialDelayMs (plus jitter) will be used, and\n * subsequent ones will increase according to the backoffFactor.\n */\n reset(): void {\n this.currentBaseMs = 0;\n }\n\n /**\n * Resets the backoff delay to the maximum delay (e.g. for use after a\n * RESOURCE_EXHAUSTED error).\n */\n resetToMax(): void {\n this.currentBaseMs = this.maxDelayMs;\n }\n\n /**\n * Returns a promise that resolves after currentDelayMs, and increases the\n * delay for any subsequent attempts. If there was a pending backoff operation\n * already, it will be canceled.\n */\n backoffAndRun(op: () => Promise<void>): void {\n // Cancel any pending backoff operation.\n this.cancel();\n\n // First schedule using the current base (which may be 0 and should be\n // honored as such).\n const desiredDelayWithJitterMs = Math.floor(\n this.currentBaseMs + this.jitterDelayMs()\n );\n\n // Guard against lastAttemptTime being in the future due to a clock change.\n const delaySoFarMs = Math.max(0, Date.now() - this.lastAttemptTime);\n\n // Guard against the backoff delay already being past.\n const remainingDelayMs = Math.max(\n 0,\n desiredDelayWithJitterMs - delaySoFarMs\n );\n\n if (remainingDelayMs > 0) {\n logDebug(\n LOG_TAG,\n `Backing off for ${remainingDelayMs} ms ` +\n `(base delay: ${this.currentBaseMs} ms, ` +\n `delay with jitter: ${desiredDelayWithJitterMs} ms, ` +\n `last attempt: ${delaySoFarMs} ms ago)`\n );\n }\n\n this.timerPromise = this.queue.enqueueAfterDelay(\n this.timerId,\n remainingDelayMs,\n () => {\n this.lastAttemptTime = Date.now();\n return op();\n }\n );\n\n // Apply backoff factor to determine next delay and ensure it is within\n // bounds.\n this.currentBaseMs *= this.backoffFactor;\n if (this.currentBaseMs < this.initialDelayMs) {\n this.currentBaseMs = this.initialDelayMs;\n }\n if (this.currentBaseMs > this.maxDelayMs) {\n this.currentBaseMs = this.maxDelayMs;\n }\n }\n\n skipBackoff(): void {\n if (this.timerPromise !== null) {\n this.timerPromise.skipDelay();\n this.timerPromise = null;\n }\n }\n\n cancel(): void {\n if (this.timerPromise !== null) {\n this.timerPromise.cancel();\n this.timerPromise = null;\n }\n }\n\n /** Returns a random value in the range [-currentBaseMs/2, currentBaseMs/2] */\n private jitterDelayMs(): number {\n return (Math.random() - 0.5) * this.currentBaseMs;\n }\n}\n","/**\n * @license\n * Copyright 2019 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { ResourcePath } from '../model/path';\nimport { debugAssert } from '../util/assert';\nimport { SortedSet } from '../util/sorted_set';\nimport { IndexManager } from './index_manager';\nimport { PersistenceTransaction } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\n\n/**\n * An in-memory implementation of IndexManager.\n */\nexport class MemoryIndexManager implements IndexManager {\n private collectionParentIndex = new MemoryCollectionParentIndex();\n\n addToCollectionParentIndex(\n transaction: PersistenceTransaction,\n collectionPath: ResourcePath\n ): PersistencePromise<void> {\n this.collectionParentIndex.add(collectionPath);\n return PersistencePromise.resolve();\n }\n\n getCollectionParents(\n transaction: PersistenceTransaction,\n collectionId: string\n ): PersistencePromise<ResourcePath[]> {\n return PersistencePromise.resolve(\n this.collectionParentIndex.getEntries(collectionId)\n );\n }\n}\n\n/**\n * Internal implementation of the collection-parent index exposed by MemoryIndexManager.\n * Also used for in-memory caching by IndexedDbIndexManager and initial index population\n * in indexeddb_schema.ts\n */\nexport class MemoryCollectionParentIndex {\n private index = {} as {\n [collectionId: string]: SortedSet<ResourcePath>;\n };\n\n // Returns false if the entry already existed.\n add(collectionPath: ResourcePath): boolean {\n debugAssert(collectionPath.length % 2 === 1, 'Expected a collection path.');\n const collectionId = collectionPath.lastSegment();\n const parentPath = collectionPath.popLast();\n const existingParents =\n this.index[collectionId] ||\n new SortedSet<ResourcePath>(ResourcePath.comparator);\n const added = !existingParents.has(parentPath);\n this.index[collectionId] = existingParents.add(parentPath);\n return added;\n }\n\n has(collectionPath: ResourcePath): boolean {\n const collectionId = collectionPath.lastSegment();\n const parentPath = collectionPath.popLast();\n const existingParents = this.index[collectionId];\n return existingParents && existingParents.has(parentPath);\n }\n\n getEntries(collectionId: string): ResourcePath[] {\n const parentPaths =\n this.index[collectionId] ||\n new SortedSet<ResourcePath>(ResourcePath.comparator);\n return parentPaths.toArray();\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { TargetId } from './types';\n\n/** Offset to ensure non-overlapping target ids. */\nconst OFFSET = 2;\n\n/**\n * Generates monotonically increasing target IDs for sending targets to the\n * watch stream.\n *\n * The client constructs two generators, one for the target cache, and one for\n * for the sync engine (to generate limbo documents targets). These\n * generators produce non-overlapping IDs (by using even and odd IDs\n * respectively).\n *\n * By separating the target ID space, the query cache can generate target IDs\n * that persist across client restarts, while sync engine can independently\n * generate in-memory target IDs that are transient and can be reused after a\n * restart.\n */\nexport class TargetIdGenerator {\n constructor(private lastId: number) {}\n\n next(): TargetId {\n this.lastId += OFFSET;\n return this.lastId;\n }\n\n static forTargetCache(): TargetIdGenerator {\n // The target cache generator must return '2' in its first call to `next()`\n // as there is no differentiation in the protocol layer between an unset\n // number and the number '0'. If we were to sent a target with target ID\n // '0', the backend would consider it unset and replace it with its own ID.\n return new TargetIdGenerator(2 - OFFSET);\n }\n\n static forSyncEngine(): TargetIdGenerator {\n // Sync engine assigns target IDs for limbo document detection.\n return new TargetIdGenerator(1 - OFFSET);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { getUA } from '@firebase/util';\nimport { debugAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { logDebug, logError } from '../util/log';\nimport { Deferred } from '../util/promise';\nimport { SCHEMA_VERSION } from './indexeddb_schema';\nimport { PersistencePromise } from './persistence_promise';\n\n// References to `window` are guarded by SimpleDb.isAvailable()\n/* eslint-disable no-restricted-globals */\n\nconst LOG_TAG = 'SimpleDb';\n\n/**\n * The maximum number of retry attempts for an IndexedDb transaction that fails\n * with a DOMException.\n */\nconst TRANSACTION_RETRY_COUNT = 3;\n\n// The different modes supported by `SimpleDb.runTransaction()`\ntype SimpleDbTransactionMode = 'readonly' | 'readwrite';\n\nexport interface SimpleDbSchemaConverter {\n createOrUpgrade(\n db: IDBDatabase,\n txn: IDBTransaction,\n fromVersion: number,\n toVersion: number\n ): PersistencePromise<void>;\n}\n\n/**\n * Provides a wrapper around IndexedDb with a simplified interface that uses\n * Promise-like return values to chain operations. Real promises cannot be used\n * since .then() continuations are executed asynchronously (e.g. via\n * .setImmediate), which would cause IndexedDB to end the transaction.\n * See PersistencePromise for more details.\n */\nexport class SimpleDb {\n /**\n * Opens the specified database, creating or upgrading it if necessary.\n *\n * Note that `version` must not be a downgrade. IndexedDB does not support downgrading the schema\n * version. We currently do not support any way to do versioning outside of IndexedDB's versioning\n * mechanism, as only version-upgrade transactions are allowed to do things like create\n * objectstores.\n */\n static openOrCreate(\n name: string,\n version: number,\n schemaConverter: SimpleDbSchemaConverter\n ): Promise<SimpleDb> {\n debugAssert(\n SimpleDb.isAvailable(),\n 'IndexedDB not supported in current environment.'\n );\n logDebug(LOG_TAG, 'Opening database:', name);\n return new PersistencePromise<SimpleDb>((resolve, reject) => {\n // TODO(mikelehen): Investigate browser compatibility.\n // https://developer.mozilla.org/en-US/docs/Web/API/IndexedDB_API/Using_IndexedDB\n // suggests IE9 and older WebKit browsers handle upgrade\n // differently. They expect setVersion, as described here:\n // https://developer.mozilla.org/en-US/docs/Web/API/IDBVersionChangeRequest/setVersion\n const request = window.indexedDB.open(name, version);\n\n request.onsuccess = (event: Event) => {\n const db = (event.target as IDBOpenDBRequest).result;\n resolve(new SimpleDb(db));\n };\n\n request.onblocked = () => {\n reject(\n new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'Cannot upgrade IndexedDB schema while another tab is open. ' +\n 'Close all tabs that access Firestore and reload this page to proceed.'\n )\n );\n };\n\n request.onerror = (event: Event) => {\n const error: DOMException = (event.target as IDBOpenDBRequest).error!;\n if (error.name === 'VersionError') {\n reject(\n new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'A newer version of the Firestore SDK was previously used and so the persisted ' +\n 'data is not compatible with the version of the SDK you are now using. The SDK ' +\n 'will operate with persistence disabled. If you need persistence, please ' +\n 're-upgrade to a newer version of the SDK or else clear the persisted IndexedDB ' +\n 'data for your app to start fresh.'\n )\n );\n } else {\n reject(error);\n }\n };\n\n request.onupgradeneeded = (event: IDBVersionChangeEvent) => {\n logDebug(\n LOG_TAG,\n 'Database \"' + name + '\" requires upgrade from version:',\n event.oldVersion\n );\n const db = (event.target as IDBOpenDBRequest).result;\n schemaConverter\n .createOrUpgrade(\n db,\n request.transaction!,\n event.oldVersion,\n SCHEMA_VERSION\n )\n .next(() => {\n logDebug(\n LOG_TAG,\n 'Database upgrade to version ' + SCHEMA_VERSION + ' complete'\n );\n });\n };\n }).toPromise();\n }\n\n /** Deletes the specified database. */\n static delete(name: string): Promise<void> {\n logDebug(LOG_TAG, 'Removing database:', name);\n return wrapRequest<void>(window.indexedDB.deleteDatabase(name)).toPromise();\n }\n\n /** Returns true if IndexedDB is available in the current environment. */\n static isAvailable(): boolean {\n if (typeof window === 'undefined' || window.indexedDB == null) {\n return false;\n }\n\n if (SimpleDb.isMockPersistence()) {\n return true;\n }\n\n // In some Node environments, `window` is defined, but `window.navigator` is\n // not. We don't support IndexedDB persistence in Node if the\n // isMockPersistence() check above returns false.\n if (window.navigator === undefined) {\n return false;\n }\n\n // We extensively use indexed array values and compound keys,\n // which IE and Edge do not support. However, they still have indexedDB\n // defined on the window, so we need to check for them here and make sure\n // to return that persistence is not enabled for those browsers.\n // For tracking support of this feature, see here:\n // https://developer.microsoft.com/en-us/microsoft-edge/platform/status/indexeddbarraysandmultientrysupport/\n\n // Check the UA string to find out the browser.\n const ua = getUA();\n\n // IE 10\n // ua = 'Mozilla/5.0 (compatible; MSIE 10.0; Windows NT 6.2; Trident/6.0)';\n\n // IE 11\n // ua = 'Mozilla/5.0 (Windows NT 6.3; Trident/7.0; rv:11.0) like Gecko';\n\n // Edge\n // ua = 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML,\n // like Gecko) Chrome/39.0.2171.71 Safari/537.36 Edge/12.0';\n\n // iOS Safari: Disable for users running iOS version < 10.\n const iOSVersion = SimpleDb.getIOSVersion(ua);\n const isUnsupportedIOS = 0 < iOSVersion && iOSVersion < 10;\n\n // Android browser: Disable for userse running version < 4.5.\n const androidVersion = SimpleDb.getAndroidVersion(ua);\n const isUnsupportedAndroid = 0 < androidVersion && androidVersion < 4.5;\n\n if (\n ua.indexOf('MSIE ') > 0 ||\n ua.indexOf('Trident/') > 0 ||\n ua.indexOf('Edge/') > 0 ||\n isUnsupportedIOS ||\n isUnsupportedAndroid\n ) {\n return false;\n } else {\n return true;\n }\n }\n\n /**\n * Returns true if the backing IndexedDB store is the Node IndexedDBShim\n * (see https://github.com/axemclion/IndexedDBShim).\n */\n static isMockPersistence(): boolean {\n return (\n typeof process !== 'undefined' &&\n process.env?.USE_MOCK_PERSISTENCE === 'YES'\n );\n }\n\n /** Helper to get a typed SimpleDbStore from a transaction. */\n static getStore<KeyType extends IDBValidKey, ValueType extends unknown>(\n txn: SimpleDbTransaction,\n store: string\n ): SimpleDbStore<KeyType, ValueType> {\n return txn.store<KeyType, ValueType>(store);\n }\n\n // visible for testing\n /** Parse User Agent to determine iOS version. Returns -1 if not found. */\n static getIOSVersion(ua: string): number {\n const iOSVersionRegex = ua.match(/i(?:phone|pad|pod) os ([\\d_]+)/i);\n const version = iOSVersionRegex\n ? iOSVersionRegex[1]\n .split('_')\n .slice(0, 2)\n .join('.')\n : '-1';\n return Number(version);\n }\n\n // visible for testing\n /** Parse User Agent to determine Android version. Returns -1 if not found. */\n static getAndroidVersion(ua: string): number {\n const androidVersionRegex = ua.match(/Android ([\\d.]+)/i);\n const version = androidVersionRegex\n ? androidVersionRegex[1]\n .split('.')\n .slice(0, 2)\n .join('.')\n : '-1';\n return Number(version);\n }\n\n constructor(private db: IDBDatabase) {\n const iOSVersion = SimpleDb.getIOSVersion(getUA());\n // NOTE: According to https://bugs.webkit.org/show_bug.cgi?id=197050, the\n // bug we're checking for should exist in iOS >= 12.2 and < 13, but for\n // whatever reason it's much harder to hit after 12.2 so we only proactively\n // log on 12.2.\n if (iOSVersion === 12.2) {\n logError(\n 'Firestore persistence suffers from a bug in iOS 12.2 ' +\n 'Safari that may cause your app to stop working. See ' +\n 'https://stackoverflow.com/q/56496296/110915 for details ' +\n 'and a potential workaround.'\n );\n }\n }\n\n setVersionChangeListener(\n versionChangeListener: (event: IDBVersionChangeEvent) => void\n ): void {\n this.db.onversionchange = (event: IDBVersionChangeEvent) => {\n return versionChangeListener(event);\n };\n }\n\n async runTransaction<T>(\n mode: SimpleDbTransactionMode,\n objectStores: string[],\n transactionFn: (transaction: SimpleDbTransaction) => PersistencePromise<T>\n ): Promise<T> {\n const readonly = mode === 'readonly';\n let attemptNumber = 0;\n\n while (true) {\n ++attemptNumber;\n\n const transaction = SimpleDbTransaction.open(\n this.db,\n readonly ? 'readonly' : 'readwrite',\n objectStores\n );\n try {\n const transactionFnResult = transactionFn(transaction)\n .catch(error => {\n // Abort the transaction if there was an error.\n transaction.abort(error);\n // We cannot actually recover, and calling `abort()` will cause the transaction's\n // completion promise to be rejected. This in turn means that we won't use\n // `transactionFnResult` below. We return a rejection here so that we don't add the\n // possibility of returning `void` to the type of `transactionFnResult`.\n return PersistencePromise.reject<T>(error);\n })\n .toPromise();\n\n // As noted above, errors are propagated by aborting the transaction. So\n // we swallow any error here to avoid the browser logging it as unhandled.\n transactionFnResult.catch(() => {});\n\n // Wait for the transaction to complete (i.e. IndexedDb's onsuccess event to\n // fire), but still return the original transactionFnResult back to the\n // caller.\n await transaction.completionPromise;\n return transactionFnResult;\n } catch (error) {\n // TODO(schmidt-sebastian): We could probably be smarter about this and\n // not retry exceptions that are likely unrecoverable (such as quota\n // exceeded errors).\n\n // Note: We cannot use an instanceof check for FirestoreException, since the\n // exception is wrapped in a generic error by our async/await handling.\n const retryable =\n error.name !== 'FirebaseError' &&\n attemptNumber < TRANSACTION_RETRY_COUNT;\n logDebug(\n LOG_TAG,\n 'Transaction failed with error: %s. Retrying: %s.',\n error.message,\n retryable\n );\n\n if (!retryable) {\n return Promise.reject(error);\n }\n }\n }\n }\n\n close(): void {\n this.db.close();\n }\n}\n\n/**\n * A controller for iterating over a key range or index. It allows an iterate\n * callback to delete the currently-referenced object, or jump to a new key\n * within the key range or index.\n */\nexport class IterationController {\n private shouldStop = false;\n private nextKey: IDBValidKey | null = null;\n\n constructor(private dbCursor: IDBCursorWithValue) {}\n\n get isDone(): boolean {\n return this.shouldStop;\n }\n\n get skipToKey(): IDBValidKey | null {\n return this.nextKey;\n }\n\n set cursor(value: IDBCursorWithValue) {\n this.dbCursor = value;\n }\n\n /**\n * This function can be called to stop iteration at any point.\n */\n done(): void {\n this.shouldStop = true;\n }\n\n /**\n * This function can be called to skip to that next key, which could be\n * an index or a primary key.\n */\n skip(key: IDBValidKey): void {\n this.nextKey = key;\n }\n\n /**\n * Delete the current cursor value from the object store.\n *\n * NOTE: You CANNOT do this with a keysOnly query.\n */\n delete(): PersistencePromise<void> {\n return wrapRequest<void>(this.dbCursor.delete());\n }\n}\n\n/**\n * Callback used with iterate() method.\n */\nexport type IterateCallback<KeyType, ValueType> = (\n key: KeyType,\n value: ValueType,\n control: IterationController\n) => void | PersistencePromise<void>;\n\n/** Options available to the iterate() method. */\nexport interface IterateOptions {\n /** Index to iterate over (else primary keys will be iterated) */\n index?: string;\n\n /** IndxedDB Range to iterate over (else entire store will be iterated) */\n range?: IDBKeyRange;\n\n /** If true, values aren't read while iterating. */\n keysOnly?: boolean;\n\n /** If true, iterate over the store in reverse. */\n reverse?: boolean;\n}\n\n/** An error that wraps exceptions that thrown during IndexedDB execution. */\nexport class IndexedDbTransactionError extends FirestoreError {\n name = 'IndexedDbTransactionError';\n\n constructor(cause: Error) {\n super(Code.UNAVAILABLE, 'IndexedDB transaction failed: ' + cause);\n }\n}\n\n/** Verifies whether `e` is an IndexedDbTransactionError. */\nexport function isIndexedDbTransactionError(e: Error): boolean {\n // Use name equality, as instanceof checks on errors don't work with errors\n // that wrap other errors.\n return e.name === 'IndexedDbTransactionError';\n}\n\n/**\n * Wraps an IDBTransaction and exposes a store() method to get a handle to a\n * specific object store.\n */\nexport class SimpleDbTransaction {\n private aborted = false;\n\n /**\n * A promise that resolves with the result of the IndexedDb transaction.\n */\n private readonly completionDeferred = new Deferred<void>();\n\n static open(\n db: IDBDatabase,\n mode: IDBTransactionMode,\n objectStoreNames: string[]\n ): SimpleDbTransaction {\n return new SimpleDbTransaction(db.transaction(objectStoreNames, mode));\n }\n\n constructor(private readonly transaction: IDBTransaction) {\n this.transaction.oncomplete = () => {\n this.completionDeferred.resolve();\n };\n this.transaction.onabort = () => {\n if (transaction.error) {\n this.completionDeferred.reject(\n new IndexedDbTransactionError(transaction.error)\n );\n } else {\n this.completionDeferred.resolve();\n }\n };\n this.transaction.onerror = (event: Event) => {\n const error = checkForAndReportiOSError(\n (event.target as IDBRequest).error!\n );\n this.completionDeferred.reject(new IndexedDbTransactionError(error));\n };\n }\n\n get completionPromise(): Promise<void> {\n return this.completionDeferred.promise;\n }\n\n abort(error?: Error): void {\n if (error) {\n this.completionDeferred.reject(error);\n }\n\n if (!this.aborted) {\n logDebug(\n LOG_TAG,\n 'Aborting transaction:',\n error ? error.message : 'Client-initiated abort'\n );\n this.aborted = true;\n this.transaction.abort();\n }\n }\n\n /**\n * Returns a SimpleDbStore<KeyType, ValueType> for the specified store. All\n * operations performed on the SimpleDbStore happen within the context of this\n * transaction and it cannot be used anymore once the transaction is\n * completed.\n *\n * Note that we can't actually enforce that the KeyType and ValueType are\n * correct, but they allow type safety through the rest of the consuming code.\n */\n store<KeyType extends IDBValidKey, ValueType extends unknown>(\n storeName: string\n ): SimpleDbStore<KeyType, ValueType> {\n const store = this.transaction.objectStore(storeName);\n debugAssert(!!store, 'Object store not part of transaction: ' + storeName);\n return new SimpleDbStore<KeyType, ValueType>(store);\n }\n}\n\n/**\n * A wrapper around an IDBObjectStore providing an API that:\n *\n * 1) Has generic KeyType / ValueType parameters to provide strongly-typed\n * methods for acting against the object store.\n * 2) Deals with IndexedDB's onsuccess / onerror event callbacks, making every\n * method return a PersistencePromise instead.\n * 3) Provides a higher-level API to avoid needing to do excessive wrapping of\n * intermediate IndexedDB types (IDBCursorWithValue, etc.)\n */\nexport class SimpleDbStore<\n KeyType extends IDBValidKey,\n ValueType extends unknown\n> {\n constructor(private store: IDBObjectStore) {}\n\n /**\n * Writes a value into the Object Store.\n *\n * @param key Optional explicit key to use when writing the object, else the\n * key will be auto-assigned (e.g. via the defined keyPath for the store).\n * @param value The object to write.\n */\n put(value: ValueType): PersistencePromise<void>;\n put(key: KeyType, value: ValueType): PersistencePromise<void>;\n put(\n keyOrValue: KeyType | ValueType,\n value?: ValueType\n ): PersistencePromise<void> {\n let request;\n if (value !== undefined) {\n logDebug(LOG_TAG, 'PUT', this.store.name, keyOrValue, value);\n request = this.store.put(value, keyOrValue as KeyType);\n } else {\n logDebug(LOG_TAG, 'PUT', this.store.name, '<auto-key>', keyOrValue);\n request = this.store.put(keyOrValue as ValueType);\n }\n return wrapRequest<void>(request);\n }\n\n /**\n * Adds a new value into an Object Store and returns the new key. Similar to\n * IndexedDb's `add()`, this method will fail on primary key collisions.\n *\n * @param value The object to write.\n * @return The key of the value to add.\n */\n add(value: ValueType): PersistencePromise<KeyType> {\n logDebug(LOG_TAG, 'ADD', this.store.name, value, value);\n const request = this.store.add(value as ValueType);\n return wrapRequest<KeyType>(request);\n }\n\n /**\n * Gets the object with the specified key from the specified store, or null\n * if no object exists with the specified key.\n *\n * @key The key of the object to get.\n * @return The object with the specified key or null if no object exists.\n */\n get(key: KeyType): PersistencePromise<ValueType | null> {\n const request = this.store.get(key);\n // We're doing an unsafe cast to ValueType.\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n return wrapRequest<any>(request).next(result => {\n // Normalize nonexistence to null.\n if (result === undefined) {\n result = null;\n }\n logDebug(LOG_TAG, 'GET', this.store.name, key, result);\n return result;\n });\n }\n\n delete(key: KeyType | IDBKeyRange): PersistencePromise<void> {\n logDebug(LOG_TAG, 'DELETE', this.store.name, key);\n const request = this.store.delete(key);\n return wrapRequest<void>(request);\n }\n\n /**\n * If we ever need more of the count variants, we can add overloads. For now,\n * all we need is to count everything in a store.\n *\n * Returns the number of rows in the store.\n */\n count(): PersistencePromise<number> {\n logDebug(LOG_TAG, 'COUNT', this.store.name);\n const request = this.store.count();\n return wrapRequest<number>(request);\n }\n\n loadAll(): PersistencePromise<ValueType[]>;\n loadAll(range: IDBKeyRange): PersistencePromise<ValueType[]>;\n loadAll(index: string, range: IDBKeyRange): PersistencePromise<ValueType[]>;\n loadAll(\n indexOrRange?: string | IDBKeyRange,\n range?: IDBKeyRange\n ): PersistencePromise<ValueType[]> {\n const cursor = this.cursor(this.options(indexOrRange, range));\n const results: ValueType[] = [];\n return this.iterateCursor(cursor, (key, value) => {\n results.push(value);\n }).next(() => {\n return results;\n });\n }\n\n deleteAll(): PersistencePromise<void>;\n deleteAll(range: IDBKeyRange): PersistencePromise<void>;\n deleteAll(index: string, range: IDBKeyRange): PersistencePromise<void>;\n deleteAll(\n indexOrRange?: string | IDBKeyRange,\n range?: IDBKeyRange\n ): PersistencePromise<void> {\n logDebug(LOG_TAG, 'DELETE ALL', this.store.name);\n const options = this.options(indexOrRange, range);\n options.keysOnly = false;\n const cursor = this.cursor(options);\n return this.iterateCursor(cursor, (key, value, control) => {\n // NOTE: Calling delete() on a cursor is documented as more efficient than\n // calling delete() on an object store with a single key\n // (https://developer.mozilla.org/en-US/docs/Web/API/IDBObjectStore/delete),\n // however, this requires us *not* to use a keysOnly cursor\n // (https://developer.mozilla.org/en-US/docs/Web/API/IDBCursor/delete). We\n // may want to compare the performance of each method.\n return control.delete();\n });\n }\n\n /**\n * Iterates over keys and values in an object store.\n *\n * @param options Options specifying how to iterate the objects in the store.\n * @param callback will be called for each iterated object. Iteration can be\n * canceled at any point by calling the doneFn passed to the callback.\n * The callback can return a PersistencePromise if it performs async\n * operations but note that iteration will continue without waiting for them\n * to complete.\n * @returns A PersistencePromise that resolves once all PersistencePromises\n * returned by callbacks resolve.\n */\n iterate(\n callback: IterateCallback<KeyType, ValueType>\n ): PersistencePromise<void>;\n iterate(\n options: IterateOptions,\n callback: IterateCallback<KeyType, ValueType>\n ): PersistencePromise<void>;\n iterate(\n optionsOrCallback: IterateOptions | IterateCallback<KeyType, ValueType>,\n callback?: IterateCallback<KeyType, ValueType>\n ): PersistencePromise<void> {\n let options;\n if (!callback) {\n options = {};\n callback = optionsOrCallback as IterateCallback<KeyType, ValueType>;\n } else {\n options = optionsOrCallback as IterateOptions;\n }\n const cursor = this.cursor(options);\n return this.iterateCursor(cursor, callback);\n }\n\n /**\n * Iterates over a store, but waits for the given callback to complete for\n * each entry before iterating the next entry. This allows the callback to do\n * asynchronous work to determine if this iteration should continue.\n *\n * The provided callback should return `true` to continue iteration, and\n * `false` otherwise.\n */\n iterateSerial(\n callback: (k: KeyType, v: ValueType) => PersistencePromise<boolean>\n ): PersistencePromise<void> {\n const cursorRequest = this.cursor({});\n return new PersistencePromise((resolve, reject) => {\n cursorRequest.onerror = (event: Event) => {\n const error = checkForAndReportiOSError(\n (event.target as IDBRequest).error!\n );\n reject(error);\n };\n cursorRequest.onsuccess = (event: Event) => {\n const cursor: IDBCursorWithValue = (event.target as IDBRequest).result;\n if (!cursor) {\n resolve();\n return;\n }\n\n callback(cursor.primaryKey as KeyType, cursor.value).next(\n shouldContinue => {\n if (shouldContinue) {\n cursor.continue();\n } else {\n resolve();\n }\n }\n );\n };\n });\n }\n\n private iterateCursor(\n cursorRequest: IDBRequest,\n fn: IterateCallback<KeyType, ValueType>\n ): PersistencePromise<void> {\n const results: Array<PersistencePromise<void>> = [];\n return new PersistencePromise((resolve, reject) => {\n cursorRequest.onerror = (event: Event) => {\n reject((event.target as IDBRequest).error!);\n };\n cursorRequest.onsuccess = (event: Event) => {\n const cursor: IDBCursorWithValue = (event.target as IDBRequest).result;\n if (!cursor) {\n resolve();\n return;\n }\n const controller = new IterationController(cursor);\n const userResult = fn(\n cursor.primaryKey as KeyType,\n cursor.value,\n controller\n );\n if (userResult instanceof PersistencePromise) {\n const userPromise: PersistencePromise<void> = userResult.catch(\n err => {\n controller.done();\n return PersistencePromise.reject(err);\n }\n );\n results.push(userPromise);\n }\n if (controller.isDone) {\n resolve();\n } else if (controller.skipToKey === null) {\n cursor.continue();\n } else {\n cursor.continue(controller.skipToKey);\n }\n };\n }).next(() => {\n return PersistencePromise.waitFor(results);\n });\n }\n\n private options(\n indexOrRange?: string | IDBKeyRange,\n range?: IDBKeyRange\n ): IterateOptions {\n let indexName: string | undefined = undefined;\n if (indexOrRange !== undefined) {\n if (typeof indexOrRange === 'string') {\n indexName = indexOrRange;\n } else {\n debugAssert(\n range === undefined,\n '3rd argument must not be defined if 2nd is a range.'\n );\n range = indexOrRange;\n }\n }\n return { index: indexName, range };\n }\n\n private cursor(options: IterateOptions): IDBRequest {\n let direction: IDBCursorDirection = 'next';\n if (options.reverse) {\n direction = 'prev';\n }\n if (options.index) {\n const index = this.store.index(options.index);\n if (options.keysOnly) {\n return index.openKeyCursor(options.range, direction);\n } else {\n return index.openCursor(options.range, direction);\n }\n } else {\n return this.store.openCursor(options.range, direction);\n }\n }\n}\n\n/**\n * Wraps an IDBRequest in a PersistencePromise, using the onsuccess / onerror\n * handlers to resolve / reject the PersistencePromise as appropriate.\n */\nfunction wrapRequest<R>(request: IDBRequest): PersistencePromise<R> {\n return new PersistencePromise<R>((resolve, reject) => {\n request.onsuccess = (event: Event) => {\n const result = (event.target as IDBRequest).result;\n resolve(result);\n };\n\n request.onerror = (event: Event) => {\n const error = checkForAndReportiOSError(\n (event.target as IDBRequest).error!\n );\n reject(error);\n };\n });\n}\n\n// Guard so we only report the error once.\nlet reportedIOSError = false;\nfunction checkForAndReportiOSError(error: DOMException): Error {\n const iOSVersion = SimpleDb.getIOSVersion(getUA());\n if (iOSVersion >= 12.2 && iOSVersion < 13) {\n const IOS_ERROR =\n 'An internal error was encountered in the Indexed Database server';\n if (error.message.indexOf(IOS_ERROR) >= 0) {\n // Wrap error in a more descriptive one.\n const newError = new FirestoreError(\n 'internal',\n `IOS_INDEXEDDB_BUG1: IndexedDb has thrown '${IOS_ERROR}'. This is likely ` +\n `due to an unavoidable bug in iOS. See https://stackoverflow.com/q/56496296/110915 ` +\n `for details and a potential workaround.`\n );\n if (!reportedIOSError) {\n reportedIOSError = true;\n // Throw a global exception outside of this promise chain, for the user to\n // potentially catch.\n setTimeout(() => {\n throw newError;\n }, 0);\n }\n return newError;\n }\n }\n return error;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert, fail } from './assert';\nimport { Code, FirestoreError } from './error';\nimport { logDebug, logError } from './log';\nimport { Deferred } from './promise';\nimport { ExponentialBackoff } from '../remote/backoff';\nimport { PlatformSupport } from '../platform/platform';\nimport { isIndexedDbTransactionError } from '../local/simple_db';\n\nconst LOG_TAG = 'AsyncQueue';\n\n// Accept any return type from setTimeout().\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype TimerHandle = any;\n\n/**\n * Wellknown \"timer\" IDs used when scheduling delayed operations on the\n * AsyncQueue. These IDs can then be used from tests to check for the presence\n * of operations or to run them early.\n *\n * The string values are used when encoding these timer IDs in JSON spec tests.\n */\nexport const enum TimerId {\n /** All can be used with runDelayedOperationsEarly() to run all timers. */\n All = 'all',\n\n /**\n * The following 4 timers are used in persistent_stream.ts for the listen and\n * write streams. The \"Idle\" timer is used to close the stream due to\n * inactivity. The \"ConnectionBackoff\" timer is used to restart a stream once\n * the appropriate backoff delay has elapsed.\n */\n ListenStreamIdle = 'listen_stream_idle',\n ListenStreamConnectionBackoff = 'listen_stream_connection_backoff',\n WriteStreamIdle = 'write_stream_idle',\n WriteStreamConnectionBackoff = 'write_stream_connection_backoff',\n\n /**\n * A timer used in online_state_tracker.ts to transition from\n * OnlineState.Unknown to Offline after a set timeout, rather than waiting\n * indefinitely for success or failure.\n */\n OnlineStateTimeout = 'online_state_timeout',\n\n /**\n * A timer used to update the client metadata in IndexedDb, which is used\n * to determine the primary leaseholder.\n */\n ClientMetadataRefresh = 'client_metadata_refresh',\n\n /** A timer used to periodically attempt LRU Garbage collection */\n LruGarbageCollection = 'lru_garbage_collection',\n\n /**\n * A timer used to retry transactions. Since there can be multiple concurrent\n * transactions, multiple of these may be in the queue at a given time.\n */\n TransactionRetry = 'transaction_retry',\n\n /**\n * A timer used to retry operations scheduled via retryable AsyncQueue\n * operations.\n */\n AsyncQueueRetry = 'async_queue_retry'\n}\n\n/**\n * Represents an operation scheduled to be run in the future on an AsyncQueue.\n *\n * It is created via DelayedOperation.createAndSchedule().\n *\n * Supports cancellation (via cancel()) and early execution (via skipDelay()).\n *\n * Note: We implement `PromiseLike` instead of `Promise`, as the `Promise` type\n * in newer versions of TypeScript defines `finally`, which is not available in\n * IE.\n */\nexport class DelayedOperation<T extends unknown> implements PromiseLike<T> {\n // handle for use with clearTimeout(), or null if the operation has been\n // executed or canceled already.\n private timerHandle: TimerHandle | null;\n\n private readonly deferred = new Deferred<T>();\n\n private constructor(\n private readonly asyncQueue: AsyncQueue,\n readonly timerId: TimerId,\n readonly targetTimeMs: number,\n private readonly op: () => Promise<T>,\n private readonly removalCallback: (op: DelayedOperation<T>) => void\n ) {\n // It's normal for the deferred promise to be canceled (due to cancellation)\n // and so we attach a dummy catch callback to avoid\n // 'UnhandledPromiseRejectionWarning' log spam.\n this.deferred.promise.catch(err => {});\n }\n\n /**\n * Creates and returns a DelayedOperation that has been scheduled to be\n * executed on the provided asyncQueue after the provided delayMs.\n *\n * @param asyncQueue The queue to schedule the operation on.\n * @param id A Timer ID identifying the type of operation this is.\n * @param delayMs The delay (ms) before the operation should be scheduled.\n * @param op The operation to run.\n * @param removalCallback A callback to be called synchronously once the\n * operation is executed or canceled, notifying the AsyncQueue to remove it\n * from its delayedOperations list.\n * PORTING NOTE: This exists to prevent making removeDelayedOperation() and\n * the DelayedOperation class public.\n */\n static createAndSchedule<R extends unknown>(\n asyncQueue: AsyncQueue,\n timerId: TimerId,\n delayMs: number,\n op: () => Promise<R>,\n removalCallback: (op: DelayedOperation<R>) => void\n ): DelayedOperation<R> {\n const targetTime = Date.now() + delayMs;\n const delayedOp = new DelayedOperation(\n asyncQueue,\n timerId,\n targetTime,\n op,\n removalCallback\n );\n delayedOp.start(delayMs);\n return delayedOp;\n }\n\n /**\n * Starts the timer. This is called immediately after construction by\n * createAndSchedule().\n */\n private start(delayMs: number): void {\n this.timerHandle = setTimeout(() => this.handleDelayElapsed(), delayMs);\n }\n\n /**\n * Queues the operation to run immediately (if it hasn't already been run or\n * canceled).\n */\n skipDelay(): void {\n return this.handleDelayElapsed();\n }\n\n /**\n * Cancels the operation if it hasn't already been executed or canceled. The\n * promise will be rejected.\n *\n * As long as the operation has not yet been run, calling cancel() provides a\n * guarantee that the operation will not be run.\n */\n cancel(reason?: string): void {\n if (this.timerHandle !== null) {\n this.clearTimeout();\n this.deferred.reject(\n new FirestoreError(\n Code.CANCELLED,\n 'Operation cancelled' + (reason ? ': ' + reason : '')\n )\n );\n }\n }\n\n then = this.deferred.promise.then.bind(this.deferred.promise);\n\n private handleDelayElapsed(): void {\n this.asyncQueue.enqueueAndForget(() => {\n if (this.timerHandle !== null) {\n this.clearTimeout();\n return this.op().then(result => {\n return this.deferred.resolve(result);\n });\n } else {\n return Promise.resolve();\n }\n });\n }\n\n private clearTimeout(): void {\n if (this.timerHandle !== null) {\n this.removalCallback(this);\n clearTimeout(this.timerHandle);\n this.timerHandle = null;\n }\n }\n}\n\nexport class AsyncQueue {\n // The last promise in the queue.\n private tail: Promise<unknown> = Promise.resolve();\n\n // The last retryable operation. Retryable operation are run in order and\n // retried with backoff.\n private retryableTail: Promise<void> = Promise.resolve();\n\n // Is this AsyncQueue being shut down? Once it is set to true, it will not\n // be changed again.\n private _isShuttingDown: boolean = false;\n\n // Operations scheduled to be queued in the future. Operations are\n // automatically removed after they are run or canceled.\n private delayedOperations: Array<DelayedOperation<unknown>> = [];\n\n // visible for testing\n failure: Error | null = null;\n\n // Flag set while there's an outstanding AsyncQueue operation, used for\n // assertion sanity-checks.\n private operationInProgress = false;\n\n // List of TimerIds to fast-forward delays for.\n private timerIdsToSkip: TimerId[] = [];\n\n // Backoff timer used to schedule retries for retryable operations\n private backoff = new ExponentialBackoff(this, TimerId.AsyncQueueRetry);\n\n // Visibility handler that triggers an immediate retry of all retryable\n // operations. Meant to speed up recovery when we regain file system access\n // after page comes into foreground.\n private visibilityHandler = (): void => this.backoff.skipBackoff();\n\n constructor() {\n const window = PlatformSupport.getPlatform().window;\n if (window && typeof window.addEventListener === 'function') {\n window.addEventListener('visibilitychange', this.visibilityHandler);\n }\n }\n\n // Is this AsyncQueue being shut down? If true, this instance will not enqueue\n // any new operations, Promises from enqueue requests will not resolve.\n get isShuttingDown(): boolean {\n return this._isShuttingDown;\n }\n\n /**\n * Adds a new operation to the queue without waiting for it to complete (i.e.\n * we ignore the Promise result).\n */\n enqueueAndForget<T extends unknown>(op: () => Promise<T>): void {\n // eslint-disable-next-line @typescript-eslint/no-floating-promises\n this.enqueue(op);\n }\n\n /**\n * Regardless if the queue has initialized shutdown, adds a new operation to the\n * queue without waiting for it to complete (i.e. we ignore the Promise result).\n */\n enqueueAndForgetEvenAfterShutdown<T extends unknown>(\n op: () => Promise<T>\n ): void {\n this.verifyNotFailed();\n // eslint-disable-next-line @typescript-eslint/no-floating-promises\n this.enqueueInternal(op);\n }\n\n /**\n * Regardless if the queue has initialized shutdown, adds a new operation to the\n * queue.\n */\n private enqueueEvenAfterShutdown<T extends unknown>(\n op: () => Promise<T>\n ): Promise<T> {\n this.verifyNotFailed();\n return this.enqueueInternal(op);\n }\n\n /**\n * Adds a new operation to the queue and initialize the shut down of this queue.\n * Returns a promise that will be resolved when the promise returned by the new\n * operation is (with its value).\n * Once this method is called, the only possible way to request running an operation\n * is through `enqueueAndForgetEvenAfterShutdown`.\n */\n async enqueueAndInitiateShutdown(op: () => Promise<void>): Promise<void> {\n this.verifyNotFailed();\n if (!this._isShuttingDown) {\n this._isShuttingDown = true;\n const window = PlatformSupport.getPlatform().window;\n if (window) {\n window.removeEventListener('visibilitychange', this.visibilityHandler);\n }\n await this.enqueueEvenAfterShutdown(op);\n }\n }\n\n /**\n * Adds a new operation to the queue. Returns a promise that will be resolved\n * when the promise returned by the new operation is (with its value).\n */\n enqueue<T extends unknown>(op: () => Promise<T>): Promise<T> {\n this.verifyNotFailed();\n if (this._isShuttingDown) {\n // Return a Promise which never resolves.\n return new Promise<T>(resolve => {});\n }\n return this.enqueueInternal(op);\n }\n\n /**\n * Enqueue a retryable operation.\n *\n * A retryable operation is rescheduled with backoff if it fails with a\n * IndexedDbTransactionError (the error type used by SimpleDb). All\n * retryable operations are executed in order and only run if all prior\n * operations were retried successfully.\n */\n enqueueRetryable(op: () => Promise<void>): void {\n this.verifyNotFailed();\n\n if (this._isShuttingDown) {\n return;\n }\n\n this.retryableTail = this.retryableTail.then(() => {\n const deferred = new Deferred<void>();\n const retryingOp = async (): Promise<void> => {\n try {\n await op();\n deferred.resolve();\n this.backoff.reset();\n } catch (e) {\n if (isIndexedDbTransactionError(e)) {\n logDebug(LOG_TAG, 'Operation failed with retryable error: ' + e);\n this.backoff.backoffAndRun(retryingOp);\n } else {\n deferred.resolve();\n throw e; // Failure will be handled by AsyncQueue\n }\n }\n };\n this.enqueueAndForget(retryingOp);\n return deferred.promise;\n });\n }\n\n private enqueueInternal<T extends unknown>(op: () => Promise<T>): Promise<T> {\n const newTail = this.tail.then(() => {\n this.operationInProgress = true;\n return op()\n .catch((error: FirestoreError) => {\n this.failure = error;\n this.operationInProgress = false;\n const message = error.stack || error.message || '';\n logError('INTERNAL UNHANDLED ERROR: ', message);\n\n // Re-throw the error so that this.tail becomes a rejected Promise and\n // all further attempts to chain (via .then) will just short-circuit\n // and return the rejected Promise.\n throw error;\n })\n .then(result => {\n this.operationInProgress = false;\n return result;\n });\n });\n this.tail = newTail;\n return newTail;\n }\n\n /**\n * Schedules an operation to be queued on the AsyncQueue once the specified\n * `delayMs` has elapsed. The returned DelayedOperation can be used to cancel\n * or fast-forward the operation prior to its running.\n */\n enqueueAfterDelay<T extends unknown>(\n timerId: TimerId,\n delayMs: number,\n op: () => Promise<T>\n ): DelayedOperation<T> {\n this.verifyNotFailed();\n\n debugAssert(\n delayMs >= 0,\n `Attempted to schedule an operation with a negative delay of ${delayMs}`\n );\n\n // Fast-forward delays for timerIds that have been overriden.\n if (this.timerIdsToSkip.indexOf(timerId) > -1) {\n delayMs = 0;\n }\n\n const delayedOp = DelayedOperation.createAndSchedule<T>(\n this,\n timerId,\n delayMs,\n op,\n removedOp =>\n this.removeDelayedOperation(removedOp as DelayedOperation<unknown>)\n );\n this.delayedOperations.push(delayedOp as DelayedOperation<unknown>);\n return delayedOp;\n }\n\n private verifyNotFailed(): void {\n if (this.failure) {\n fail(\n 'AsyncQueue is already failed: ' +\n (this.failure.stack || this.failure.message)\n );\n }\n }\n\n /**\n * Verifies there's an operation currently in-progress on the AsyncQueue.\n * Unfortunately we can't verify that the running code is in the promise chain\n * of that operation, so this isn't a foolproof check, but it should be enough\n * to catch some bugs.\n */\n verifyOperationInProgress(): void {\n debugAssert(\n this.operationInProgress,\n 'verifyOpInProgress() called when no op in progress on this queue.'\n );\n }\n\n /**\n * Waits until all currently queued tasks are finished executing. Delayed\n * operations are not run.\n */\n async drain(): Promise<void> {\n // Operations in the queue prior to draining may have enqueued additional\n // operations. Keep draining the queue until the tail is no longer advanced,\n // which indicates that no more new operations were enqueued and that all\n // operations were executed.\n let currentTail: Promise<unknown>;\n do {\n currentTail = this.tail;\n await currentTail;\n } while (currentTail !== this.tail);\n }\n\n /**\n * For Tests: Determine if a delayed operation with a particular TimerId\n * exists.\n */\n containsDelayedOperation(timerId: TimerId): boolean {\n for (const op of this.delayedOperations) {\n if (op.timerId === timerId) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * For Tests: Runs some or all delayed operations early.\n *\n * @param lastTimerId Delayed operations up to and including this TimerId will\n * be drained. Pass TimerId.All to run all delayed operations.\n * @returns a Promise that resolves once all operations have been run.\n */\n runAllDelayedOperationsUntil(lastTimerId: TimerId): Promise<void> {\n // Note that draining may generate more delayed ops, so we do that first.\n return this.drain().then(() => {\n // Run ops in the same order they'd run if they ran naturally.\n this.delayedOperations.sort((a, b) => a.targetTimeMs - b.targetTimeMs);\n\n for (const op of this.delayedOperations) {\n op.skipDelay();\n if (lastTimerId !== TimerId.All && op.timerId === lastTimerId) {\n break;\n }\n }\n\n return this.drain();\n });\n }\n\n /**\n * For Tests: Skip all subsequent delays for a timer id.\n */\n skipDelaysForTimerId(timerId: TimerId): void {\n this.timerIdsToSkip.push(timerId);\n }\n\n /** Called once a DelayedOperation is run or canceled. */\n private removeDelayedOperation(op: DelayedOperation<unknown>): void {\n // NOTE: indexOf / slice are O(n), but delayedOperations is expected to be small.\n const index = this.delayedOperations.indexOf(op);\n debugAssert(index >= 0, 'Delayed operation not found.');\n this.delayedOperations.splice(index, 1);\n }\n}\n\n/**\n * Returns a FirestoreError that can be surfaced to the user if the provided\n * error is an IndexedDbTransactionError. Re-throws the error otherwise.\n */\nexport function wrapInUserErrorIfRecoverable(\n e: Error,\n msg: string\n): FirestoreError {\n logError(LOG_TAG, `${msg}: ${e}`);\n if (isIndexedDbTransactionError(e)) {\n return new FirestoreError(Code.UNAVAILABLE, `${msg}: ${e}`);\n } else {\n throw e;\n }\n}\n","/**\n * @license\n * Copyright 2018 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { ListenSequence } from '../core/listen_sequence';\nimport { ListenSequenceNumber, TargetId } from '../core/types';\nimport { debugAssert } from '../util/assert';\nimport { AsyncQueue, DelayedOperation, TimerId } from '../util/async_queue';\nimport { getLogLevel, logDebug, LogLevel } from '../util/log';\nimport { primitiveComparator } from '../util/misc';\nimport { SortedMap } from '../util/sorted_map';\nimport { SortedSet } from '../util/sorted_set';\nimport { ignoreIfPrimaryLeaseLoss, LocalStore } from './local_store';\nimport {\n GarbageCollectionScheduler,\n PersistenceTransaction\n} from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { TargetData } from './target_data';\nimport { isIndexedDbTransactionError } from './simple_db';\n\nconst LOG_TAG = 'LruGarbageCollector';\n\n/**\n * Persistence layers intending to use LRU Garbage collection should have reference delegates that\n * implement this interface. This interface defines the operations that the LRU garbage collector\n * needs from the persistence layer.\n */\nexport interface LruDelegate {\n readonly garbageCollector: LruGarbageCollector;\n\n /** Enumerates all the targets in the TargetCache. */\n forEachTarget(\n txn: PersistenceTransaction,\n f: (target: TargetData) => void\n ): PersistencePromise<void>;\n\n getSequenceNumberCount(\n txn: PersistenceTransaction\n ): PersistencePromise<number>;\n\n /**\n * Enumerates sequence numbers for documents not associated with a target.\n * Note that this may include duplicate sequence numbers.\n */\n forEachOrphanedDocumentSequenceNumber(\n txn: PersistenceTransaction,\n f: (sequenceNumber: ListenSequenceNumber) => void\n ): PersistencePromise<void>;\n\n /**\n * Removes all targets that have a sequence number less than or equal to `upperBound`, and are not\n * present in the `activeTargetIds` set.\n *\n * @return the number of targets removed.\n */\n removeTargets(\n txn: PersistenceTransaction,\n upperBound: ListenSequenceNumber,\n activeTargetIds: ActiveTargets\n ): PersistencePromise<number>;\n\n /**\n * Removes all unreferenced documents from the cache that have a sequence number less than or\n * equal to the given `upperBound`.\n *\n * @return the number of documents removed.\n */\n removeOrphanedDocuments(\n txn: PersistenceTransaction,\n upperBound: ListenSequenceNumber\n ): PersistencePromise<number>;\n\n getCacheSize(txn: PersistenceTransaction): PersistencePromise<number>;\n}\n\n/**\n * Describes a map whose keys are active target ids. We do not care about the type of the\n * values.\n */\nexport type ActiveTargets = SortedMap<TargetId, unknown>;\n\n// The type and comparator for the items contained in the SortedSet used in\n// place of a priority queue for the RollingSequenceNumberBuffer.\ntype BufferEntry = [ListenSequenceNumber, number];\nfunction bufferEntryComparator(\n [aSequence, aIndex]: BufferEntry,\n [bSequence, bIndex]: BufferEntry\n): number {\n const seqCmp = primitiveComparator(aSequence, bSequence);\n if (seqCmp === 0) {\n // This order doesn't matter, but we can bias against churn by sorting\n // entries created earlier as less than newer entries.\n return primitiveComparator(aIndex, bIndex);\n } else {\n return seqCmp;\n }\n}\n\n/**\n * Used to calculate the nth sequence number. Keeps a rolling buffer of the\n * lowest n values passed to `addElement`, and finally reports the largest of\n * them in `maxValue`.\n */\nclass RollingSequenceNumberBuffer {\n private buffer: SortedSet<BufferEntry> = new SortedSet<BufferEntry>(\n bufferEntryComparator\n );\n\n private previousIndex = 0;\n\n constructor(private readonly maxElements: number) {}\n\n private nextIndex(): number {\n return ++this.previousIndex;\n }\n\n addElement(sequenceNumber: ListenSequenceNumber): void {\n const entry: BufferEntry = [sequenceNumber, this.nextIndex()];\n if (this.buffer.size < this.maxElements) {\n this.buffer = this.buffer.add(entry);\n } else {\n const highestValue = this.buffer.last()!;\n if (bufferEntryComparator(entry, highestValue) < 0) {\n this.buffer = this.buffer.delete(highestValue).add(entry);\n }\n }\n }\n\n get maxValue(): ListenSequenceNumber {\n // Guaranteed to be non-empty. If we decide we are not collecting any\n // sequence numbers, nthSequenceNumber below short-circuits. If we have\n // decided that we are collecting n sequence numbers, it's because n is some\n // percentage of the existing sequence numbers. That means we should never\n // be in a situation where we are collecting sequence numbers but don't\n // actually have any.\n return this.buffer.last()![0];\n }\n}\n\n/**\n * Describes the results of a garbage collection run. `didRun` will be set to\n * `false` if collection was skipped (either it is disabled or the cache size\n * has not hit the threshold). If collection ran, the other fields will be\n * filled in with the details of the results.\n */\nexport interface LruResults {\n readonly didRun: boolean;\n readonly sequenceNumbersCollected: number;\n readonly targetsRemoved: number;\n readonly documentsRemoved: number;\n}\n\nconst GC_DID_NOT_RUN: LruResults = {\n didRun: false,\n sequenceNumbersCollected: 0,\n targetsRemoved: 0,\n documentsRemoved: 0\n};\n\nexport class LruParams {\n static readonly COLLECTION_DISABLED = -1;\n static readonly MINIMUM_CACHE_SIZE_BYTES = 1 * 1024 * 1024;\n static readonly DEFAULT_CACHE_SIZE_BYTES = 40 * 1024 * 1024;\n private static readonly DEFAULT_COLLECTION_PERCENTILE = 10;\n private static readonly DEFAULT_MAX_SEQUENCE_NUMBERS_TO_COLLECT = 1000;\n\n static withCacheSize(cacheSize: number): LruParams {\n return new LruParams(\n cacheSize,\n LruParams.DEFAULT_COLLECTION_PERCENTILE,\n LruParams.DEFAULT_MAX_SEQUENCE_NUMBERS_TO_COLLECT\n );\n }\n\n static readonly DEFAULT: LruParams = new LruParams(\n LruParams.DEFAULT_CACHE_SIZE_BYTES,\n LruParams.DEFAULT_COLLECTION_PERCENTILE,\n LruParams.DEFAULT_MAX_SEQUENCE_NUMBERS_TO_COLLECT\n );\n\n static readonly DISABLED: LruParams = new LruParams(\n LruParams.COLLECTION_DISABLED,\n 0,\n 0\n );\n\n constructor(\n // When we attempt to collect, we will only do so if the cache size is greater than this\n // threshold. Passing `COLLECTION_DISABLED` here will cause collection to always be skipped.\n readonly cacheSizeCollectionThreshold: number,\n // The percentage of sequence numbers that we will attempt to collect\n readonly percentileToCollect: number,\n // A cap on the total number of sequence numbers that will be collected. This prevents\n // us from collecting a huge number of sequence numbers if the cache has grown very large.\n readonly maximumSequenceNumbersToCollect: number\n ) {}\n}\n\n/** How long we wait to try running LRU GC after SDK initialization. */\nconst INITIAL_GC_DELAY_MS = 1 * 60 * 1000;\n/** Minimum amount of time between GC checks, after the first one. */\nconst REGULAR_GC_DELAY_MS = 5 * 60 * 1000;\n\n/**\n * This class is responsible for the scheduling of LRU garbage collection. It handles checking\n * whether or not GC is enabled, as well as which delay to use before the next run.\n */\nexport class LruScheduler implements GarbageCollectionScheduler {\n private hasRun: boolean = false;\n private gcTask: DelayedOperation<void> | null;\n\n constructor(\n private readonly garbageCollector: LruGarbageCollector,\n private readonly asyncQueue: AsyncQueue\n ) {\n this.gcTask = null;\n }\n\n start(localStore: LocalStore): void {\n debugAssert(\n this.gcTask === null,\n 'Cannot start an already started LruScheduler'\n );\n if (\n this.garbageCollector.params.cacheSizeCollectionThreshold !==\n LruParams.COLLECTION_DISABLED\n ) {\n this.scheduleGC(localStore);\n }\n }\n\n stop(): void {\n if (this.gcTask) {\n this.gcTask.cancel();\n this.gcTask = null;\n }\n }\n\n get started(): boolean {\n return this.gcTask !== null;\n }\n\n private scheduleGC(localStore: LocalStore): void {\n debugAssert(\n this.gcTask === null,\n 'Cannot schedule GC while a task is pending'\n );\n const delay = this.hasRun ? REGULAR_GC_DELAY_MS : INITIAL_GC_DELAY_MS;\n logDebug(\n 'LruGarbageCollector',\n `Garbage collection scheduled in ${delay}ms`\n );\n this.gcTask = this.asyncQueue.enqueueAfterDelay(\n TimerId.LruGarbageCollection,\n delay,\n async () => {\n this.gcTask = null;\n this.hasRun = true;\n try {\n await localStore.collectGarbage(this.garbageCollector);\n } catch (e) {\n if (isIndexedDbTransactionError(e)) {\n logDebug(\n LOG_TAG,\n 'Ignoring IndexedDB error during garbage collection: ',\n e\n );\n } else {\n await ignoreIfPrimaryLeaseLoss(e);\n }\n }\n await this.scheduleGC(localStore);\n }\n );\n }\n}\n\n/** Implements the steps for LRU garbage collection. */\nexport class LruGarbageCollector {\n constructor(\n private readonly delegate: LruDelegate,\n readonly params: LruParams\n ) {}\n\n /** Given a percentile of target to collect, returns the number of targets to collect. */\n calculateTargetCount(\n txn: PersistenceTransaction,\n percentile: number\n ): PersistencePromise<number> {\n return this.delegate.getSequenceNumberCount(txn).next(targetCount => {\n return Math.floor((percentile / 100.0) * targetCount);\n });\n }\n\n /** Returns the nth sequence number, counting in order from the smallest. */\n nthSequenceNumber(\n txn: PersistenceTransaction,\n n: number\n ): PersistencePromise<ListenSequenceNumber> {\n if (n === 0) {\n return PersistencePromise.resolve(ListenSequence.INVALID);\n }\n\n const buffer = new RollingSequenceNumberBuffer(n);\n return this.delegate\n .forEachTarget(txn, target => buffer.addElement(target.sequenceNumber))\n .next(() => {\n return this.delegate.forEachOrphanedDocumentSequenceNumber(\n txn,\n sequenceNumber => buffer.addElement(sequenceNumber)\n );\n })\n .next(() => buffer.maxValue);\n }\n\n /**\n * Removes targets with a sequence number equal to or less than the given upper bound, and removes\n * document associations with those targets.\n */\n removeTargets(\n txn: PersistenceTransaction,\n upperBound: ListenSequenceNumber,\n activeTargetIds: ActiveTargets\n ): PersistencePromise<number> {\n return this.delegate.removeTargets(txn, upperBound, activeTargetIds);\n }\n\n /**\n * Removes documents that have a sequence number equal to or less than the upper bound and are not\n * otherwise pinned.\n */\n removeOrphanedDocuments(\n txn: PersistenceTransaction,\n upperBound: ListenSequenceNumber\n ): PersistencePromise<number> {\n return this.delegate.removeOrphanedDocuments(txn, upperBound);\n }\n\n collect(\n txn: PersistenceTransaction,\n activeTargetIds: ActiveTargets\n ): PersistencePromise<LruResults> {\n if (\n this.params.cacheSizeCollectionThreshold === LruParams.COLLECTION_DISABLED\n ) {\n logDebug('LruGarbageCollector', 'Garbage collection skipped; disabled');\n return PersistencePromise.resolve(GC_DID_NOT_RUN);\n }\n\n return this.getCacheSize(txn).next(cacheSize => {\n if (cacheSize < this.params.cacheSizeCollectionThreshold) {\n logDebug(\n 'LruGarbageCollector',\n `Garbage collection skipped; Cache size ${cacheSize} ` +\n `is lower than threshold ${this.params.cacheSizeCollectionThreshold}`\n );\n return GC_DID_NOT_RUN;\n } else {\n return this.runGarbageCollection(txn, activeTargetIds);\n }\n });\n }\n\n getCacheSize(txn: PersistenceTransaction): PersistencePromise<number> {\n return this.delegate.getCacheSize(txn);\n }\n\n private runGarbageCollection(\n txn: PersistenceTransaction,\n activeTargetIds: ActiveTargets\n ): PersistencePromise<LruResults> {\n let upperBoundSequenceNumber: number;\n let sequenceNumbersToCollect: number, targetsRemoved: number;\n // Timestamps for various pieces of the process\n let countedTargetsTs: number,\n foundUpperBoundTs: number,\n removedTargetsTs: number,\n removedDocumentsTs: number;\n const startTs = Date.now();\n return this.calculateTargetCount(txn, this.params.percentileToCollect)\n .next(sequenceNumbers => {\n // Cap at the configured max\n if (sequenceNumbers > this.params.maximumSequenceNumbersToCollect) {\n logDebug(\n 'LruGarbageCollector',\n 'Capping sequence numbers to collect down ' +\n `to the maximum of ${this.params.maximumSequenceNumbersToCollect} ` +\n `from ${sequenceNumbers}`\n );\n sequenceNumbersToCollect = this.params\n .maximumSequenceNumbersToCollect;\n } else {\n sequenceNumbersToCollect = sequenceNumbers;\n }\n countedTargetsTs = Date.now();\n\n return this.nthSequenceNumber(txn, sequenceNumbersToCollect);\n })\n .next(upperBound => {\n upperBoundSequenceNumber = upperBound;\n foundUpperBoundTs = Date.now();\n\n return this.removeTargets(\n txn,\n upperBoundSequenceNumber,\n activeTargetIds\n );\n })\n .next(numTargetsRemoved => {\n targetsRemoved = numTargetsRemoved;\n removedTargetsTs = Date.now();\n\n return this.removeOrphanedDocuments(txn, upperBoundSequenceNumber);\n })\n .next(documentsRemoved => {\n removedDocumentsTs = Date.now();\n\n if (getLogLevel() <= LogLevel.DEBUG) {\n const desc =\n 'LRU Garbage Collection\\n' +\n `\\tCounted targets in ${countedTargetsTs - startTs}ms\\n` +\n `\\tDetermined least recently used ${sequenceNumbersToCollect} in ` +\n `${foundUpperBoundTs - countedTargetsTs}ms\\n` +\n `\\tRemoved ${targetsRemoved} targets in ` +\n `${removedTargetsTs - foundUpperBoundTs}ms\\n` +\n `\\tRemoved ${documentsRemoved} documents in ` +\n `${removedDocumentsTs - removedTargetsTs}ms\\n` +\n `Total Duration: ${removedDocumentsTs - startTs}ms`;\n logDebug('LruGarbageCollector', desc);\n }\n\n return PersistencePromise.resolve<LruResults>({\n didRun: true,\n sequenceNumbersCollected: sequenceNumbersToCollect,\n targetsRemoved,\n documentsRemoved\n });\n });\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Timestamp } from '../api/timestamp';\nimport { User } from '../auth/user';\nimport { Query } from '../core/query';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { Target } from '../core/target';\nimport { BatchId, TargetId } from '../core/types';\nimport {\n DocumentKeySet,\n documentKeySet,\n DocumentMap,\n maybeDocumentMap,\n MaybeDocumentMap\n} from '../model/collections';\nimport { MaybeDocument, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { Mutation, PatchMutation, Precondition } from '../model/mutation';\nimport {\n BATCHID_UNKNOWN,\n MutationBatch,\n MutationBatchResult\n} from '../model/mutation_batch';\nimport { RemoteEvent, TargetChange } from '../remote/remote_event';\nimport { hardAssert, debugAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { logDebug } from '../util/log';\nimport { primitiveComparator } from '../util/misc';\nimport { ObjectMap } from '../util/obj_map';\nimport { SortedMap } from '../util/sorted_map';\n\nimport { LocalDocumentsView } from './local_documents_view';\nimport { LocalViewChanges } from './local_view_changes';\nimport { LruGarbageCollector, LruResults } from './lru_garbage_collector';\nimport { MutationQueue } from './mutation_queue';\nimport {\n Persistence,\n PersistenceTransaction,\n PRIMARY_LEASE_LOST_ERROR_MSG\n} from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { TargetCache } from './target_cache';\nimport { QueryEngine } from './query_engine';\nimport { RemoteDocumentCache } from './remote_document_cache';\nimport { RemoteDocumentChangeBuffer } from './remote_document_change_buffer';\nimport { ClientId } from './shared_client_state';\nimport { TargetData, TargetPurpose } from './target_data';\nimport { ByteString } from '../util/byte_string';\nimport { IndexedDbPersistence } from './indexeddb_persistence';\nimport { IndexedDbMutationQueue } from './indexeddb_mutation_queue';\nimport { IndexedDbRemoteDocumentCache } from './indexeddb_remote_document_cache';\nimport { IndexedDbTargetCache } from './indexeddb_target_cache';\nimport { extractFieldMask } from '../model/object_value';\nimport { isIndexedDbTransactionError } from './simple_db';\n\nconst LOG_TAG = 'LocalStore';\n\n/** The result of a write to the local store. */\nexport interface LocalWriteResult {\n batchId: BatchId;\n changes: MaybeDocumentMap;\n}\n\n/** The result of a user-change operation in the local store. */\nexport interface UserChangeResult {\n readonly affectedDocuments: MaybeDocumentMap;\n readonly removedBatchIds: BatchId[];\n readonly addedBatchIds: BatchId[];\n}\n\n/** The result of executing a query against the local store. */\nexport interface QueryResult {\n readonly documents: DocumentMap;\n readonly remoteKeys: DocumentKeySet;\n}\n\n/**\n * Local storage in the Firestore client. Coordinates persistence components\n * like the mutation queue and remote document cache to present a\n * latency-compensated view of stored data.\n *\n * The LocalStore is responsible for accepting mutations from the Sync Engine.\n * Writes from the client are put into a queue as provisional Mutations until\n * they are processed by the RemoteStore and confirmed as having been written\n * to the server.\n *\n * The local store provides the local version of documents that have been\n * modified locally. It maintains the constraint:\n *\n * LocalDocument = RemoteDocument + Active(LocalMutations)\n *\n * (Active mutations are those that are enqueued and have not been previously\n * acknowledged or rejected).\n *\n * The RemoteDocument (\"ground truth\") state is provided via the\n * applyChangeBatch method. It will be some version of a server-provided\n * document OR will be a server-provided document PLUS acknowledged mutations:\n *\n * RemoteDocument' = RemoteDocument + Acknowledged(LocalMutations)\n *\n * Note that this \"dirty\" version of a RemoteDocument will not be identical to a\n * server base version, since it has LocalMutations added to it pending getting\n * an authoritative copy from the server.\n *\n * Since LocalMutations can be rejected by the server, we have to be able to\n * revert a LocalMutation that has already been applied to the LocalDocument\n * (typically done by replaying all remaining LocalMutations to the\n * RemoteDocument to re-apply).\n *\n * The LocalStore is responsible for the garbage collection of the documents it\n * contains. For now, it every doc referenced by a view, the mutation queue, or\n * the RemoteStore.\n *\n * It also maintains the persistence of mapping queries to resume tokens and\n * target ids. It needs to know this data about queries to properly know what\n * docs it would be allowed to garbage collect.\n *\n * The LocalStore must be able to efficiently execute queries against its local\n * cache of the documents, to provide the initial set of results before any\n * remote changes have been received.\n *\n * Note: In TypeScript, most methods return Promises since the implementation\n * may rely on fetching data from IndexedDB which is async.\n * These Promises will only be rejected on an I/O error or other internal\n * (unexpected) failure (e.g. failed assert) and always represent an\n * unrecoverable error (should be caught / reported by the async_queue).\n */\nexport class LocalStore {\n /**\n * The maximum time to leave a resume token buffered without writing it out.\n * This value is arbitrary: it's long enough to avoid several writes\n * (possibly indefinitely if updates come more frequently than this) but\n * short enough that restarting after crashing will still have a pretty\n * recent resume token.\n */\n private static readonly RESUME_TOKEN_MAX_AGE_MICROS = 5 * 60 * 1e6;\n\n /**\n * The set of all mutations that have been sent but not yet been applied to\n * the backend.\n */\n protected mutationQueue: MutationQueue;\n\n /** The set of all cached remote documents. */\n protected remoteDocuments: RemoteDocumentCache;\n\n /**\n * The \"local\" view of all documents (layering mutationQueue on top of\n * remoteDocumentCache).\n */\n protected localDocuments: LocalDocumentsView;\n\n /** Maps a target to its `TargetData`. */\n protected targetCache: TargetCache;\n\n /**\n * Maps a targetID to data about its target.\n *\n * PORTING NOTE: We are using an immutable data structure on Web to make re-runs\n * of `applyRemoteEvent()` idempotent.\n */\n protected targetDataByTarget = new SortedMap<TargetId, TargetData>(\n primitiveComparator\n );\n\n /** Maps a target to its targetID. */\n // TODO(wuandy): Evaluate if TargetId can be part of Target.\n private targetIdByTarget = new ObjectMap<Target, TargetId>(t =>\n t.canonicalId()\n );\n\n /**\n * The read time of the last entry processed by `getNewDocumentChanges()`.\n *\n * PORTING NOTE: This is only used for multi-tab synchronization.\n */\n protected lastDocumentChangeReadTime = SnapshotVersion.min();\n\n constructor(\n /** Manages our in-memory or durable persistence. */\n protected persistence: Persistence,\n private queryEngine: QueryEngine,\n initialUser: User\n ) {\n debugAssert(\n persistence.started,\n 'LocalStore was passed an unstarted persistence implementation'\n );\n this.mutationQueue = persistence.getMutationQueue(initialUser);\n this.remoteDocuments = persistence.getRemoteDocumentCache();\n this.targetCache = persistence.getTargetCache();\n this.localDocuments = new LocalDocumentsView(\n this.remoteDocuments,\n this.mutationQueue,\n this.persistence.getIndexManager()\n );\n this.queryEngine.setLocalDocumentsView(this.localDocuments);\n }\n\n /** Starts the LocalStore. */\n start(): Promise<void> {\n return Promise.resolve();\n }\n\n /**\n * Tells the LocalStore that the currently authenticated user has changed.\n *\n * In response the local store switches the mutation queue to the new user and\n * returns any resulting document changes.\n */\n // PORTING NOTE: Android and iOS only return the documents affected by the\n // change.\n async handleUserChange(user: User): Promise<UserChangeResult> {\n let newMutationQueue = this.mutationQueue;\n let newLocalDocuments = this.localDocuments;\n\n const result = await this.persistence.runTransaction(\n 'Handle user change',\n 'readonly',\n txn => {\n // Swap out the mutation queue, grabbing the pending mutation batches\n // before and after.\n let oldBatches: MutationBatch[];\n return this.mutationQueue\n .getAllMutationBatches(txn)\n .next(promisedOldBatches => {\n oldBatches = promisedOldBatches;\n\n newMutationQueue = this.persistence.getMutationQueue(user);\n\n // Recreate our LocalDocumentsView using the new\n // MutationQueue.\n newLocalDocuments = new LocalDocumentsView(\n this.remoteDocuments,\n newMutationQueue,\n this.persistence.getIndexManager()\n );\n return newMutationQueue.getAllMutationBatches(txn);\n })\n .next(newBatches => {\n const removedBatchIds: BatchId[] = [];\n const addedBatchIds: BatchId[] = [];\n\n // Union the old/new changed keys.\n let changedKeys = documentKeySet();\n\n for (const batch of oldBatches) {\n removedBatchIds.push(batch.batchId);\n for (const mutation of batch.mutations) {\n changedKeys = changedKeys.add(mutation.key);\n }\n }\n\n for (const batch of newBatches) {\n addedBatchIds.push(batch.batchId);\n for (const mutation of batch.mutations) {\n changedKeys = changedKeys.add(mutation.key);\n }\n }\n\n // Return the set of all (potentially) changed documents and the list\n // of mutation batch IDs that were affected by change.\n return newLocalDocuments\n .getDocuments(txn, changedKeys)\n .next(affectedDocuments => {\n return {\n affectedDocuments,\n removedBatchIds,\n addedBatchIds\n };\n });\n });\n }\n );\n\n this.mutationQueue = newMutationQueue;\n this.localDocuments = newLocalDocuments;\n this.queryEngine.setLocalDocumentsView(this.localDocuments);\n\n return result;\n }\n\n /* Accept locally generated Mutations and commit them to storage. */\n localWrite(mutations: Mutation[]): Promise<LocalWriteResult> {\n const localWriteTime = Timestamp.now();\n const keys = mutations.reduce(\n (keys, m) => keys.add(m.key),\n documentKeySet()\n );\n\n let existingDocs: MaybeDocumentMap;\n\n return this.persistence\n .runTransaction('Locally write mutations', 'readwrite', txn => {\n // Load and apply all existing mutations. This lets us compute the\n // current base state for all non-idempotent transforms before applying\n // any additional user-provided writes.\n return this.localDocuments.getDocuments(txn, keys).next(docs => {\n existingDocs = docs;\n\n // For non-idempotent mutations (such as `FieldValue.increment()`),\n // we record the base state in a separate patch mutation. This is\n // later used to guarantee consistent values and prevents flicker\n // even if the backend sends us an update that already includes our\n // transform.\n const baseMutations: Mutation[] = [];\n\n for (const mutation of mutations) {\n const baseValue = mutation.extractBaseValue(\n existingDocs.get(mutation.key)\n );\n if (baseValue != null) {\n // NOTE: The base state should only be applied if there's some\n // existing document to override, so use a Precondition of\n // exists=true\n baseMutations.push(\n new PatchMutation(\n mutation.key,\n baseValue,\n extractFieldMask(baseValue.proto.mapValue!),\n Precondition.exists(true)\n )\n );\n }\n }\n\n return this.mutationQueue.addMutationBatch(\n txn,\n localWriteTime,\n baseMutations,\n mutations\n );\n });\n })\n .then(batch => {\n const changes = batch.applyToLocalDocumentSet(existingDocs);\n return { batchId: batch.batchId, changes };\n });\n }\n\n /**\n * Acknowledge the given batch.\n *\n * On the happy path when a batch is acknowledged, the local store will\n *\n * + remove the batch from the mutation queue;\n * + apply the changes to the remote document cache;\n * + recalculate the latency compensated view implied by those changes (there\n * may be mutations in the queue that affect the documents but haven't been\n * acknowledged yet); and\n * + give the changed documents back the sync engine\n *\n * @returns The resulting (modified) documents.\n */\n acknowledgeBatch(\n batchResult: MutationBatchResult\n ): Promise<MaybeDocumentMap> {\n return this.persistence.runTransaction(\n 'Acknowledge batch',\n 'readwrite-primary',\n txn => {\n const affected = batchResult.batch.keys();\n const documentBuffer = this.remoteDocuments.newChangeBuffer({\n trackRemovals: true // Make sure document removals show up in `getNewDocumentChanges()`\n });\n return this.mutationQueue\n .acknowledgeBatch(txn, batchResult.batch, batchResult.streamToken)\n .next(() =>\n this.applyWriteToRemoteDocuments(txn, batchResult, documentBuffer)\n )\n .next(() => documentBuffer.apply(txn))\n .next(() => this.mutationQueue.performConsistencyCheck(txn))\n .next(() => this.localDocuments.getDocuments(txn, affected));\n }\n );\n }\n\n /**\n * Remove mutations from the MutationQueue for the specified batch;\n * LocalDocuments will be recalculated.\n *\n * @returns The resulting modified documents.\n */\n rejectBatch(batchId: BatchId): Promise<MaybeDocumentMap> {\n return this.persistence.runTransaction(\n 'Reject batch',\n 'readwrite-primary',\n txn => {\n let affectedKeys: DocumentKeySet;\n return this.mutationQueue\n .lookupMutationBatch(txn, batchId)\n .next((batch: MutationBatch | null) => {\n hardAssert(batch !== null, 'Attempt to reject nonexistent batch!');\n affectedKeys = batch.keys();\n return this.mutationQueue.removeMutationBatch(txn, batch);\n })\n .next(() => {\n return this.mutationQueue.performConsistencyCheck(txn);\n })\n .next(() => {\n return this.localDocuments.getDocuments(txn, affectedKeys);\n });\n }\n );\n }\n\n /**\n * Returns the largest (latest) batch id in mutation queue that is pending server response.\n * Returns `BATCHID_UNKNOWN` if the queue is empty.\n */\n getHighestUnacknowledgedBatchId(): Promise<BatchId> {\n return this.persistence.runTransaction(\n 'Get highest unacknowledged batch id',\n 'readonly',\n txn => {\n return this.mutationQueue.getHighestUnacknowledgedBatchId(txn);\n }\n );\n }\n\n /** Returns the last recorded stream token for the current user. */\n getLastStreamToken(): Promise<ByteString> {\n return this.persistence.runTransaction(\n 'Get last stream token',\n 'readonly',\n txn => {\n return this.mutationQueue.getLastStreamToken(txn);\n }\n );\n }\n\n /**\n * Sets the stream token for the current user without acknowledging any\n * mutation batch. This is usually only useful after a stream handshake or in\n * response to an error that requires clearing the stream token.\n */\n setLastStreamToken(streamToken: ByteString): Promise<void> {\n return this.persistence.runTransaction(\n 'Set last stream token',\n 'readwrite-primary',\n txn => {\n return this.mutationQueue.setLastStreamToken(txn, streamToken);\n }\n );\n }\n\n /**\n * Returns the last consistent snapshot processed (used by the RemoteStore to\n * determine whether to buffer incoming snapshots from the backend).\n */\n getLastRemoteSnapshotVersion(): Promise<SnapshotVersion> {\n return this.persistence.runTransaction(\n 'Get last remote snapshot version',\n 'readonly',\n txn => this.targetCache.getLastRemoteSnapshotVersion(txn)\n );\n }\n\n /**\n * Update the \"ground-state\" (remote) documents. We assume that the remote\n * event reflects any write batches that have been acknowledged or rejected\n * (i.e. we do not re-apply local mutations to updates from this event).\n *\n * LocalDocuments are re-calculated if there are remaining mutations in the\n * queue.\n */\n applyRemoteEvent(remoteEvent: RemoteEvent): Promise<MaybeDocumentMap> {\n const remoteVersion = remoteEvent.snapshotVersion;\n let newTargetDataByTargetMap = this.targetDataByTarget;\n\n return this.persistence\n .runTransaction('Apply remote event', 'readwrite-primary', txn => {\n const documentBuffer = this.remoteDocuments.newChangeBuffer({\n trackRemovals: true // Make sure document removals show up in `getNewDocumentChanges()`\n });\n\n // Reset newTargetDataByTargetMap in case this transaction gets re-run.\n newTargetDataByTargetMap = this.targetDataByTarget;\n\n const promises = [] as Array<PersistencePromise<void>>;\n remoteEvent.targetChanges.forEach((change, targetId) => {\n const oldTargetData = newTargetDataByTargetMap.get(targetId);\n if (!oldTargetData) {\n return;\n }\n\n // Only update the remote keys if the target is still active. This\n // ensures that we can persist the updated target data along with\n // the updated assignment.\n promises.push(\n this.targetCache\n .removeMatchingKeys(txn, change.removedDocuments, targetId)\n .next(() => {\n return this.targetCache.addMatchingKeys(\n txn,\n change.addedDocuments,\n targetId\n );\n })\n );\n\n const resumeToken = change.resumeToken;\n // Update the resume token if the change includes one.\n if (resumeToken.approximateByteSize() > 0) {\n const newTargetData = oldTargetData\n .withResumeToken(resumeToken, remoteVersion)\n .withSequenceNumber(txn.currentSequenceNumber);\n newTargetDataByTargetMap = newTargetDataByTargetMap.insert(\n targetId,\n newTargetData\n );\n\n // Update the target data if there are target changes (or if\n // sufficient time has passed since the last update).\n if (\n LocalStore.shouldPersistTargetData(\n oldTargetData,\n newTargetData,\n change\n )\n ) {\n promises.push(\n this.targetCache.updateTargetData(txn, newTargetData)\n );\n }\n }\n });\n\n let changedDocs = maybeDocumentMap();\n let updatedKeys = documentKeySet();\n remoteEvent.documentUpdates.forEach((key, doc) => {\n updatedKeys = updatedKeys.add(key);\n });\n\n // Each loop iteration only affects its \"own\" doc, so it's safe to get all the remote\n // documents in advance in a single call.\n promises.push(\n documentBuffer.getEntries(txn, updatedKeys).next(existingDocs => {\n remoteEvent.documentUpdates.forEach((key, doc) => {\n const existingDoc = existingDocs.get(key);\n\n // Note: The order of the steps below is important, since we want\n // to ensure that rejected limbo resolutions (which fabricate\n // NoDocuments with SnapshotVersion.min()) never add documents to\n // cache.\n if (\n doc instanceof NoDocument &&\n doc.version.isEqual(SnapshotVersion.min())\n ) {\n // NoDocuments with SnapshotVersion.min() are used in manufactured\n // events. We remove these documents from cache since we lost\n // access.\n documentBuffer.removeEntry(key, remoteVersion);\n changedDocs = changedDocs.insert(key, doc);\n } else if (\n existingDoc == null ||\n doc.version.compareTo(existingDoc.version) > 0 ||\n (doc.version.compareTo(existingDoc.version) === 0 &&\n existingDoc.hasPendingWrites)\n ) {\n debugAssert(\n !SnapshotVersion.min().isEqual(remoteVersion),\n 'Cannot add a document when the remote version is zero'\n );\n documentBuffer.addEntry(doc, remoteVersion);\n changedDocs = changedDocs.insert(key, doc);\n } else {\n logDebug(\n LOG_TAG,\n 'Ignoring outdated watch update for ',\n key,\n '. Current version:',\n existingDoc.version,\n ' Watch version:',\n doc.version\n );\n }\n\n if (remoteEvent.resolvedLimboDocuments.has(key)) {\n promises.push(\n this.persistence.referenceDelegate.updateLimboDocument(\n txn,\n key\n )\n );\n }\n });\n })\n );\n\n // HACK: The only reason we allow a null snapshot version is so that we\n // can synthesize remote events when we get permission denied errors while\n // trying to resolve the state of a locally cached document that is in\n // limbo.\n if (!remoteVersion.isEqual(SnapshotVersion.min())) {\n const updateRemoteVersion = this.targetCache\n .getLastRemoteSnapshotVersion(txn)\n .next(lastRemoteSnapshotVersion => {\n debugAssert(\n remoteVersion.compareTo(lastRemoteSnapshotVersion) >= 0,\n 'Watch stream reverted to previous snapshot?? ' +\n remoteVersion +\n ' < ' +\n lastRemoteSnapshotVersion\n );\n return this.targetCache.setTargetsMetadata(\n txn,\n txn.currentSequenceNumber,\n remoteVersion\n );\n });\n promises.push(updateRemoteVersion);\n }\n\n return PersistencePromise.waitFor(promises)\n .next(() => documentBuffer.apply(txn))\n .next(() => {\n return this.localDocuments.getLocalViewOfDocuments(\n txn,\n changedDocs\n );\n });\n })\n .then(changedDocs => {\n this.targetDataByTarget = newTargetDataByTargetMap;\n return changedDocs;\n });\n }\n\n /**\n * Returns true if the newTargetData should be persisted during an update of\n * an active target. TargetData should always be persisted when a target is\n * being released and should not call this function.\n *\n * While the target is active, TargetData updates can be omitted when nothing\n * about the target has changed except metadata like the resume token or\n * snapshot version. Occasionally it's worth the extra write to prevent these\n * values from getting too stale after a crash, but this doesn't have to be\n * too frequent.\n */\n private static shouldPersistTargetData(\n oldTargetData: TargetData,\n newTargetData: TargetData,\n change: TargetChange\n ): boolean {\n hardAssert(\n newTargetData.resumeToken.approximateByteSize() > 0,\n 'Attempted to persist target data with no resume token'\n );\n\n // Always persist target data if we don't already have a resume token.\n if (oldTargetData.resumeToken.approximateByteSize() === 0) {\n return true;\n }\n\n // Don't allow resume token changes to be buffered indefinitely. This\n // allows us to be reasonably up-to-date after a crash and avoids needing\n // to loop over all active queries on shutdown. Especially in the browser\n // we may not get time to do anything interesting while the current tab is\n // closing.\n const timeDelta =\n newTargetData.snapshotVersion.toMicroseconds() -\n oldTargetData.snapshotVersion.toMicroseconds();\n if (timeDelta >= this.RESUME_TOKEN_MAX_AGE_MICROS) {\n return true;\n }\n\n // Otherwise if the only thing that has changed about a target is its resume\n // token it's not worth persisting. Note that the RemoteStore keeps an\n // in-memory view of the currently active targets which includes the current\n // resume token, so stream failure or user changes will still use an\n // up-to-date resume token regardless of what we do here.\n const changes =\n change.addedDocuments.size +\n change.modifiedDocuments.size +\n change.removedDocuments.size;\n return changes > 0;\n }\n\n /**\n * Notify local store of the changed views to locally pin documents.\n */\n async notifyLocalViewChanges(viewChanges: LocalViewChanges[]): Promise<void> {\n try {\n await this.persistence.runTransaction(\n 'notifyLocalViewChanges',\n 'readwrite',\n txn => {\n return PersistencePromise.forEach(\n viewChanges,\n (viewChange: LocalViewChanges) => {\n return PersistencePromise.forEach(\n viewChange.addedKeys,\n (key: DocumentKey) =>\n this.persistence.referenceDelegate.addReference(\n txn,\n viewChange.targetId,\n key\n )\n ).next(() =>\n PersistencePromise.forEach(\n viewChange.removedKeys,\n (key: DocumentKey) =>\n this.persistence.referenceDelegate.removeReference(\n txn,\n viewChange.targetId,\n key\n )\n )\n );\n }\n );\n }\n );\n } catch (e) {\n if (isIndexedDbTransactionError(e)) {\n // If `notifyLocalViewChanges` fails, we did not advance the sequence\n // number for the documents that were included in this transaction.\n // This might trigger them to be deleted earlier than they otherwise\n // would have, but it should not invalidate the integrity of the data.\n logDebug(LOG_TAG, 'Failed to update sequence numbers: ' + e);\n } else {\n throw e;\n }\n }\n\n for (const viewChange of viewChanges) {\n const targetId = viewChange.targetId;\n\n if (!viewChange.fromCache) {\n const targetData = this.targetDataByTarget.get(targetId);\n debugAssert(\n targetData !== null,\n `Can't set limbo-free snapshot version for unknown target: ${targetId}`\n );\n\n // Advance the last limbo free snapshot version\n const lastLimboFreeSnapshotVersion = targetData.snapshotVersion;\n const updatedTargetData = targetData.withLastLimboFreeSnapshotVersion(\n lastLimboFreeSnapshotVersion\n );\n this.targetDataByTarget = this.targetDataByTarget.insert(\n targetId,\n updatedTargetData\n );\n }\n }\n }\n\n /**\n * Gets the mutation batch after the passed in batchId in the mutation queue\n * or null if empty.\n * @param afterBatchId If provided, the batch to search after.\n * @returns The next mutation or null if there wasn't one.\n */\n nextMutationBatch(afterBatchId?: BatchId): Promise<MutationBatch | null> {\n return this.persistence.runTransaction(\n 'Get next mutation batch',\n 'readonly',\n txn => {\n if (afterBatchId === undefined) {\n afterBatchId = BATCHID_UNKNOWN;\n }\n return this.mutationQueue.getNextMutationBatchAfterBatchId(\n txn,\n afterBatchId\n );\n }\n );\n }\n\n /**\n * Read the current value of a Document with a given key or null if not\n * found - used for testing.\n */\n readDocument(key: DocumentKey): Promise<MaybeDocument | null> {\n return this.persistence.runTransaction('read document', 'readonly', txn => {\n return this.localDocuments.getDocument(txn, key);\n });\n }\n\n /**\n * Assigns the given target an internal ID so that its results can be pinned so\n * they don't get GC'd. A target must be allocated in the local store before\n * the store can be used to manage its view.\n *\n * Allocating an already allocated `Target` will return the existing `TargetData`\n * for that `Target`.\n */\n allocateTarget(target: Target): Promise<TargetData> {\n return this.persistence\n .runTransaction('Allocate target', 'readwrite', txn => {\n let targetData: TargetData;\n return this.targetCache\n .getTargetData(txn, target)\n .next((cached: TargetData | null) => {\n if (cached) {\n // This target has been listened to previously, so reuse the\n // previous targetID.\n // TODO(mcg): freshen last accessed date?\n targetData = cached;\n return PersistencePromise.resolve(targetData);\n } else {\n return this.targetCache.allocateTargetId(txn).next(targetId => {\n targetData = new TargetData(\n target,\n targetId,\n TargetPurpose.Listen,\n txn.currentSequenceNumber\n );\n return this.targetCache\n .addTargetData(txn, targetData)\n .next(() => targetData);\n });\n }\n });\n })\n .then(targetData => {\n if (this.targetDataByTarget.get(targetData.targetId) === null) {\n this.targetDataByTarget = this.targetDataByTarget.insert(\n targetData.targetId,\n targetData\n );\n this.targetIdByTarget.set(target, targetData.targetId);\n }\n return targetData;\n });\n }\n\n /**\n * Returns the TargetData as seen by the LocalStore, including updates that may\n * have not yet been persisted to the TargetCache.\n */\n // Visible for testing.\n getTargetData(\n transaction: PersistenceTransaction,\n target: Target\n ): PersistencePromise<TargetData | null> {\n const targetId = this.targetIdByTarget.get(target);\n if (targetId !== undefined) {\n return PersistencePromise.resolve<TargetData | null>(\n this.targetDataByTarget.get(targetId)\n );\n } else {\n return this.targetCache.getTargetData(transaction, target);\n }\n }\n\n /**\n * Unpin all the documents associated with the given target. If\n * `keepPersistedTargetData` is set to false and Eager GC enabled, the method\n * directly removes the associated target data from the target cache.\n *\n * Releasing a non-existing `Target` is a no-op.\n */\n // PORTING NOTE: `keepPersistedTargetData` is multi-tab only.\n releaseTarget(\n targetId: number,\n keepPersistedTargetData: boolean\n ): Promise<void> {\n const targetData = this.targetDataByTarget.get(targetId);\n debugAssert(\n targetData !== null,\n `Tried to release nonexistent target: ${targetId}`\n );\n\n const mode = keepPersistedTargetData ? 'readwrite' : 'readwrite-primary';\n return this.persistence\n .runTransaction('Release target', mode, txn => {\n if (!keepPersistedTargetData) {\n return this.persistence.referenceDelegate.removeTarget(\n txn,\n targetData!\n );\n } else {\n return PersistencePromise.resolve();\n }\n })\n .then(() => {\n this.targetDataByTarget = this.targetDataByTarget.remove(targetId);\n this.targetIdByTarget.delete(targetData!.target);\n });\n }\n\n /**\n * Runs the specified query against the local store and returns the results,\n * potentially taking advantage of query data from previous executions (such\n * as the set of remote keys).\n *\n * @param usePreviousResults Whether results from previous executions can\n * be used to optimize this query execution.\n */\n executeQuery(\n query: Query,\n usePreviousResults: boolean\n ): Promise<QueryResult> {\n let lastLimboFreeSnapshotVersion = SnapshotVersion.min();\n let remoteKeys = documentKeySet();\n\n return this.persistence.runTransaction('Execute query', 'readonly', txn => {\n return this.getTargetData(txn, query.toTarget())\n .next(targetData => {\n if (targetData) {\n lastLimboFreeSnapshotVersion =\n targetData.lastLimboFreeSnapshotVersion;\n return this.targetCache\n .getMatchingKeysForTargetId(txn, targetData.targetId)\n .next(result => {\n remoteKeys = result;\n });\n }\n })\n .next(() =>\n this.queryEngine.getDocumentsMatchingQuery(\n txn,\n query,\n usePreviousResults\n ? lastLimboFreeSnapshotVersion\n : SnapshotVersion.min(),\n usePreviousResults ? remoteKeys : documentKeySet()\n )\n )\n .next(documents => {\n return { documents, remoteKeys };\n });\n });\n }\n\n private applyWriteToRemoteDocuments(\n txn: PersistenceTransaction,\n batchResult: MutationBatchResult,\n documentBuffer: RemoteDocumentChangeBuffer\n ): PersistencePromise<void> {\n const batch = batchResult.batch;\n const docKeys = batch.keys();\n let promiseChain = PersistencePromise.resolve();\n docKeys.forEach(docKey => {\n promiseChain = promiseChain\n .next(() => {\n return documentBuffer.getEntry(txn, docKey);\n })\n .next((remoteDoc: MaybeDocument | null) => {\n let doc = remoteDoc;\n const ackVersion = batchResult.docVersions.get(docKey);\n hardAssert(\n ackVersion !== null,\n 'ackVersions should contain every doc in the write.'\n );\n if (!doc || doc.version.compareTo(ackVersion!) < 0) {\n doc = batch.applyToRemoteDocument(docKey, doc, batchResult);\n if (!doc) {\n debugAssert(\n !remoteDoc,\n 'Mutation batch ' +\n batch +\n ' applied to document ' +\n remoteDoc +\n ' resulted in null'\n );\n } else {\n // We use the commitVersion as the readTime rather than the\n // document's updateTime since the updateTime is not advanced\n // for updates that do not modify the underlying document.\n documentBuffer.addEntry(doc, batchResult.commitVersion);\n }\n }\n });\n });\n return promiseChain.next(() =>\n this.mutationQueue.removeMutationBatch(txn, batch)\n );\n }\n\n collectGarbage(garbageCollector: LruGarbageCollector): Promise<LruResults> {\n return this.persistence.runTransaction(\n 'Collect garbage',\n 'readwrite-primary',\n txn => garbageCollector.collect(txn, this.targetDataByTarget)\n );\n }\n}\n\n/**\n * An implementation of LocalStore that provides additional functionality\n * for MultiTabSyncEngine.\n */\n// PORTING NOTE: Web only.\nexport class MultiTabLocalStore extends LocalStore {\n protected mutationQueue: IndexedDbMutationQueue;\n protected remoteDocuments: IndexedDbRemoteDocumentCache;\n protected targetCache: IndexedDbTargetCache;\n\n constructor(\n protected persistence: IndexedDbPersistence,\n queryEngine: QueryEngine,\n initialUser: User\n ) {\n super(persistence, queryEngine, initialUser);\n\n this.mutationQueue = persistence.getMutationQueue(initialUser);\n this.remoteDocuments = persistence.getRemoteDocumentCache();\n this.targetCache = persistence.getTargetCache();\n }\n\n /** Starts the LocalStore. */\n start(): Promise<void> {\n return this.synchronizeLastDocumentChangeReadTime();\n }\n\n /** Returns the local view of the documents affected by a mutation batch. */\n lookupMutationDocuments(batchId: BatchId): Promise<MaybeDocumentMap | null> {\n return this.persistence.runTransaction(\n 'Lookup mutation documents',\n 'readonly',\n txn => {\n return this.mutationQueue\n .lookupMutationKeys(txn, batchId)\n .next(keys => {\n if (keys) {\n return this.localDocuments.getDocuments(\n txn,\n keys\n ) as PersistencePromise<MaybeDocumentMap | null>;\n } else {\n return PersistencePromise.resolve<MaybeDocumentMap | null>(null);\n }\n });\n }\n );\n }\n\n removeCachedMutationBatchMetadata(batchId: BatchId): void {\n this.mutationQueue.removeCachedMutationKeys(batchId);\n }\n\n setNetworkEnabled(networkEnabled: boolean): void {\n this.persistence.setNetworkEnabled(networkEnabled);\n }\n\n getActiveClients(): Promise<ClientId[]> {\n return this.persistence.getActiveClients();\n }\n\n getTarget(targetId: TargetId): Promise<Target | null> {\n const cachedTargetData = this.targetDataByTarget.get(targetId);\n\n if (cachedTargetData) {\n return Promise.resolve(cachedTargetData.target);\n } else {\n return this.persistence.runTransaction(\n 'Get target data',\n 'readonly',\n txn => {\n return this.targetCache\n .getTargetDataForTarget(txn, targetId)\n .next(targetData => (targetData ? targetData.target : null));\n }\n );\n }\n }\n\n /**\n * Returns the set of documents that have been updated since the last call.\n * If this is the first call, returns the set of changes since client\n * initialization. Further invocations will return document changes since\n * the point of rejection.\n */\n getNewDocumentChanges(): Promise<MaybeDocumentMap> {\n return this.persistence\n .runTransaction('Get new document changes', 'readonly', txn =>\n this.remoteDocuments.getNewDocumentChanges(\n txn,\n this.lastDocumentChangeReadTime\n )\n )\n .then(({ changedDocs, readTime }) => {\n this.lastDocumentChangeReadTime = readTime;\n return changedDocs;\n });\n }\n\n /**\n * Reads the newest document change from persistence and forwards the internal\n * synchronization marker so that calls to `getNewDocumentChanges()`\n * only return changes that happened after client initialization.\n */\n async synchronizeLastDocumentChangeReadTime(): Promise<void> {\n this.lastDocumentChangeReadTime = await this.persistence.runTransaction(\n 'Synchronize last document change read time',\n 'readonly',\n txn => this.remoteDocuments.getLastReadTime(txn)\n );\n }\n}\n\n/**\n * Verifies the error thrown by a LocalStore operation. If a LocalStore\n * operation fails because the primary lease has been taken by another client,\n * we ignore the error (the persistence layer will immediately call\n * `applyPrimaryLease` to propagate the primary state change). All other errors\n * are re-thrown.\n *\n * @param err An error returned by a LocalStore operation.\n * @return A Promise that resolves after we recovered, or the original error.\n */\nexport async function ignoreIfPrimaryLeaseLoss(\n err: FirestoreError\n): Promise<void> {\n if (\n err.code === Code.FAILED_PRECONDITION &&\n err.message === PRIMARY_LEASE_LOST_ERROR_MSG\n ) {\n logDebug(LOG_TAG, 'Unexpectedly lost primary lease');\n } else {\n throw err;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { User } from '../auth/user';\nimport { ListenSequenceNumber, TargetId } from '../core/types';\nimport { DocumentKey } from '../model/document_key';\nimport { IndexManager } from './index_manager';\nimport { LocalStore } from './local_store';\nimport { MutationQueue } from './mutation_queue';\nimport { PersistencePromise } from './persistence_promise';\nimport { TargetCache } from './target_cache';\nimport { RemoteDocumentCache } from './remote_document_cache';\nimport { TargetData } from './target_data';\n\nexport const PRIMARY_LEASE_LOST_ERROR_MSG =\n 'The current tab is not in the required state to perform this operation. ' +\n 'It might be necessary to refresh the browser tab.';\n\n/**\n * A base class representing a persistence transaction, encapsulating both the\n * transaction's sequence numbers as well as a list of onCommitted listeners.\n *\n * When you call Persistence.runTransaction(), it will create a transaction and\n * pass it to your callback. You then pass it to any method that operates\n * on persistence.\n */\nexport abstract class PersistenceTransaction {\n private readonly onCommittedListeners: Array<() => void> = [];\n\n abstract readonly currentSequenceNumber: ListenSequenceNumber;\n\n addOnCommittedListener(listener: () => void): void {\n this.onCommittedListeners.push(listener);\n }\n\n raiseOnCommittedEvent(): void {\n this.onCommittedListeners.forEach(listener => listener());\n }\n}\n\n/** The different modes supported by `IndexedDbPersistence.runTransaction()`. */\nexport type PersistenceTransactionMode =\n | 'readonly'\n | 'readwrite'\n | 'readwrite-primary';\n\n/**\n * Callback type for primary state notifications. This callback can be\n * registered with the persistence layer to get notified when we transition from\n * primary to secondary state and vice versa.\n *\n * Note: Instances can only toggle between Primary and Secondary state if\n * IndexedDB persistence is enabled and multiple clients are active. If this\n * listener is registered with MemoryPersistence, the callback will be called\n * exactly once marking the current instance as Primary.\n */\nexport type PrimaryStateListener = (isPrimary: boolean) => Promise<void>;\n\n/**\n * A ReferenceDelegate instance handles all of the hooks into the document-reference lifecycle. This\n * includes being added to a target, being removed from a target, being subject to mutation, and\n * being mutated by the user.\n *\n * Different implementations may do different things with each of these events. Not every\n * implementation needs to do something with every lifecycle hook.\n *\n * PORTING NOTE: since sequence numbers are attached to transactions in this\n * client, the ReferenceDelegate does not need to deal in transactional\n * semantics (onTransactionStarted/Committed()), nor does it need to track and\n * generate sequence numbers (getCurrentSequenceNumber()).\n */\nexport interface ReferenceDelegate {\n /** Notify the delegate that the given document was added to a target. */\n addReference(\n txn: PersistenceTransaction,\n targetId: TargetId,\n doc: DocumentKey\n ): PersistencePromise<void>;\n\n /** Notify the delegate that the given document was removed from a target. */\n removeReference(\n txn: PersistenceTransaction,\n targetId: TargetId,\n doc: DocumentKey\n ): PersistencePromise<void>;\n\n /**\n * Notify the delegate that a target was removed. The delegate may, but is not obligated to,\n * actually delete the target and associated data.\n */\n removeTarget(\n txn: PersistenceTransaction,\n targetData: TargetData\n ): PersistencePromise<void>;\n\n /**\n * Notify the delegate that a document may no longer be part of any views or\n * have any mutations associated.\n */\n markPotentiallyOrphaned(\n txn: PersistenceTransaction,\n doc: DocumentKey\n ): PersistencePromise<void>;\n\n /** Notify the delegate that a limbo document was updated. */\n updateLimboDocument(\n txn: PersistenceTransaction,\n doc: DocumentKey\n ): PersistencePromise<void>;\n}\n\n/**\n * Persistence is the lowest-level shared interface to persistent storage in\n * Firestore.\n *\n * Persistence is used to create MutationQueue and RemoteDocumentCache\n * instances backed by persistence (which might be in-memory or LevelDB).\n *\n * Persistence also exposes an API to create and run PersistenceTransactions\n * against persistence. All read / write operations must be wrapped in a\n * transaction. Implementations of PersistenceTransaction / Persistence only\n * need to guarantee that writes made against the transaction are not made to\n * durable storage until the transaction resolves its PersistencePromise.\n * Since memory-only storage components do not alter durable storage, they are\n * free to ignore the transaction.\n *\n * This contract is enough to allow the LocalStore be be written\n * independently of whether or not the stored state actually is durably\n * persisted. If persistent storage is enabled, writes are grouped together to\n * avoid inconsistent state that could cause crashes.\n *\n * Concretely, when persistent storage is enabled, the persistent versions of\n * MutationQueue, RemoteDocumentCache, and others (the mutators) will\n * defer their writes into a transaction. Once the local store has completed\n * one logical operation, it commits the transaction.\n *\n * When persistent storage is disabled, the non-persistent versions of the\n * mutators ignore the transaction. This short-cut is allowed because\n * memory-only storage leaves no state so it cannot be inconsistent.\n *\n * This simplifies the implementations of the mutators and allows memory-only\n * implementations to supplement the persistent ones without requiring any\n * special dual-store implementation of Persistence. The cost is that the\n * LocalStore needs to be slightly careful about the order of its reads and\n * writes in order to avoid relying on being able to read back uncommitted\n * writes.\n */\nexport interface Persistence {\n /**\n * Whether or not this persistence instance has been started.\n */\n readonly started: boolean;\n\n readonly referenceDelegate: ReferenceDelegate;\n\n /** Starts persistence. */\n start(): Promise<void>;\n\n /**\n * Releases any resources held during eager shutdown.\n */\n shutdown(): Promise<void>;\n\n /**\n * Registers a listener that gets called when the database receives a\n * version change event indicating that it has deleted.\n *\n * PORTING NOTE: This is only used for Web multi-tab.\n */\n setDatabaseDeletedListener(\n databaseDeletedListener: () => Promise<void>\n ): void;\n\n /**\n * Returns a MutationQueue representing the persisted mutations for the\n * given user.\n *\n * Note: The implementation is free to return the same instance every time\n * this is called for a given user. In particular, the memory-backed\n * implementation does this to emulate the persisted implementation to the\n * extent possible (e.g. in the case of uid switching from\n * sally=>jack=>sally, sally's mutation queue will be preserved).\n */\n getMutationQueue(user: User): MutationQueue;\n\n /**\n * Returns a TargetCache representing the persisted cache of targets.\n *\n * Note: The implementation is free to return the same instance every time\n * this is called. In particular, the memory-backed implementation does this\n * to emulate the persisted implementation to the extent possible.\n */\n getTargetCache(): TargetCache;\n\n /**\n * Returns a RemoteDocumentCache representing the persisted cache of remote\n * documents.\n *\n * Note: The implementation is free to return the same instance every time\n * this is called. In particular, the memory-backed implementation does this\n * to emulate the persisted implementation to the extent possible.\n */\n getRemoteDocumentCache(): RemoteDocumentCache;\n\n /**\n * Returns an IndexManager instance that manages our persisted query indexes.\n *\n * Note: The implementation is free to return the same instance every time\n * this is called. In particular, the memory-backed implementation does this\n * to emulate the persisted implementation to the extent possible.\n */\n getIndexManager(): IndexManager;\n\n /**\n * Performs an operation inside a persistence transaction. Any reads or writes\n * against persistence must be performed within a transaction. Writes will be\n * committed atomically once the transaction completes.\n *\n * Persistence operations are asynchronous and therefore the provided\n * transactionOperation must return a PersistencePromise. When it is resolved,\n * the transaction will be committed and the Promise returned by this method\n * will resolve.\n *\n * @param action A description of the action performed by this transaction,\n * used for logging.\n * @param mode The underlying mode of the IndexedDb transaction. Can be\n * 'readonly`, 'readwrite' or 'readwrite-primary'. Transactions marked\n * 'readwrite-primary' can only be executed by the primary client. In this\n * mode, the transactionOperation will not be run if the primary lease cannot\n * be acquired and the returned promise will be rejected with a\n * FAILED_PRECONDITION error.\n * @param transactionOperation The operation to run inside a transaction.\n * @return A promise that is resolved once the transaction completes.\n */\n runTransaction<T>(\n action: string,\n mode: PersistenceTransactionMode,\n transactionOperation: (\n transaction: PersistenceTransaction\n ) => PersistencePromise<T>\n ): Promise<T>;\n}\n\n/**\n * Interface implemented by the LRU scheduler to start(), stop() and restart\n * garbage collection.\n */\nexport interface GarbageCollectionScheduler {\n readonly started: boolean;\n start(localStore: LocalStore): void;\n stop(): void;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { BatchId, TargetId } from '../core/types';\nimport { documentKeySet, DocumentKeySet } from '../model/collections';\nimport { DocumentKey } from '../model/document_key';\nimport { primitiveComparator } from '../util/misc';\nimport { SortedSet } from '../util/sorted_set';\n\n/**\n * A collection of references to a document from some kind of numbered entity\n * (either a target ID or batch ID). As references are added to or removed from\n * the set corresponding events are emitted to a registered garbage collector.\n *\n * Each reference is represented by a DocumentReference object. Each of them\n * contains enough information to uniquely identify the reference. They are all\n * stored primarily in a set sorted by key. A document is considered garbage if\n * there's no references in that set (this can be efficiently checked thanks to\n * sorting by key).\n *\n * ReferenceSet also keeps a secondary set that contains references sorted by\n * IDs. This one is used to efficiently implement removal of all references by\n * some target ID.\n */\nexport class ReferenceSet {\n // A set of outstanding references to a document sorted by key.\n private refsByKey = new SortedSet(DocReference.compareByKey);\n\n // A set of outstanding references to a document sorted by target id.\n private refsByTarget = new SortedSet(DocReference.compareByTargetId);\n\n /** Returns true if the reference set contains no references. */\n isEmpty(): boolean {\n return this.refsByKey.isEmpty();\n }\n\n /** Adds a reference to the given document key for the given ID. */\n addReference(key: DocumentKey, id: TargetId | BatchId): void {\n const ref = new DocReference(key, id);\n this.refsByKey = this.refsByKey.add(ref);\n this.refsByTarget = this.refsByTarget.add(ref);\n }\n\n /** Add references to the given document keys for the given ID. */\n addReferences(keys: DocumentKeySet, id: TargetId | BatchId): void {\n keys.forEach(key => this.addReference(key, id));\n }\n\n /**\n * Removes a reference to the given document key for the given\n * ID.\n */\n removeReference(key: DocumentKey, id: TargetId | BatchId): void {\n this.removeRef(new DocReference(key, id));\n }\n\n removeReferences(keys: DocumentKeySet, id: TargetId | BatchId): void {\n keys.forEach(key => this.removeReference(key, id));\n }\n\n /**\n * Clears all references with a given ID. Calls removeRef() for each key\n * removed.\n */\n removeReferencesForId(id: TargetId | BatchId): DocumentKey[] {\n const emptyKey = DocumentKey.EMPTY;\n const startRef = new DocReference(emptyKey, id);\n const endRef = new DocReference(emptyKey, id + 1);\n const keys: DocumentKey[] = [];\n this.refsByTarget.forEachInRange([startRef, endRef], ref => {\n this.removeRef(ref);\n keys.push(ref.key);\n });\n return keys;\n }\n\n removeAllReferences(): void {\n this.refsByKey.forEach(ref => this.removeRef(ref));\n }\n\n private removeRef(ref: DocReference): void {\n this.refsByKey = this.refsByKey.delete(ref);\n this.refsByTarget = this.refsByTarget.delete(ref);\n }\n\n referencesForId(id: TargetId | BatchId): DocumentKeySet {\n const emptyKey = DocumentKey.EMPTY;\n const startRef = new DocReference(emptyKey, id);\n const endRef = new DocReference(emptyKey, id + 1);\n let keys = documentKeySet();\n this.refsByTarget.forEachInRange([startRef, endRef], ref => {\n keys = keys.add(ref.key);\n });\n return keys;\n }\n\n containsKey(key: DocumentKey): boolean {\n const ref = new DocReference(key, 0);\n const firstRef = this.refsByKey.firstAfterOrEqual(ref);\n return firstRef !== null && key.isEqual(firstRef.key);\n }\n}\n\nexport class DocReference {\n constructor(\n public key: DocumentKey,\n public targetOrBatchId: TargetId | BatchId\n ) {}\n\n /** Compare by key then by ID */\n static compareByKey(left: DocReference, right: DocReference): number {\n return (\n DocumentKey.comparator(left.key, right.key) ||\n primitiveComparator(left.targetOrBatchId, right.targetOrBatchId)\n );\n }\n\n /** Compare by ID then by key */\n static compareByTargetId(left: DocReference, right: DocReference): number {\n return (\n primitiveComparator(left.targetOrBatchId, right.targetOrBatchId) ||\n DocumentKey.comparator(left.key, right.key)\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\nimport { fail } from './assert';\nimport { Code, FirestoreError } from './error';\nimport { Dict, forEach } from './obj';\n\n/** Types accepted by validateType() and related methods for validation. */\nexport type ValidationType =\n | 'undefined'\n | 'object'\n | 'function'\n | 'boolean'\n | 'number'\n | 'string'\n | 'non-empty string';\n\n/**\n * Validates that no arguments were passed in the invocation of functionName.\n *\n * Forward the magic \"arguments\" variable as second parameter on which the\n * parameter validation is performed:\n * validateNoArgs('myFunction', arguments);\n */\nexport function validateNoArgs(functionName: string, args: IArguments): void {\n if (args.length !== 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() does not support arguments, ` +\n 'but was called with ' +\n formatPlural(args.length, 'argument') +\n '.'\n );\n }\n}\n\n/**\n * Validates the invocation of functionName has the exact number of arguments.\n *\n * Forward the magic \"arguments\" variable as second parameter on which the\n * parameter validation is performed:\n * validateExactNumberOfArgs('myFunction', arguments, 2);\n */\nexport function validateExactNumberOfArgs(\n functionName: string,\n args: IArguments,\n numberOfArgs: number\n): void {\n if (args.length !== numberOfArgs) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires ` +\n formatPlural(numberOfArgs, 'argument') +\n ', but was called with ' +\n formatPlural(args.length, 'argument') +\n '.'\n );\n }\n}\n\n/**\n * Validates the invocation of functionName has at least the provided number of\n * arguments (but can have many more).\n *\n * Forward the magic \"arguments\" variable as second parameter on which the\n * parameter validation is performed:\n * validateAtLeastNumberOfArgs('myFunction', arguments, 2);\n */\nexport function validateAtLeastNumberOfArgs(\n functionName: string,\n args: IArguments,\n minNumberOfArgs: number\n): void {\n if (args.length < minNumberOfArgs) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires at least ` +\n formatPlural(minNumberOfArgs, 'argument') +\n ', but was called with ' +\n formatPlural(args.length, 'argument') +\n '.'\n );\n }\n}\n\n/**\n * Validates the invocation of functionName has number of arguments between\n * the values provided.\n *\n * Forward the magic \"arguments\" variable as second parameter on which the\n * parameter validation is performed:\n * validateBetweenNumberOfArgs('myFunction', arguments, 2, 3);\n */\nexport function validateBetweenNumberOfArgs(\n functionName: string,\n args: IArguments,\n minNumberOfArgs: number,\n maxNumberOfArgs: number\n): void {\n if (args.length < minNumberOfArgs || args.length > maxNumberOfArgs) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires between ${minNumberOfArgs} and ` +\n `${maxNumberOfArgs} arguments, but was called with ` +\n formatPlural(args.length, 'argument') +\n '.'\n );\n }\n}\n\n/**\n * Validates the provided argument is an array and has as least the expected\n * number of elements.\n */\nexport function validateNamedArrayAtLeastNumberOfElements<T>(\n functionName: string,\n value: T[],\n name: string,\n minNumberOfElements: number\n): void {\n if (!(value instanceof Array) || value.length < minNumberOfElements) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires its ${name} argument to be an ` +\n 'array with at least ' +\n `${formatPlural(minNumberOfElements, 'element')}.`\n );\n }\n}\n\n/**\n * Validates the provided positional argument has the native JavaScript type\n * using typeof checks.\n */\nexport function validateArgType(\n functionName: string,\n type: ValidationType,\n position: number,\n argument: unknown\n): void {\n validateType(functionName, type, `${ordinal(position)} argument`, argument);\n}\n\n/**\n * Validates the provided argument has the native JavaScript type using\n * typeof checks or is undefined.\n */\nexport function validateOptionalArgType(\n functionName: string,\n type: ValidationType,\n position: number,\n argument: unknown\n): void {\n if (argument !== undefined) {\n validateArgType(functionName, type, position, argument);\n }\n}\n\n/**\n * Validates the provided named option has the native JavaScript type using\n * typeof checks.\n */\nexport function validateNamedType(\n functionName: string,\n type: ValidationType,\n optionName: string,\n argument: unknown\n): void {\n validateType(functionName, type, `${optionName} option`, argument);\n}\n\n/**\n * Validates the provided named option has the native JavaScript type using\n * typeof checks or is undefined.\n */\nexport function validateNamedOptionalType(\n functionName: string,\n type: ValidationType,\n optionName: string,\n argument: unknown\n): void {\n if (argument !== undefined) {\n validateNamedType(functionName, type, optionName, argument);\n }\n}\n\nexport function validateArrayElements<T>(\n functionName: string,\n optionName: string,\n typeDescription: string,\n argument: T[],\n validator: (arg0: T) => boolean\n): void {\n if (!(argument instanceof Array)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires its ${optionName} ` +\n `option to be an array, but it was: ${valueDescription(argument)}`\n );\n }\n\n for (let i = 0; i < argument.length; ++i) {\n if (!validator(argument[i])) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires all ${optionName} ` +\n `elements to be ${typeDescription}, but the value at index ${i} ` +\n `was: ${valueDescription(argument[i])}`\n );\n }\n }\n}\n\nexport function validateOptionalArrayElements<T>(\n functionName: string,\n optionName: string,\n typeDescription: string,\n argument: T[] | undefined,\n validator: (arg0: T) => boolean\n): void {\n if (argument !== undefined) {\n validateArrayElements(\n functionName,\n optionName,\n typeDescription,\n argument,\n validator\n );\n }\n}\n\n/**\n * Validates that the provided named option equals one of the expected values.\n */\nexport function validateNamedPropertyEquals<T>(\n functionName: string,\n inputName: string,\n optionName: string,\n input: T,\n expected: T[]\n): void {\n const expectedDescription: string[] = [];\n\n for (const val of expected) {\n if (val === input) {\n return;\n }\n expectedDescription.push(valueDescription(val));\n }\n\n const actualDescription = valueDescription(input);\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid value ${actualDescription} provided to function ${functionName}() for option ` +\n `\"${optionName}\". Acceptable values: ${expectedDescription.join(', ')}`\n );\n}\n\n/**\n * Validates that the provided named option equals one of the expected values or\n * is undefined.\n */\nexport function validateNamedOptionalPropertyEquals<T>(\n functionName: string,\n inputName: string,\n optionName: string,\n input: T,\n expected: T[]\n): void {\n if (input !== undefined) {\n validateNamedPropertyEquals(\n functionName,\n inputName,\n optionName,\n input,\n expected\n );\n }\n}\n\n/**\n * Validates that the provided argument is a valid enum.\n *\n * @param functionName Function making the validation call.\n * @param enums Array containing all possible values for the enum.\n * @param position Position of the argument in `functionName`.\n * @param argument Argument to validate.\n * @return The value as T if the argument can be converted.\n */\nexport function validateStringEnum<T>(\n functionName: string,\n enums: T[],\n position: number,\n argument: unknown\n): T {\n if (!enums.some(element => element === argument)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid value ${valueDescription(argument)} provided to function ` +\n `${functionName}() for its ${ordinal(position)} argument. Acceptable ` +\n `values: ${enums.join(', ')}`\n );\n }\n return argument as T;\n}\n\n/** Helper to validate the type of a provided input. */\nfunction validateType(\n functionName: string,\n type: ValidationType,\n inputName: string,\n input: unknown\n): void {\n let valid = false;\n if (type === 'object') {\n valid = isPlainObject(input);\n } else if (type === 'non-empty string') {\n valid = typeof input === 'string' && input !== '';\n } else {\n valid = typeof input === type;\n }\n\n if (!valid) {\n const description = valueDescription(input);\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires its ${inputName} ` +\n `to be of type ${type}, but it was: ${description}`\n );\n }\n}\n\n/**\n * Returns true if it's a non-null object without a custom prototype\n * (i.e. excludes Array, Date, etc.).\n */\nexport function isPlainObject(input: unknown): boolean {\n return (\n typeof input === 'object' &&\n input !== null &&\n (Object.getPrototypeOf(input) === Object.prototype ||\n Object.getPrototypeOf(input) === null)\n );\n}\n\n/** Returns a string describing the type / value of the provided input. */\nexport function valueDescription(input: unknown): string {\n if (input === undefined) {\n return 'undefined';\n } else if (input === null) {\n return 'null';\n } else if (typeof input === 'string') {\n if (input.length > 20) {\n input = `${input.substring(0, 20)}...`;\n }\n return JSON.stringify(input);\n } else if (typeof input === 'number' || typeof input === 'boolean') {\n return '' + input;\n } else if (typeof input === 'object') {\n if (input instanceof Array) {\n return 'an array';\n } else {\n const customObjectName = tryGetCustomObjectType(input!);\n if (customObjectName) {\n return `a custom ${customObjectName} object`;\n } else {\n return 'an object';\n }\n }\n } else if (typeof input === 'function') {\n return 'a function';\n } else {\n return fail('Unknown wrong type: ' + typeof input);\n }\n}\n\n/** Hacky method to try to get the constructor name for an object. */\nexport function tryGetCustomObjectType(input: object): string | null {\n if (input.constructor) {\n const funcNameRegex = /function\\s+([^\\s(]+)\\s*\\(/;\n const results = funcNameRegex.exec(input.constructor.toString());\n if (results && results.length > 1) {\n return results[1];\n }\n }\n return null;\n}\n\n/** Validates the provided argument is defined. */\nexport function validateDefined(\n functionName: string,\n position: number,\n argument: unknown\n): void {\n if (argument === undefined) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires a valid ${ordinal(position)} ` +\n `argument, but it was undefined.`\n );\n }\n}\n\n/**\n * Validates the provided positional argument is an object, and its keys and\n * values match the expected keys and types provided in optionTypes.\n */\nexport function validateOptionNames(\n functionName: string,\n options: object,\n optionNames: string[]\n): void {\n forEach(options as Dict<unknown>, (key, _) => {\n if (optionNames.indexOf(key) < 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Unknown option '${key}' passed to function ${functionName}(). ` +\n 'Available options: ' +\n optionNames.join(', ')\n );\n }\n });\n}\n\n/**\n * Helper method to throw an error that the provided argument did not pass\n * an instanceof check.\n */\nexport function invalidClassError(\n functionName: string,\n type: string,\n position: number,\n argument: unknown\n): Error {\n const description = valueDescription(argument);\n return new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires its ${ordinal(position)} ` +\n `argument to be a ${type}, but it was: ${description}`\n );\n}\n\nexport function validatePositiveNumber(\n functionName: string,\n position: number,\n n: number\n): void {\n if (n <= 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${functionName}() requires its ${ordinal(\n position\n )} argument to be a positive number, but it was: ${n}.`\n );\n }\n}\n\n/** Converts a number to its english word representation */\nfunction ordinal(num: number): string {\n switch (num) {\n case 1:\n return 'first';\n case 2:\n return 'second';\n case 3:\n return 'third';\n default:\n return num + 'th';\n }\n}\n\n/**\n * Formats the given word as plural conditionally given the preceding number.\n */\nfunction formatPlural(num: number, str: string): string {\n return `${num} ${str}` + (num === 1 ? '' : 's');\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { PlatformSupport } from '../platform/platform';\nimport { Code, FirestoreError } from '../util/error';\nimport {\n invalidClassError,\n validateArgType,\n validateExactNumberOfArgs\n} from '../util/input_validation';\nimport { ByteString } from '../util/byte_string';\n\n/** Helper function to assert Uint8Array is available at runtime. */\nfunction assertUint8ArrayAvailable(): void {\n if (typeof Uint8Array === 'undefined') {\n throw new FirestoreError(\n Code.UNIMPLEMENTED,\n 'Uint8Arrays are not available in this environment.'\n );\n }\n}\n\n/** Helper function to assert Base64 functions are available at runtime. */\nfunction assertBase64Available(): void {\n if (!PlatformSupport.getPlatform().base64Available) {\n throw new FirestoreError(\n Code.UNIMPLEMENTED,\n 'Blobs are unavailable in Firestore in this environment.'\n );\n }\n}\n\n/**\n * Immutable class holding a blob (binary data).\n * This class is directly exposed in the public API.\n *\n * Note that while you can't hide the constructor in JavaScript code, we are\n * using the hack above to make sure no-one outside this module can call it.\n */\nexport class Blob {\n // Prefix with underscore to signal that we consider this not part of the\n // public API and to prevent it from showing up for autocompletion.\n _byteString: ByteString;\n\n constructor(byteString: ByteString) {\n assertBase64Available();\n this._byteString = byteString;\n }\n\n static fromBase64String(base64: string): Blob {\n validateExactNumberOfArgs('Blob.fromBase64String', arguments, 1);\n validateArgType('Blob.fromBase64String', 'string', 1, base64);\n assertBase64Available();\n try {\n return new Blob(ByteString.fromBase64String(base64));\n } catch (e) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Failed to construct Blob from Base64 string: ' + e\n );\n }\n }\n\n static fromUint8Array(array: Uint8Array): Blob {\n validateExactNumberOfArgs('Blob.fromUint8Array', arguments, 1);\n assertUint8ArrayAvailable();\n if (!(array instanceof Uint8Array)) {\n throw invalidClassError('Blob.fromUint8Array', 'Uint8Array', 1, array);\n }\n return new Blob(ByteString.fromUint8Array(array));\n }\n\n toBase64(): string {\n validateExactNumberOfArgs('Blob.toBase64', arguments, 0);\n assertBase64Available();\n return this._byteString.toBase64();\n }\n\n toUint8Array(): Uint8Array {\n validateExactNumberOfArgs('Blob.toUint8Array', arguments, 0);\n assertUint8ArrayAvailable();\n return this._byteString.toUint8Array();\n }\n\n toString(): string {\n return 'Blob(base64: ' + this.toBase64() + ')';\n }\n\n isEqual(other: Blob): boolean {\n return this._byteString.isEqual(other._byteString);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as firestore from '@firebase/firestore-types';\n\nimport { FieldPath as InternalFieldPath } from '../model/path';\nimport { Code, FirestoreError } from '../util/error';\nimport {\n invalidClassError,\n validateArgType,\n validateNamedArrayAtLeastNumberOfElements\n} from '../util/input_validation';\n\n// The objects that are a part of this API are exposed to third-parties as\n// compiled javascript so we want to flag our private members with a leading\n// underscore to discourage their use.\n\n/**\n * A FieldPath refers to a field in a document. The path may consist of a single\n * field name (referring to a top-level field in the document), or a list of\n * field names (referring to a nested field in the document).\n */\nexport class FieldPath implements firestore.FieldPath {\n /** Internal representation of a Firestore field path. */\n _internalPath: InternalFieldPath;\n\n /**\n * Creates a FieldPath from the provided field names. If more than one field\n * name is provided, the path will point to a nested field in a document.\n *\n * @param fieldNames A list of field names.\n */\n constructor(...fieldNames: string[]) {\n validateNamedArrayAtLeastNumberOfElements(\n 'FieldPath',\n fieldNames,\n 'fieldNames',\n 1\n );\n\n for (let i = 0; i < fieldNames.length; ++i) {\n validateArgType('FieldPath', 'string', i, fieldNames[i]);\n if (fieldNames[i].length === 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid field name at argument $(i + 1). ` +\n 'Field names must not be empty.'\n );\n }\n }\n\n this._internalPath = new InternalFieldPath(fieldNames);\n }\n\n /**\n * Internal Note: The backend doesn't technically support querying by\n * document ID. Instead it queries by the entire document name (full path\n * included), but in the cases we currently support documentId(), the net\n * effect is the same.\n */\n private static readonly _DOCUMENT_ID = new FieldPath(\n InternalFieldPath.keyField().canonicalString()\n );\n\n static documentId(): FieldPath {\n return FieldPath._DOCUMENT_ID;\n }\n\n isEqual(other: firestore.FieldPath): boolean {\n if (!(other instanceof FieldPath)) {\n throw invalidClassError('isEqual', 'FieldPath', 1, other);\n }\n return this._internalPath.isEqual(other._internalPath);\n }\n}\n\n/**\n * Matches any characters in a field path string that are reserved.\n */\nconst RESERVED = new RegExp('[~\\\\*/\\\\[\\\\]]');\n\n/**\n * Parses a field path string into a FieldPath, treating dots as separators.\n */\nexport function fromDotSeparatedString(path: string): FieldPath {\n const found = path.search(RESERVED);\n if (found >= 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid field path (${path}). Paths must not contain ` +\n `'~', '*', '/', '[', or ']'`\n );\n }\n try {\n return new FieldPath(...path.split('.'));\n } catch (e) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid field path (${path}). Paths must not be empty, ` +\n `begin with '.', end with '.', or contain '..'`\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as firestore from '@firebase/firestore-types';\nimport {\n validateArgType,\n validateAtLeastNumberOfArgs,\n validateExactNumberOfArgs,\n validateNoArgs\n} from '../util/input_validation';\nimport { FieldTransform } from '../model/mutation';\nimport {\n ArrayRemoveTransformOperation,\n ArrayUnionTransformOperation,\n NumericIncrementTransformOperation,\n ServerTimestampTransform\n} from '../model/transform_operation';\nimport { ParseContext, parseData, UserDataSource } from './user_data_reader';\nimport { debugAssert } from '../util/assert';\n\n/**\n * An opaque base class for FieldValue sentinel objects in our public API,\n * with public static methods for creating said sentinel objects.\n */\nexport abstract class FieldValueImpl {\n protected constructor(readonly _methodName: string) {}\n\n abstract toFieldTransform(context: ParseContext): FieldTransform | null;\n\n abstract isEqual(other: FieldValue): boolean;\n}\n\nexport class DeleteFieldValueImpl extends FieldValueImpl {\n constructor() {\n super('FieldValue.delete');\n }\n\n toFieldTransform(context: ParseContext): null {\n if (context.dataSource === UserDataSource.MergeSet) {\n // No transform to add for a delete, but we need to add it to our\n // fieldMask so it gets deleted.\n context.fieldMask.push(context.path!);\n } else if (context.dataSource === UserDataSource.Update) {\n debugAssert(\n context.path!.length > 0,\n 'FieldValue.delete() at the top level should have already' +\n ' been handled.'\n );\n throw context.createError(\n 'FieldValue.delete() can only appear at the top level ' +\n 'of your update data'\n );\n } else {\n // We shouldn't encounter delete sentinels for queries or non-merge set() calls.\n throw context.createError(\n 'FieldValue.delete() cannot be used with set() unless you pass ' +\n '{merge:true}'\n );\n }\n return null;\n }\n\n isEqual(other: FieldValue): boolean {\n return other instanceof DeleteFieldValueImpl;\n }\n}\n\nexport class ServerTimestampFieldValueImpl extends FieldValueImpl {\n constructor() {\n super('FieldValue.serverTimestamp');\n }\n\n toFieldTransform(context: ParseContext): FieldTransform {\n return new FieldTransform(context.path!, ServerTimestampTransform.instance);\n }\n\n isEqual(other: FieldValue): boolean {\n return other instanceof ServerTimestampFieldValueImpl;\n }\n}\n\nexport class ArrayUnionFieldValueImpl extends FieldValueImpl {\n constructor(private readonly _elements: unknown[]) {\n super('FieldValue.arrayUnion');\n }\n\n toFieldTransform(context: ParseContext): FieldTransform {\n // Although array transforms are used with writes, the actual elements\n // being uniomed or removed are not considered writes since they cannot\n // contain any FieldValue sentinels, etc.\n const parseContext = new ParseContext(\n {\n dataSource: UserDataSource.Argument,\n methodName: this._methodName,\n arrayElement: true\n },\n context.databaseId,\n context.serializer,\n context.ignoreUndefinedProperties\n );\n const parsedElements = this._elements.map(\n element => parseData(element, parseContext)!\n );\n const arrayUnion = new ArrayUnionTransformOperation(parsedElements);\n return new FieldTransform(context.path!, arrayUnion);\n }\n\n isEqual(other: FieldValue): boolean {\n // TODO(mrschmidt): Implement isEquals\n return this === other;\n }\n}\n\nexport class ArrayRemoveFieldValueImpl extends FieldValueImpl {\n constructor(readonly _elements: unknown[]) {\n super('FieldValue.arrayRemove');\n }\n\n toFieldTransform(context: ParseContext): FieldTransform {\n // Although array transforms are used with writes, the actual elements\n // being unioned or removed are not considered writes since they cannot\n // contain any FieldValue sentinels, etc.\n const parseContext = new ParseContext(\n {\n dataSource: UserDataSource.Argument,\n methodName: this._methodName,\n arrayElement: true\n },\n context.databaseId,\n context.serializer,\n context.ignoreUndefinedProperties\n );\n const parsedElements = this._elements.map(\n element => parseData(element, parseContext)!\n );\n const arrayUnion = new ArrayRemoveTransformOperation(parsedElements);\n return new FieldTransform(context.path!, arrayUnion);\n }\n\n isEqual(other: FieldValue): boolean {\n // TODO(mrschmidt): Implement isEquals\n return this === other;\n }\n}\n\nexport class NumericIncrementFieldValueImpl extends FieldValueImpl {\n constructor(private readonly _operand: number) {\n super('FieldValue.increment');\n }\n\n toFieldTransform(context: ParseContext): FieldTransform {\n const parseContext = new ParseContext(\n {\n dataSource: UserDataSource.Argument,\n methodName: this._methodName\n },\n context.databaseId,\n context.serializer,\n context.ignoreUndefinedProperties\n );\n const operand = parseData(this._operand, parseContext)!;\n const numericIncrement = new NumericIncrementTransformOperation(\n context.serializer,\n operand\n );\n return new FieldTransform(context.path!, numericIncrement);\n }\n\n isEqual(other: FieldValue): boolean {\n // TODO(mrschmidt): Implement isEquals\n return this === other;\n }\n}\n\nexport class FieldValue implements firestore.FieldValue {\n static delete(): FieldValueImpl {\n validateNoArgs('FieldValue.delete', arguments);\n return new DeleteFieldValueImpl();\n }\n\n static serverTimestamp(): FieldValueImpl {\n validateNoArgs('FieldValue.serverTimestamp', arguments);\n return new ServerTimestampFieldValueImpl();\n }\n\n static arrayUnion(...elements: unknown[]): FieldValueImpl {\n validateAtLeastNumberOfArgs('FieldValue.arrayUnion', arguments, 1);\n // NOTE: We don't actually parse the data until it's used in set() or\n // update() since we need access to the Firestore instance.\n return new ArrayUnionFieldValueImpl(elements);\n }\n\n static arrayRemove(...elements: unknown[]): FieldValueImpl {\n validateAtLeastNumberOfArgs('FieldValue.arrayRemove', arguments, 1);\n // NOTE: We don't actually parse the data until it's used in set() or\n // update() since we need access to the Firestore instance.\n return new ArrayRemoveFieldValueImpl(elements);\n }\n\n static increment(n: number): FieldValueImpl {\n validateArgType('FieldValue.increment', 'number', 1, n);\n validateExactNumberOfArgs('FieldValue.increment', arguments, 1);\n return new NumericIncrementFieldValueImpl(n);\n }\n\n isEqual(other: FieldValue): boolean {\n return this === other;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Code, FirestoreError } from '../util/error';\nimport {\n validateArgType,\n validateExactNumberOfArgs\n} from '../util/input_validation';\nimport { primitiveComparator } from '../util/misc';\n\n/**\n * Immutable class representing a geo point as latitude-longitude pair.\n * This class is directly exposed in the public API, including its constructor.\n */\nexport class GeoPoint {\n // Prefix with underscore to signal this is a private variable in JS and\n // prevent it showing up for autocompletion when typing latitude or longitude.\n private _lat: number;\n private _long: number;\n\n constructor(latitude: number, longitude: number) {\n validateExactNumberOfArgs('GeoPoint', arguments, 2);\n validateArgType('GeoPoint', 'number', 1, latitude);\n validateArgType('GeoPoint', 'number', 2, longitude);\n if (!isFinite(latitude) || latitude < -90 || latitude > 90) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Latitude must be a number between -90 and 90, but was: ' + latitude\n );\n }\n if (!isFinite(longitude) || longitude < -180 || longitude > 180) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Longitude must be a number between -180 and 180, but was: ' + longitude\n );\n }\n\n this._lat = latitude;\n this._long = longitude;\n }\n\n /**\n * Returns the latitude of this geo point, a number between -90 and 90.\n */\n get latitude(): number {\n return this._lat;\n }\n\n /**\n * Returns the longitude of this geo point, a number between -180 and 180.\n */\n get longitude(): number {\n return this._long;\n }\n\n isEqual(other: GeoPoint): boolean {\n return this._lat === other._lat && this._long === other._long;\n }\n\n /**\n * Actually private to JS consumers of our API, so this function is prefixed\n * with an underscore.\n */\n _compareTo(other: GeoPoint): number {\n return (\n primitiveComparator(this._lat, other._lat) ||\n primitiveComparator(this._long, other._long)\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as firestore from '@firebase/firestore-types';\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { Timestamp } from './timestamp';\nimport { DatabaseId } from '../core/database_info';\nimport { DocumentKey } from '../model/document_key';\nimport {\n FieldMask,\n FieldTransform,\n Mutation,\n PatchMutation,\n Precondition,\n SetMutation,\n TransformMutation\n} from '../model/mutation';\nimport { FieldPath } from '../model/path';\nimport { debugAssert, fail } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { isPlainObject, valueDescription } from '../util/input_validation';\nimport { Dict, forEach, isEmpty } from '../util/obj';\nimport { ObjectValue, ObjectValueBuilder } from '../model/object_value';\nimport { JsonProtoSerializer } from '../remote/serializer';\nimport { Blob } from './blob';\nimport {\n FieldPath as ExternalFieldPath,\n fromDotSeparatedString\n} from './field_path';\nimport { DeleteFieldValueImpl, FieldValueImpl } from './field_value';\nimport { GeoPoint } from './geo_point';\nimport { PlatformSupport } from '../platform/platform';\nimport { DocumentReference } from './database';\n\nconst RESERVED_FIELD_REGEX = /^__.*__$/;\n\n/** The result of parsing document data (e.g. for a setData call). */\nexport class ParsedSetData {\n constructor(\n readonly data: ObjectValue,\n readonly fieldMask: FieldMask | null,\n readonly fieldTransforms: FieldTransform[]\n ) {}\n\n toMutations(key: DocumentKey, precondition: Precondition): Mutation[] {\n const mutations = [] as Mutation[];\n if (this.fieldMask !== null) {\n mutations.push(\n new PatchMutation(key, this.data, this.fieldMask, precondition)\n );\n } else {\n mutations.push(new SetMutation(key, this.data, precondition));\n }\n if (this.fieldTransforms.length > 0) {\n mutations.push(new TransformMutation(key, this.fieldTransforms));\n }\n return mutations;\n }\n}\n\n/** The result of parsing \"update\" data (i.e. for an updateData call). */\nexport class ParsedUpdateData {\n constructor(\n readonly data: ObjectValue,\n readonly fieldMask: FieldMask,\n readonly fieldTransforms: FieldTransform[]\n ) {}\n\n toMutations(key: DocumentKey, precondition: Precondition): Mutation[] {\n const mutations = [\n new PatchMutation(key, this.data, this.fieldMask, precondition)\n ] as Mutation[];\n if (this.fieldTransforms.length > 0) {\n mutations.push(new TransformMutation(key, this.fieldTransforms));\n }\n return mutations;\n }\n}\n\n/*\n * Represents what type of API method provided the data being parsed; useful\n * for determining which error conditions apply during parsing and providing\n * better error messages.\n */\nexport const enum UserDataSource {\n Set,\n Update,\n MergeSet,\n /**\n * Indicates the source is a where clause, cursor bound, arrayUnion()\n * element, etc. Of note, isWrite(source) will return false.\n */\n Argument,\n /**\n * Indicates that the source is an Argument that may directly contain nested\n * arrays (e.g. the operand of an `in` query).\n */\n ArrayArgument\n}\n\nfunction isWrite(dataSource: UserDataSource): boolean {\n switch (dataSource) {\n case UserDataSource.Set: // fall through\n case UserDataSource.MergeSet: // fall through\n case UserDataSource.Update:\n return true;\n case UserDataSource.Argument:\n case UserDataSource.ArrayArgument:\n return false;\n default:\n throw fail(`Unexpected case for UserDataSource: ${dataSource}`);\n }\n}\n\n/** Contains the settings that are mutated as we parse user data. */\ninterface ContextSettings {\n /** Indicates what kind of API method this data came from. */\n readonly dataSource: UserDataSource;\n /** The name of the method the user called to create the ParseContext. */\n readonly methodName: string;\n /**\n * A path within the object being parsed. This could be an empty path (in\n * which case the context represents the root of the data being parsed), or a\n * nonempty path (indicating the context represents a nested location within\n * the data).\n */\n readonly path?: FieldPath;\n /**\n * Whether or not this context corresponds to an element of an array.\n * If not set, elements are treated as if they were outside of arrays.\n */\n readonly arrayElement?: boolean;\n}\n\n/** A \"context\" object passed around while parsing user data. */\nexport class ParseContext {\n readonly fieldTransforms: FieldTransform[];\n readonly fieldMask: FieldPath[];\n /**\n * Initializes a ParseContext with the given source and path.\n *\n * @param settings The settings for the parser.\n * @param databaseId The database ID of the Firestore instance.\n * @param serializer The serializer to use to generate the Value proto.\n * @param ignoreUndefinedProperties Whether to ignore undefined properties\n * rather than throw.\n * @param fieldTransforms A mutable list of field transforms encountered while\n * parsing the data.\n * @param fieldMask A mutable list of field paths encountered while parsing\n * the data.\n *\n * TODO(b/34871131): We don't support array paths right now, so path can be\n * null to indicate the context represents any location within an array (in\n * which case certain features will not work and errors will be somewhat\n * compromised).\n */\n constructor(\n readonly settings: ContextSettings,\n readonly databaseId: DatabaseId,\n readonly serializer: JsonProtoSerializer,\n readonly ignoreUndefinedProperties: boolean,\n fieldTransforms?: FieldTransform[],\n fieldMask?: FieldPath[]\n ) {\n // Minor hack: If fieldTransforms is undefined, we assume this is an\n // external call and we need to validate the entire path.\n if (fieldTransforms === undefined) {\n this.validatePath();\n }\n this.fieldTransforms = fieldTransforms || [];\n this.fieldMask = fieldMask || [];\n }\n\n get path(): FieldPath | undefined {\n return this.settings.path;\n }\n\n get dataSource(): UserDataSource {\n return this.settings.dataSource;\n }\n\n /** Returns a new context with the specified settings overwritten. */\n contextWith(configuration: Partial<ContextSettings>): ParseContext {\n return new ParseContext(\n { ...this.settings, ...configuration },\n this.databaseId,\n this.serializer,\n this.ignoreUndefinedProperties,\n this.fieldTransforms,\n this.fieldMask\n );\n }\n\n childContextForField(field: string): ParseContext {\n const childPath = this.path?.child(field);\n const context = this.contextWith({ path: childPath, arrayElement: false });\n context.validatePathSegment(field);\n return context;\n }\n\n childContextForFieldPath(field: FieldPath): ParseContext {\n const childPath = this.path?.child(field);\n const context = this.contextWith({ path: childPath, arrayElement: false });\n context.validatePath();\n return context;\n }\n\n childContextForArray(index: number): ParseContext {\n // TODO(b/34871131): We don't support array paths right now; so make path\n // undefined.\n return this.contextWith({ path: undefined, arrayElement: true });\n }\n\n createError(reason: string): Error {\n const fieldDescription =\n !this.path || this.path.isEmpty()\n ? ''\n : ` (found in field ${this.path.toString()})`;\n return new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${this.settings.methodName}() called with invalid data. ` +\n reason +\n fieldDescription\n );\n }\n\n /** Returns 'true' if 'fieldPath' was traversed when creating this context. */\n contains(fieldPath: FieldPath): boolean {\n return (\n this.fieldMask.find(field => fieldPath.isPrefixOf(field)) !== undefined ||\n this.fieldTransforms.find(transform =>\n fieldPath.isPrefixOf(transform.field)\n ) !== undefined\n );\n }\n\n private validatePath(): void {\n // TODO(b/34871131): Remove null check once we have proper paths for fields\n // within arrays.\n if (!this.path) {\n return;\n }\n for (let i = 0; i < this.path.length; i++) {\n this.validatePathSegment(this.path.get(i));\n }\n }\n\n private validatePathSegment(segment: string): void {\n if (segment.length === 0) {\n throw this.createError('Document fields must not be empty');\n }\n if (isWrite(this.dataSource) && RESERVED_FIELD_REGEX.test(segment)) {\n throw this.createError('Document fields cannot begin and end with \"__\"');\n }\n }\n}\n\n/**\n * Helper for parsing raw user input (provided via the API) into internal model\n * classes.\n */\nexport class UserDataReader {\n private readonly serializer: JsonProtoSerializer;\n\n constructor(\n private readonly databaseId: DatabaseId,\n private readonly ignoreUndefinedProperties: boolean,\n serializer?: JsonProtoSerializer\n ) {\n this.serializer =\n serializer || PlatformSupport.getPlatform().newSerializer(databaseId);\n }\n\n /** Parse document data from a non-merge set() call. */\n parseSetData(methodName: string, input: unknown): ParsedSetData {\n const context = this.createContext(UserDataSource.Set, methodName);\n validatePlainObject('Data must be an object, but it was:', context, input);\n const updateData = parseObject(input, context)!;\n\n return new ParsedSetData(\n new ObjectValue(updateData),\n /* fieldMask= */ null,\n context.fieldTransforms\n );\n }\n\n /** Parse document data from a set() call with '{merge:true}'. */\n parseMergeData(\n methodName: string,\n input: unknown,\n fieldPaths?: Array<string | firestore.FieldPath>\n ): ParsedSetData {\n const context = this.createContext(UserDataSource.MergeSet, methodName);\n validatePlainObject('Data must be an object, but it was:', context, input);\n const updateData = parseObject(input, context);\n\n let fieldMask: FieldMask;\n let fieldTransforms: FieldTransform[];\n\n if (!fieldPaths) {\n fieldMask = new FieldMask(context.fieldMask);\n fieldTransforms = context.fieldTransforms;\n } else {\n const validatedFieldPaths: FieldPath[] = [];\n\n for (const stringOrFieldPath of fieldPaths) {\n let fieldPath: FieldPath;\n\n if (stringOrFieldPath instanceof ExternalFieldPath) {\n fieldPath = stringOrFieldPath._internalPath;\n } else if (typeof stringOrFieldPath === 'string') {\n fieldPath = fieldPathFromDotSeparatedString(\n methodName,\n stringOrFieldPath\n );\n } else {\n throw fail(\n 'Expected stringOrFieldPath to be a string or a FieldPath'\n );\n }\n\n if (!context.contains(fieldPath)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Field '${fieldPath}' is specified in your field mask but missing from your input data.`\n );\n }\n\n if (!fieldMaskContains(validatedFieldPaths, fieldPath)) {\n validatedFieldPaths.push(fieldPath);\n }\n }\n\n fieldMask = new FieldMask(validatedFieldPaths);\n fieldTransforms = context.fieldTransforms.filter(transform =>\n fieldMask.covers(transform.field)\n );\n }\n return new ParsedSetData(\n new ObjectValue(updateData),\n fieldMask,\n fieldTransforms\n );\n }\n\n /** Parse update data from an update() call. */\n parseUpdateData(methodName: string, input: unknown): ParsedUpdateData {\n const context = this.createContext(UserDataSource.Update, methodName);\n validatePlainObject('Data must be an object, but it was:', context, input);\n\n const fieldMaskPaths: FieldPath[] = [];\n const updateData = new ObjectValueBuilder();\n forEach(input as Dict<unknown>, (key, value) => {\n const path = fieldPathFromDotSeparatedString(methodName, key);\n\n const childContext = context.childContextForFieldPath(path);\n if (value instanceof DeleteFieldValueImpl) {\n // Add it to the field mask, but don't add anything to updateData.\n fieldMaskPaths.push(path);\n } else {\n const parsedValue = parseData(value, childContext);\n if (parsedValue != null) {\n fieldMaskPaths.push(path);\n updateData.set(path, parsedValue);\n }\n }\n });\n\n const mask = new FieldMask(fieldMaskPaths);\n return new ParsedUpdateData(\n updateData.build(),\n mask,\n context.fieldTransforms\n );\n }\n\n /** Parse update data from a list of field/value arguments. */\n parseUpdateVarargs(\n methodName: string,\n field: string | ExternalFieldPath,\n value: unknown,\n moreFieldsAndValues: unknown[]\n ): ParsedUpdateData {\n const context = this.createContext(UserDataSource.Update, methodName);\n const keys = [fieldPathFromArgument(methodName, field)];\n const values = [value];\n\n if (moreFieldsAndValues.length % 2 !== 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${methodName}() needs to be called with an even number ` +\n 'of arguments that alternate between field names and values.'\n );\n }\n\n for (let i = 0; i < moreFieldsAndValues.length; i += 2) {\n keys.push(\n fieldPathFromArgument(\n methodName,\n moreFieldsAndValues[i] as string | ExternalFieldPath\n )\n );\n values.push(moreFieldsAndValues[i + 1]);\n }\n\n const fieldMaskPaths: FieldPath[] = [];\n const updateData = new ObjectValueBuilder();\n\n // We iterate in reverse order to pick the last value for a field if the\n // user specified the field multiple times.\n for (let i = keys.length - 1; i >= 0; --i) {\n if (!fieldMaskContains(fieldMaskPaths, keys[i])) {\n const path = keys[i];\n const value = values[i];\n const childContext = context.childContextForFieldPath(path);\n if (value instanceof DeleteFieldValueImpl) {\n // Add it to the field mask, but don't add anything to updateData.\n fieldMaskPaths.push(path);\n } else {\n const parsedValue = parseData(value, childContext);\n if (parsedValue != null) {\n fieldMaskPaths.push(path);\n updateData.set(path, parsedValue);\n }\n }\n }\n }\n\n const mask = new FieldMask(fieldMaskPaths);\n return new ParsedUpdateData(\n updateData.build(),\n mask,\n context.fieldTransforms\n );\n }\n\n /** Creates a new top-level parse context. */\n private createContext(\n dataSource: UserDataSource,\n methodName: string\n ): ParseContext {\n return new ParseContext(\n {\n dataSource,\n methodName,\n path: FieldPath.EMPTY_PATH,\n arrayElement: false\n },\n this.databaseId,\n this.serializer,\n this.ignoreUndefinedProperties\n );\n }\n\n /**\n * Parse a \"query value\" (e.g. value in a where filter or a value in a cursor\n * bound).\n *\n * @param allowArrays Whether the query value is an array that may directly\n * contain additional arrays (e.g. the operand of an `in` query).\n */\n parseQueryValue(\n methodName: string,\n input: unknown,\n allowArrays = false\n ): api.Value {\n const context = this.createContext(\n allowArrays ? UserDataSource.ArrayArgument : UserDataSource.Argument,\n methodName\n );\n const parsed = parseData(input, context);\n debugAssert(parsed != null, 'Parsed data should not be null.');\n debugAssert(\n context.fieldTransforms.length === 0,\n 'Field transforms should have been disallowed.'\n );\n return parsed;\n }\n}\n\n/**\n * Parses user data to Protobuf Values.\n *\n * @param input Data to be parsed.\n * @param context A context object representing the current path being parsed,\n * the source of the data being parsed, etc.\n * @return The parsed value, or null if the value was a FieldValue sentinel\n * that should not be included in the resulting parsed data.\n */\nexport function parseData(\n input: unknown,\n context: ParseContext\n): api.Value | null {\n if (looksLikeJsonObject(input)) {\n validatePlainObject('Unsupported field value:', context, input);\n return parseObject(input, context);\n } else if (input instanceof FieldValueImpl) {\n // FieldValues usually parse into transforms (except FieldValue.delete())\n // in which case we do not want to include this field in our parsed data\n // (as doing so will overwrite the field directly prior to the transform\n // trying to transform it). So we don't add this location to\n // context.fieldMask and we return null as our parsing result.\n parseSentinelFieldValue(input, context);\n return null;\n } else {\n // If context.path is null we are inside an array and we don't support\n // field mask paths more granular than the top-level array.\n if (context.path) {\n context.fieldMask.push(context.path);\n }\n\n if (input instanceof Array) {\n // TODO(b/34871131): Include the path containing the array in the error\n // message.\n // In the case of IN queries, the parsed data is an array (representing\n // the set of values to be included for the IN query) that may directly\n // contain additional arrays (each representing an individual field\n // value), so we disable this validation.\n if (\n context.settings.arrayElement &&\n context.dataSource !== UserDataSource.ArrayArgument\n ) {\n throw context.createError('Nested arrays are not supported');\n }\n return parseArray(input as unknown[], context);\n } else {\n return parseScalarValue(input, context);\n }\n }\n}\n\nfunction parseObject(\n obj: Dict<unknown>,\n context: ParseContext\n): { mapValue: api.MapValue } {\n const fields: Dict<api.Value> = {};\n\n if (isEmpty(obj)) {\n // If we encounter an empty object, we explicitly add it to the update\n // mask to ensure that the server creates a map entry.\n if (context.path && context.path.length > 0) {\n context.fieldMask.push(context.path);\n }\n } else {\n forEach(obj, (key: string, val: unknown) => {\n const parsedValue = parseData(val, context.childContextForField(key));\n if (parsedValue != null) {\n fields[key] = parsedValue;\n }\n });\n }\n\n return { mapValue: { fields } };\n}\n\nfunction parseArray(array: unknown[], context: ParseContext): api.Value {\n const values: api.Value[] = [];\n let entryIndex = 0;\n for (const entry of array) {\n let parsedEntry = parseData(\n entry,\n context.childContextForArray(entryIndex)\n );\n if (parsedEntry == null) {\n // Just include nulls in the array for fields being replaced with a\n // sentinel.\n parsedEntry = { nullValue: 'NULL_VALUE' };\n }\n values.push(parsedEntry);\n entryIndex++;\n }\n return { arrayValue: { values } };\n}\n\n/**\n * \"Parses\" the provided FieldValueImpl, adding any necessary transforms to\n * context.fieldTransforms.\n */\nfunction parseSentinelFieldValue(\n value: FieldValueImpl,\n context: ParseContext\n): void {\n // Sentinels are only supported with writes, and not within arrays.\n if (!isWrite(context.dataSource)) {\n throw context.createError(\n `${value._methodName}() can only be used with update() and set()`\n );\n }\n if (context.path === null) {\n throw context.createError(\n `${value._methodName}() is not currently supported inside arrays`\n );\n }\n\n const fieldTransform = value.toFieldTransform(context);\n if (fieldTransform) {\n context.fieldTransforms.push(fieldTransform);\n }\n}\n\n/**\n * Helper to parse a scalar value (i.e. not an Object, Array, or FieldValue)\n *\n * @return The parsed value\n */\nfunction parseScalarValue(\n value: unknown,\n context: ParseContext\n): api.Value | null {\n if (value === null) {\n return { nullValue: 'NULL_VALUE' };\n } else if (typeof value === 'number') {\n return context.serializer.toNumber(value);\n } else if (typeof value === 'boolean') {\n return { booleanValue: value };\n } else if (typeof value === 'string') {\n return { stringValue: value };\n } else if (value instanceof Date) {\n const timestamp = Timestamp.fromDate(value);\n return { timestampValue: context.serializer.toTimestamp(timestamp) };\n } else if (value instanceof Timestamp) {\n // Firestore backend truncates precision down to microseconds. To ensure\n // offline mode works the same with regards to truncation, perform the\n // truncation immediately without waiting for the backend to do that.\n const timestamp = new Timestamp(\n value.seconds,\n Math.floor(value.nanoseconds / 1000) * 1000\n );\n return { timestampValue: context.serializer.toTimestamp(timestamp) };\n } else if (value instanceof GeoPoint) {\n return {\n geoPointValue: {\n latitude: value.latitude,\n longitude: value.longitude\n }\n };\n } else if (value instanceof Blob) {\n return { bytesValue: context.serializer.toBytes(value) };\n } else if (value instanceof DocumentReference) {\n const thisDb = context.databaseId;\n const otherDb = value.firestore._databaseId;\n if (!otherDb.isEqual(thisDb)) {\n throw context.createError(\n 'Document reference is for database ' +\n `${otherDb.projectId}/${otherDb.database} but should be ` +\n `for database ${thisDb.projectId}/${thisDb.database}`\n );\n }\n return {\n referenceValue: context.serializer.toResourceName(\n value._key.path,\n value.firestore._databaseId\n )\n };\n } else if (value === undefined && context.ignoreUndefinedProperties) {\n return null;\n } else {\n throw context.createError(\n `Unsupported field value: ${valueDescription(value)}`\n );\n }\n}\n\n/**\n * Checks whether an object looks like a JSON object that should be converted\n * into a struct. Normal class/prototype instances are considered to look like\n * JSON objects since they should be converted to a struct value. Arrays, Dates,\n * GeoPoints, etc. are not considered to look like JSON objects since they map\n * to specific FieldValue types other than ObjectValue.\n */\nfunction looksLikeJsonObject(input: unknown): boolean {\n return (\n typeof input === 'object' &&\n input !== null &&\n !(input instanceof Array) &&\n !(input instanceof Date) &&\n !(input instanceof Timestamp) &&\n !(input instanceof GeoPoint) &&\n !(input instanceof Blob) &&\n !(input instanceof DocumentReference) &&\n !(input instanceof FieldValueImpl)\n );\n}\n\nfunction validatePlainObject(\n message: string,\n context: ParseContext,\n input: unknown\n): asserts input is Dict<unknown> {\n if (!looksLikeJsonObject(input) || !isPlainObject(input)) {\n const description = valueDescription(input);\n if (description === 'an object') {\n // Massage the error if it was an object.\n throw context.createError(message + ' a custom object');\n } else {\n throw context.createError(message + ' ' + description);\n }\n }\n}\n\n/**\n * Helper that calls fromDotSeparatedString() but wraps any error thrown.\n */\nexport function fieldPathFromArgument(\n methodName: string,\n path: string | ExternalFieldPath\n): FieldPath {\n if (path instanceof ExternalFieldPath) {\n return path._internalPath;\n } else if (typeof path === 'string') {\n return fieldPathFromDotSeparatedString(methodName, path);\n } else {\n const message = 'Field path arguments must be of type string or FieldPath.';\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${methodName}() called with invalid data. ${message}`\n );\n }\n}\n\n/**\n * Wraps fromDotSeparatedString with an error message about the method that\n * was thrown.\n * @param methodName The publicly visible method name\n * @param path The dot-separated string form of a field path which will be split\n * on dots.\n */\nfunction fieldPathFromDotSeparatedString(\n methodName: string,\n path: string\n): FieldPath {\n try {\n return fromDotSeparatedString(path)._internalPath;\n } catch (e) {\n const message = errorMessage(e);\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function ${methodName}() called with invalid data. ${message}`\n );\n }\n}\n\n/**\n * Extracts the message from a caught exception, which should be an Error object\n * though JS doesn't guarantee that.\n */\nfunction errorMessage(error: Error | object): string {\n return error instanceof Error ? error.message : error.toString();\n}\n\n/** Checks `haystack` if FieldPath `needle` is present. Runs in O(n). */\nfunction fieldMaskContains(haystack: FieldPath[], needle: FieldPath): boolean {\n return haystack.some(v => v.isEqual(needle));\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { CredentialsProvider, Token } from '../api/credentials';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { TargetId } from '../core/types';\nimport { TargetData } from '../local/target_data';\nimport { Mutation, MutationResult } from '../model/mutation';\nimport * as api from '../protos/firestore_proto_api';\nimport { hardAssert, debugAssert } from '../util/assert';\nimport { AsyncQueue, DelayedOperation, TimerId } from '../util/async_queue';\nimport { Code, FirestoreError } from '../util/error';\nimport { logError, logDebug } from '../util/log';\n\nimport { isNullOrUndefined } from '../util/types';\nimport { ExponentialBackoff } from './backoff';\nimport { Connection, Stream } from './connection';\nimport { JsonProtoSerializer } from './serializer';\nimport { WatchChange } from './watch_change';\nimport { ByteString } from '../util/byte_string';\n\nconst LOG_TAG = 'PersistentStream';\n\n// The generated proto interfaces for these class are missing the database\n// field. So we add it here.\n// TODO(b/36015800): Remove this once the api generator is fixed.\ninterface ListenRequest extends api.ListenRequest {\n database?: string;\n}\nexport interface WriteRequest extends api.WriteRequest {\n database?: string;\n}\n/**\n * PersistentStream can be in one of 5 states (each described in detail below)\n * based on the following state transition diagram:\n *\n * start() called auth & connection succeeded\n * INITIAL ----------------> STARTING -----------------------------> OPEN\n * ^ | |\n * | | error occurred |\n * | \\-----------------------------v-----/\n * | |\n * backoff | |\n * elapsed | start() called |\n * \\--- BACKOFF <---------------- ERROR\n *\n * [any state] --------------------------> INITIAL\n * stop() called or\n * idle timer expired\n */\nconst enum PersistentStreamState {\n /**\n * The streaming RPC is not yet running and there's no error condition.\n * Calling start() will start the stream immediately without backoff.\n * While in this state isStarted() will return false.\n */\n Initial,\n\n /**\n * The stream is starting, either waiting for an auth token or for the stream\n * to successfully open. While in this state, isStarted() will return true but\n * isOpen() will return false.\n */\n Starting,\n\n /**\n * The streaming RPC is up and running. Requests and responses can flow\n * freely. Both isStarted() and isOpen() will return true.\n */\n Open,\n\n /**\n * The stream encountered an error. The next start attempt will back off.\n * While in this state isStarted() will return false.\n */\n Error,\n\n /**\n * An in-between state after an error where the stream is waiting before\n * re-starting. After waiting is complete, the stream will try to open.\n * While in this state isStarted() will return true but isOpen() will return\n * false.\n */\n Backoff\n}\n\n/**\n * Provides a common interface that is shared by the listeners for stream\n * events by the concrete implementation classes.\n */\nexport interface PersistentStreamListener {\n /**\n * Called after the stream was established and can accept outgoing\n * messages\n */\n onOpen: () => Promise<void>;\n /**\n * Called after the stream has closed. If there was an error, the\n * FirestoreError will be set.\n */\n onClose: (err?: FirestoreError) => Promise<void>;\n}\n\n/** The time a stream stays open after it is marked idle. */\nconst IDLE_TIMEOUT_MS = 60 * 1000;\n\n/**\n * A PersistentStream is an abstract base class that represents a streaming RPC\n * to the Firestore backend. It's built on top of the connections own support\n * for streaming RPCs, and adds several critical features for our clients:\n *\n * - Exponential backoff on failure\n * - Authentication via CredentialsProvider\n * - Dispatching all callbacks into the shared worker queue\n * - Closing idle streams after 60 seconds of inactivity\n *\n * Subclasses of PersistentStream implement serialization of models to and\n * from the JSON representation of the protocol buffers for a specific\n * streaming RPC.\n *\n * ## Starting and Stopping\n *\n * Streaming RPCs are stateful and need to be start()ed before messages can\n * be sent and received. The PersistentStream will call the onOpen() function\n * of the listener once the stream is ready to accept requests.\n *\n * Should a start() fail, PersistentStream will call the registered onClose()\n * listener with a FirestoreError indicating what went wrong.\n *\n * A PersistentStream can be started and stopped repeatedly.\n *\n * Generic types:\n * SendType: The type of the outgoing message of the underlying\n * connection stream\n * ReceiveType: The type of the incoming message of the underlying\n * connection stream\n * ListenerType: The type of the listener that will be used for callbacks\n */\nexport abstract class PersistentStream<\n SendType,\n ReceiveType,\n ListenerType extends PersistentStreamListener\n> {\n private state = PersistentStreamState.Initial;\n /**\n * A close count that's incremented every time the stream is closed; used by\n * getCloseGuardedDispatcher() to invalidate callbacks that happen after\n * close.\n */\n private closeCount = 0;\n\n private idleTimer: DelayedOperation<void> | null = null;\n private stream: Stream<SendType, ReceiveType> | null = null;\n\n protected backoff: ExponentialBackoff;\n\n constructor(\n private queue: AsyncQueue,\n connectionTimerId: TimerId,\n private idleTimerId: TimerId,\n protected connection: Connection,\n private credentialsProvider: CredentialsProvider,\n protected listener: ListenerType\n ) {\n this.backoff = new ExponentialBackoff(queue, connectionTimerId);\n }\n\n /**\n * Returns true if start() has been called and no error has occurred. True\n * indicates the stream is open or in the process of opening (which\n * encompasses respecting backoff, getting auth tokens, and starting the\n * actual RPC). Use isOpen() to determine if the stream is open and ready for\n * outbound requests.\n */\n isStarted(): boolean {\n return (\n this.state === PersistentStreamState.Starting ||\n this.state === PersistentStreamState.Open ||\n this.state === PersistentStreamState.Backoff\n );\n }\n\n /**\n * Returns true if the underlying RPC is open (the onOpen() listener has been\n * called) and the stream is ready for outbound requests.\n */\n isOpen(): boolean {\n return this.state === PersistentStreamState.Open;\n }\n\n /**\n * Starts the RPC. Only allowed if isStarted() returns false. The stream is\n * not immediately ready for use: onOpen() will be invoked when the RPC is\n * ready for outbound requests, at which point isOpen() will return true.\n *\n * When start returns, isStarted() will return true.\n */\n start(): void {\n if (this.state === PersistentStreamState.Error) {\n this.performBackoff();\n return;\n }\n\n debugAssert(\n this.state === PersistentStreamState.Initial,\n 'Already started'\n );\n this.auth();\n }\n\n /**\n * Stops the RPC. This call is idempotent and allowed regardless of the\n * current isStarted() state.\n *\n * When stop returns, isStarted() and isOpen() will both return false.\n */\n async stop(): Promise<void> {\n if (this.isStarted()) {\n await this.close(PersistentStreamState.Initial);\n }\n }\n\n /**\n * After an error the stream will usually back off on the next attempt to\n * start it. If the error warrants an immediate restart of the stream, the\n * sender can use this to indicate that the receiver should not back off.\n *\n * Each error will call the onClose() listener. That function can decide to\n * inhibit backoff if required.\n */\n inhibitBackoff(): void {\n debugAssert(\n !this.isStarted(),\n 'Can only inhibit backoff in a stopped state'\n );\n\n this.state = PersistentStreamState.Initial;\n this.backoff.reset();\n }\n\n /**\n * Marks this stream as idle. If no further actions are performed on the\n * stream for one minute, the stream will automatically close itself and\n * notify the stream's onClose() handler with Status.OK. The stream will then\n * be in a !isStarted() state, requiring the caller to start the stream again\n * before further use.\n *\n * Only streams that are in state 'Open' can be marked idle, as all other\n * states imply pending network operations.\n */\n markIdle(): void {\n // Starts the idle time if we are in state 'Open' and are not yet already\n // running a timer (in which case the previous idle timeout still applies).\n if (this.isOpen() && this.idleTimer === null) {\n this.idleTimer = this.queue.enqueueAfterDelay(\n this.idleTimerId,\n IDLE_TIMEOUT_MS,\n () => this.handleIdleCloseTimer()\n );\n }\n }\n\n /** Sends a message to the underlying stream. */\n protected sendRequest(msg: SendType): void {\n this.cancelIdleCheck();\n this.stream!.send(msg);\n }\n\n /** Called by the idle timer when the stream should close due to inactivity. */\n private async handleIdleCloseTimer(): Promise<void> {\n if (this.isOpen()) {\n // When timing out an idle stream there's no reason to force the stream into backoff when\n // it restarts so set the stream state to Initial instead of Error.\n return this.close(PersistentStreamState.Initial);\n }\n }\n\n /** Marks the stream as active again. */\n private cancelIdleCheck(): void {\n if (this.idleTimer) {\n this.idleTimer.cancel();\n this.idleTimer = null;\n }\n }\n\n /**\n * Closes the stream and cleans up as necessary:\n *\n * * closes the underlying GRPC stream;\n * * calls the onClose handler with the given 'error';\n * * sets internal stream state to 'finalState';\n * * adjusts the backoff timer based on the error\n *\n * A new stream can be opened by calling start().\n *\n * @param finalState the intended state of the stream after closing.\n * @param error the error the connection was closed with.\n */\n private async close(\n finalState: PersistentStreamState,\n error?: FirestoreError\n ): Promise<void> {\n debugAssert(this.isStarted(), 'Only started streams should be closed.');\n debugAssert(\n finalState === PersistentStreamState.Error || isNullOrUndefined(error),\n \"Can't provide an error when not in an error state.\"\n );\n\n // Cancel any outstanding timers (they're guaranteed not to execute).\n this.cancelIdleCheck();\n this.backoff.cancel();\n\n // Invalidates any stream-related callbacks (e.g. from auth or the\n // underlying stream), guaranteeing they won't execute.\n this.closeCount++;\n\n if (finalState !== PersistentStreamState.Error) {\n // If this is an intentional close ensure we don't delay our next connection attempt.\n this.backoff.reset();\n } else if (error && error.code === Code.RESOURCE_EXHAUSTED) {\n // Log the error. (Probably either 'quota exceeded' or 'max queue length reached'.)\n logError(error.toString());\n logError(\n 'Using maximum backoff delay to prevent overloading the backend.'\n );\n this.backoff.resetToMax();\n } else if (error && error.code === Code.UNAUTHENTICATED) {\n // \"unauthenticated\" error means the token was rejected. Try force refreshing it in case it\n // just expired.\n this.credentialsProvider.invalidateToken();\n }\n\n // Clean up the underlying stream because we are no longer interested in events.\n if (this.stream !== null) {\n this.tearDown();\n this.stream.close();\n this.stream = null;\n }\n\n // This state must be assigned before calling onClose() to allow the callback to\n // inhibit backoff or otherwise manipulate the state in its non-started state.\n this.state = finalState;\n\n // Notify the listener that the stream closed.\n await this.listener.onClose(error);\n }\n\n /**\n * Can be overridden to perform additional cleanup before the stream is closed.\n * Calling super.tearDown() is not required.\n */\n protected tearDown(): void {}\n\n /**\n * Used by subclasses to start the concrete RPC and return the underlying\n * connection stream.\n */\n protected abstract startRpc(\n token: Token | null\n ): Stream<SendType, ReceiveType>;\n\n /**\n * Called after the stream has received a message. The function will be\n * called on the right queue and must return a Promise.\n * @param message The message received from the stream.\n */\n protected abstract onMessage(message: ReceiveType): Promise<void>;\n\n private auth(): void {\n debugAssert(\n this.state === PersistentStreamState.Initial,\n 'Must be in initial state to auth'\n );\n\n this.state = PersistentStreamState.Starting;\n\n const dispatchIfNotClosed = this.getCloseGuardedDispatcher(this.closeCount);\n\n // TODO(mikelehen): Just use dispatchIfNotClosed, but see TODO below.\n const closeCount = this.closeCount;\n\n this.credentialsProvider.getToken().then(\n token => {\n // Stream can be stopped while waiting for authentication.\n // TODO(mikelehen): We really should just use dispatchIfNotClosed\n // and let this dispatch onto the queue, but that opened a spec test can\n // of worms that I don't want to deal with in this PR.\n if (this.closeCount === closeCount) {\n // Normally we'd have to schedule the callback on the AsyncQueue.\n // However, the following calls are safe to be called outside the\n // AsyncQueue since they don't chain asynchronous calls\n this.startStream(token);\n }\n },\n (error: Error) => {\n dispatchIfNotClosed(() => {\n const rpcError = new FirestoreError(\n Code.UNKNOWN,\n 'Fetching auth token failed: ' + error.message\n );\n return this.handleStreamClose(rpcError);\n });\n }\n );\n }\n\n private startStream(token: Token | null): void {\n debugAssert(\n this.state === PersistentStreamState.Starting,\n 'Trying to start stream in a non-starting state'\n );\n\n const dispatchIfNotClosed = this.getCloseGuardedDispatcher(this.closeCount);\n\n this.stream = this.startRpc(token);\n this.stream.onOpen(() => {\n dispatchIfNotClosed(() => {\n debugAssert(\n this.state === PersistentStreamState.Starting,\n 'Expected stream to be in state Starting, but was ' + this.state\n );\n this.state = PersistentStreamState.Open;\n return this.listener!.onOpen();\n });\n });\n this.stream.onClose((error?: FirestoreError) => {\n dispatchIfNotClosed(() => {\n return this.handleStreamClose(error);\n });\n });\n this.stream.onMessage((msg: ReceiveType) => {\n dispatchIfNotClosed(() => {\n return this.onMessage(msg);\n });\n });\n }\n\n private performBackoff(): void {\n debugAssert(\n this.state === PersistentStreamState.Error,\n 'Should only perform backoff when in Error state'\n );\n this.state = PersistentStreamState.Backoff;\n\n this.backoff.backoffAndRun(async () => {\n debugAssert(\n this.state === PersistentStreamState.Backoff,\n 'Backoff elapsed but state is now: ' + this.state\n );\n\n this.state = PersistentStreamState.Initial;\n this.start();\n debugAssert(this.isStarted(), 'PersistentStream should have started');\n });\n }\n\n // Visible for tests\n handleStreamClose(error?: FirestoreError): Promise<void> {\n debugAssert(\n this.isStarted(),\n \"Can't handle server close on non-started stream\"\n );\n logDebug(LOG_TAG, `close with error: ${error}`);\n\n this.stream = null;\n\n // In theory the stream could close cleanly, however, in our current model\n // we never expect this to happen because if we stop a stream ourselves,\n // this callback will never be called. To prevent cases where we retry\n // without a backoff accidentally, we set the stream to error in all cases.\n return this.close(PersistentStreamState.Error, error);\n }\n\n /**\n * Returns a \"dispatcher\" function that dispatches operations onto the\n * AsyncQueue but only runs them if closeCount remains unchanged. This allows\n * us to turn auth / stream callbacks into no-ops if the stream is closed /\n * re-opened, etc.\n */\n private getCloseGuardedDispatcher(\n startCloseCount: number\n ): (fn: () => Promise<void>) => void {\n return (fn: () => Promise<void>): void => {\n this.queue.enqueueAndForget(() => {\n if (this.closeCount === startCloseCount) {\n return fn();\n } else {\n logDebug(\n LOG_TAG,\n 'stream callback skipped by getCloseGuardedDispatcher.'\n );\n return Promise.resolve();\n }\n });\n };\n }\n}\n\n/** Listener for the PersistentWatchStream */\nexport interface WatchStreamListener extends PersistentStreamListener {\n /**\n * Called on a watchChange. The snapshot parameter will be MIN if the watch\n * change did not have a snapshot associated with it.\n */\n onWatchChange: (\n watchChange: WatchChange,\n snapshot: SnapshotVersion\n ) => Promise<void>;\n}\n\n/**\n * A PersistentStream that implements the Listen RPC.\n *\n * Once the Listen stream has called the onOpen() listener, any number of\n * listen() and unlisten() calls can be made to control what changes will be\n * sent from the server for ListenResponses.\n */\nexport class PersistentListenStream extends PersistentStream<\n api.ListenRequest,\n api.ListenResponse,\n WatchStreamListener\n> {\n constructor(\n queue: AsyncQueue,\n connection: Connection,\n credentials: CredentialsProvider,\n private serializer: JsonProtoSerializer,\n listener: WatchStreamListener\n ) {\n super(\n queue,\n TimerId.ListenStreamConnectionBackoff,\n TimerId.ListenStreamIdle,\n connection,\n credentials,\n listener\n );\n }\n\n protected startRpc(\n token: Token | null\n ): Stream<api.ListenRequest, api.ListenResponse> {\n return this.connection.openStream<api.ListenRequest, api.ListenResponse>(\n 'Listen',\n token\n );\n }\n\n protected onMessage(watchChangeProto: api.ListenResponse): Promise<void> {\n // A successful response means the stream is healthy\n this.backoff.reset();\n\n const watchChange = this.serializer.fromWatchChange(watchChangeProto);\n const snapshot = this.serializer.versionFromListenResponse(\n watchChangeProto\n );\n return this.listener!.onWatchChange(watchChange, snapshot);\n }\n\n /**\n * Registers interest in the results of the given target. If the target\n * includes a resumeToken it will be included in the request. Results that\n * affect the target will be streamed back as WatchChange messages that\n * reference the targetId.\n */\n watch(targetData: TargetData): void {\n const request: ListenRequest = {};\n request.database = this.serializer.encodedDatabaseId;\n request.addTarget = this.serializer.toTarget(targetData);\n\n const labels = this.serializer.toListenRequestLabels(targetData);\n if (labels) {\n request.labels = labels;\n }\n\n this.sendRequest(request);\n }\n\n /**\n * Unregisters interest in the results of the target associated with the\n * given targetId.\n */\n unwatch(targetId: TargetId): void {\n const request: ListenRequest = {};\n request.database = this.serializer.encodedDatabaseId;\n request.removeTarget = targetId;\n this.sendRequest(request);\n }\n}\n\n/** Listener for the PersistentWriteStream */\nexport interface WriteStreamListener extends PersistentStreamListener {\n /**\n * Called by the PersistentWriteStream upon a successful handshake response\n * from the server, which is the receiver's cue to send any pending writes.\n */\n onHandshakeComplete: () => Promise<void>;\n\n /**\n * Called by the PersistentWriteStream upon receiving a StreamingWriteResponse\n * from the server that contains a mutation result.\n */\n onMutationResult: (\n commitVersion: SnapshotVersion,\n results: MutationResult[]\n ) => Promise<void>;\n}\n\n/**\n * A Stream that implements the Write RPC.\n *\n * The Write RPC requires the caller to maintain special streamToken\n * state in between calls, to help the server understand which responses the\n * client has processed by the time the next request is made. Every response\n * will contain a streamToken; this value must be passed to the next\n * request.\n *\n * After calling start() on this stream, the next request must be a handshake,\n * containing whatever streamToken is on hand. Once a response to this\n * request is received, all pending mutations may be submitted. When\n * submitting multiple batches of mutations at the same time, it's\n * okay to use the same streamToken for the calls to writeMutations.\n *\n * TODO(b/33271235): Use proto types\n */\nexport class PersistentWriteStream extends PersistentStream<\n api.WriteRequest,\n api.WriteResponse,\n WriteStreamListener\n> {\n private handshakeComplete_ = false;\n\n constructor(\n queue: AsyncQueue,\n connection: Connection,\n credentials: CredentialsProvider,\n private serializer: JsonProtoSerializer,\n listener: WriteStreamListener\n ) {\n super(\n queue,\n TimerId.WriteStreamConnectionBackoff,\n TimerId.WriteStreamIdle,\n connection,\n credentials,\n listener\n );\n }\n\n /**\n * The last received stream token from the server, used to acknowledge which\n * responses the client has processed. Stream tokens are opaque checkpoint\n * markers whose only real value is their inclusion in the next request.\n *\n * PersistentWriteStream manages propagating this value from responses to the\n * next request.\n */\n lastStreamToken: ByteString = ByteString.EMPTY_BYTE_STRING;\n\n /**\n * Tracks whether or not a handshake has been successfully exchanged and\n * the stream is ready to accept mutations.\n */\n get handshakeComplete(): boolean {\n return this.handshakeComplete_;\n }\n\n // Override of PersistentStream.start\n start(): void {\n this.handshakeComplete_ = false;\n super.start();\n }\n\n protected tearDown(): void {\n if (this.handshakeComplete_) {\n this.writeMutations([]);\n }\n }\n\n protected startRpc(\n token: Token | null\n ): Stream<api.WriteRequest, api.WriteResponse> {\n return this.connection.openStream<api.WriteRequest, api.WriteResponse>(\n 'Write',\n token\n );\n }\n\n protected onMessage(responseProto: api.WriteResponse): Promise<void> {\n // Always capture the last stream token.\n hardAssert(\n !!responseProto.streamToken,\n 'Got a write response without a stream token'\n );\n this.lastStreamToken = this.serializer.fromBytes(responseProto.streamToken);\n\n if (!this.handshakeComplete_) {\n // The first response is always the handshake response\n hardAssert(\n !responseProto.writeResults || responseProto.writeResults.length === 0,\n 'Got mutation results for handshake'\n );\n this.handshakeComplete_ = true;\n return this.listener!.onHandshakeComplete();\n } else {\n // A successful first write response means the stream is healthy,\n // Note, that we could consider a successful handshake healthy, however,\n // the write itself might be causing an error we want to back off from.\n this.backoff.reset();\n\n const results = this.serializer.fromWriteResults(\n responseProto.writeResults,\n responseProto.commitTime\n );\n const commitVersion = this.serializer.fromVersion(\n responseProto.commitTime!\n );\n return this.listener!.onMutationResult(commitVersion, results);\n }\n }\n\n /**\n * Sends an initial streamToken to the server, performing the handshake\n * required to make the StreamingWrite RPC work. Subsequent\n * calls should wait until onHandshakeComplete was called.\n */\n writeHandshake(): void {\n debugAssert(this.isOpen(), 'Writing handshake requires an opened stream');\n debugAssert(!this.handshakeComplete_, 'Handshake already completed');\n // TODO(dimond): Support stream resumption. We intentionally do not set the\n // stream token on the handshake, ignoring any stream token we might have.\n const request: WriteRequest = {};\n request.database = this.serializer.encodedDatabaseId;\n this.sendRequest(request);\n }\n\n /** Sends a group of mutations to the Firestore backend to apply. */\n writeMutations(mutations: Mutation[]): void {\n debugAssert(this.isOpen(), 'Writing mutations requires an opened stream');\n debugAssert(\n this.handshakeComplete_,\n 'Handshake must be complete before writing mutations'\n );\n debugAssert(\n this.lastStreamToken.approximateByteSize() > 0,\n 'Trying to write mutation without a token'\n );\n\n const request: WriteRequest = {\n streamToken: this.serializer.toBytes(this.lastStreamToken),\n writes: mutations.map(mutation => this.serializer.toMutation(mutation))\n };\n\n this.sendRequest(request);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { CredentialsProvider } from '../api/credentials';\nimport { MaybeDocument, Document } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { Mutation, MutationResult } from '../model/mutation';\nimport * as api from '../protos/firestore_proto_api';\nimport { debugCast, hardAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { Connection } from './connection';\nimport { JsonProtoSerializer } from './serializer';\nimport {\n PersistentListenStream,\n PersistentWriteStream,\n WatchStreamListener,\n WriteStreamListener\n} from './persistent_stream';\nimport { AsyncQueue } from '../util/async_queue';\nimport { Query } from '../core/query';\n\n/**\n * Datastore and its related methods are a wrapper around the external Google\n * Cloud Datastore grpc API, which provides an interface that is more convenient\n * for the rest of the client SDK architecture to consume.\n */\nexport class Datastore {\n // Make sure that the structural type of `Datastore` is unique.\n // See https://github.com/microsoft/TypeScript/issues/5451\n private _ = undefined;\n}\n\n/**\n * An implementation of Datastore that exposes additional state for internal\n * consumption.\n */\nclass DatastoreImpl extends Datastore {\n constructor(\n public readonly connection: Connection,\n public readonly credentials: CredentialsProvider,\n public readonly serializer: JsonProtoSerializer\n ) {\n super();\n }\n\n /** Gets an auth token and invokes the provided RPC. */\n invokeRPC<Req, Resp>(rpcName: string, request: Req): Promise<Resp> {\n return this.credentials\n .getToken()\n .then(token => {\n return this.connection.invokeRPC<Req, Resp>(rpcName, request, token);\n })\n .catch((error: FirestoreError) => {\n if (error.code === Code.UNAUTHENTICATED) {\n this.credentials.invalidateToken();\n }\n throw error;\n });\n }\n\n /** Gets an auth token and invokes the provided RPC with streamed results. */\n invokeStreamingRPC<Req, Resp>(\n rpcName: string,\n request: Req\n ): Promise<Resp[]> {\n return this.credentials\n .getToken()\n .then(token => {\n return this.connection.invokeStreamingRPC<Req, Resp>(\n rpcName,\n request,\n token\n );\n })\n .catch((error: FirestoreError) => {\n if (error.code === Code.UNAUTHENTICATED) {\n this.credentials.invalidateToken();\n }\n throw error;\n });\n }\n}\n\nexport function newDatastore(\n connection: Connection,\n credentials: CredentialsProvider,\n serializer: JsonProtoSerializer\n): Datastore {\n return new DatastoreImpl(connection, credentials, serializer);\n}\n\nexport async function invokeCommitRpc(\n datastore: Datastore,\n mutations: Mutation[]\n): Promise<MutationResult[]> {\n const datastoreImpl = debugCast(datastore, DatastoreImpl);\n const params = {\n database: datastoreImpl.serializer.encodedDatabaseId,\n writes: mutations.map(m => datastoreImpl.serializer.toMutation(m))\n };\n const response = await datastoreImpl.invokeRPC<\n api.CommitRequest,\n api.CommitResponse\n >('Commit', params);\n return datastoreImpl.serializer.fromWriteResults(\n response.writeResults,\n response.commitTime\n );\n}\n\nexport async function invokeBatchGetDocumentsRpc(\n datastore: Datastore,\n keys: DocumentKey[]\n): Promise<MaybeDocument[]> {\n const datastoreImpl = debugCast(datastore, DatastoreImpl);\n const params = {\n database: datastoreImpl.serializer.encodedDatabaseId,\n documents: keys.map(k => datastoreImpl.serializer.toName(k))\n };\n const response = await datastoreImpl.invokeStreamingRPC<\n api.BatchGetDocumentsRequest,\n api.BatchGetDocumentsResponse\n >('BatchGetDocuments', params);\n\n const docs = new Map<string, MaybeDocument>();\n response.forEach(proto => {\n const doc = datastoreImpl.serializer.fromMaybeDocument(proto);\n docs.set(doc.key.toString(), doc);\n });\n const result: MaybeDocument[] = [];\n keys.forEach(key => {\n const doc = docs.get(key.toString());\n hardAssert(!!doc, 'Missing entity in write response for ' + key);\n result.push(doc);\n });\n return result;\n}\n\nexport async function invokeRunQueryRpc(\n datastore: Datastore,\n query: Query\n): Promise<Document[]> {\n const datastoreImpl = debugCast(datastore, DatastoreImpl);\n const { structuredQuery, parent } = datastoreImpl.serializer.toQueryTarget(\n query.toTarget()\n );\n const params = {\n database: datastoreImpl.serializer.encodedDatabaseId,\n parent,\n structuredQuery\n };\n\n const response = await datastoreImpl.invokeStreamingRPC<\n api.RunQueryRequest,\n api.RunQueryResponse\n >('RunQuery', params);\n\n return (\n response\n // Omit RunQueryResponses that only contain readTimes.\n .filter(proto => !!proto.document)\n .map(proto => datastoreImpl.serializer.fromDocument(proto.document!))\n );\n}\n\nexport function newPersistentWriteStream(\n datastore: Datastore,\n queue: AsyncQueue,\n listener: WriteStreamListener\n): PersistentWriteStream {\n const datastoreImpl = debugCast(datastore, DatastoreImpl);\n return new PersistentWriteStream(\n queue,\n datastoreImpl.connection,\n datastoreImpl.credentials,\n datastoreImpl.serializer,\n listener\n );\n}\n\nexport function newPersistentWatchStream(\n datastore: Datastore,\n queue: AsyncQueue,\n listener: WatchStreamListener\n): PersistentListenStream {\n const datastoreImpl = debugCast(datastore, DatastoreImpl);\n return new PersistentListenStream(\n queue,\n datastoreImpl.connection,\n datastoreImpl.credentials,\n datastoreImpl.serializer,\n listener\n );\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { ParsedSetData, ParsedUpdateData } from '../api/user_data_reader';\nimport { documentVersionMap } from '../model/collections';\nimport { Document, MaybeDocument, NoDocument } from '../model/document';\n\nimport { DocumentKey } from '../model/document_key';\nimport {\n DeleteMutation,\n Mutation,\n Precondition,\n VerifyMutation\n} from '../model/mutation';\nimport {\n Datastore,\n invokeBatchGetDocumentsRpc,\n invokeCommitRpc\n} from '../remote/datastore';\nimport { fail, debugAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { SnapshotVersion } from './snapshot_version';\n\n/**\n * Internal transaction object responsible for accumulating the mutations to\n * perform and the base versions for any documents read.\n */\nexport class Transaction {\n // The version of each document that was read during this transaction.\n private readVersions = documentVersionMap();\n private mutations: Mutation[] = [];\n private committed = false;\n\n /**\n * A deferred usage error that occurred previously in this transaction that\n * will cause the transaction to fail once it actually commits.\n */\n private lastWriteError: FirestoreError | null = null;\n\n /**\n * Set of documents that have been written in the transaction.\n *\n * When there's more than one write to the same key in a transaction, any\n * writes after the first are handled differently.\n */\n private writtenDocs: Set<DocumentKey> = new Set();\n\n constructor(private datastore: Datastore) {}\n\n async lookup(keys: DocumentKey[]): Promise<MaybeDocument[]> {\n this.ensureCommitNotCalled();\n\n if (this.mutations.length > 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Firestore transactions require all reads to be executed before all writes.'\n );\n }\n const docs = await invokeBatchGetDocumentsRpc(this.datastore, keys);\n docs.forEach(doc => {\n if (doc instanceof NoDocument || doc instanceof Document) {\n this.recordVersion(doc);\n } else {\n fail('Document in a transaction was a ' + doc.constructor.name);\n }\n });\n return docs;\n }\n\n set(key: DocumentKey, data: ParsedSetData): void {\n this.write(data.toMutations(key, this.precondition(key)));\n this.writtenDocs.add(key);\n }\n\n update(key: DocumentKey, data: ParsedUpdateData): void {\n try {\n this.write(data.toMutations(key, this.preconditionForUpdate(key)));\n } catch (e) {\n this.lastWriteError = e;\n }\n this.writtenDocs.add(key);\n }\n\n delete(key: DocumentKey): void {\n this.write([new DeleteMutation(key, this.precondition(key))]);\n this.writtenDocs.add(key);\n }\n\n async commit(): Promise<void> {\n this.ensureCommitNotCalled();\n\n if (this.lastWriteError) {\n throw this.lastWriteError;\n }\n let unwritten = this.readVersions;\n // For each mutation, note that the doc was written.\n this.mutations.forEach(mutation => {\n unwritten = unwritten.remove(mutation.key);\n });\n // For each document that was read but not written to, we want to perform\n // a `verify` operation.\n unwritten.forEach((key, _version) => {\n this.mutations.push(new VerifyMutation(key, this.precondition(key)));\n });\n await invokeCommitRpc(this.datastore, this.mutations);\n this.committed = true;\n }\n\n private recordVersion(doc: MaybeDocument): void {\n let docVersion: SnapshotVersion;\n\n if (doc instanceof Document) {\n docVersion = doc.version;\n } else if (doc instanceof NoDocument) {\n // For deleted docs, we must use baseVersion 0 when we overwrite them.\n docVersion = SnapshotVersion.min();\n } else {\n throw fail('Document in a transaction was a ' + doc.constructor.name);\n }\n\n const existingVersion = this.readVersions.get(doc.key);\n if (existingVersion !== null) {\n if (!docVersion.isEqual(existingVersion)) {\n // This transaction will fail no matter what.\n throw new FirestoreError(\n Code.ABORTED,\n 'Document version changed between two reads.'\n );\n }\n } else {\n this.readVersions = this.readVersions.insert(doc.key, docVersion);\n }\n }\n\n /**\n * Returns the version of this document when it was read in this transaction,\n * as a precondition, or no precondition if it was not read.\n */\n private precondition(key: DocumentKey): Precondition {\n const version = this.readVersions.get(key);\n if (!this.writtenDocs.has(key) && version) {\n return Precondition.updateTime(version);\n } else {\n return Precondition.none();\n }\n }\n\n /**\n * Returns the precondition for a document if the operation is an update.\n */\n private preconditionForUpdate(key: DocumentKey): Precondition {\n const version = this.readVersions.get(key);\n // The first time a document is written, we want to take into account the\n // read time and existence\n if (!this.writtenDocs.has(key) && version) {\n if (version.isEqual(SnapshotVersion.min())) {\n // The document doesn't exist, so fail the transaction.\n\n // This has to be validated locally because you can't send a\n // precondition that a document does not exist without changing the\n // semantics of the backend write to be an insert. This is the reverse\n // of what we want, since we want to assert that the document doesn't\n // exist but then send the update and have it fail. Since we can't\n // express that to the backend, we have to validate locally.\n\n // Note: this can change once we can send separate verify writes in the\n // transaction.\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n \"Can't update a document that doesn't exist.\"\n );\n }\n // Document exists, base precondition on document update time.\n return Precondition.updateTime(version);\n } else {\n // Document was not read, so we just use the preconditions for a blind\n // update.\n return Precondition.exists(true);\n }\n }\n\n private write(mutations: Mutation[]): void {\n this.ensureCommitNotCalled();\n this.mutations = this.mutations.concat(mutations);\n }\n\n private ensureCommitNotCalled(): void {\n debugAssert(\n !this.committed,\n 'A transaction object cannot be used after its update callback has been invoked.'\n );\n }\n}\n","/**\n * @license\n * Copyright 2018 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { OnlineState } from '../core/types';\nimport { debugAssert } from '../util/assert';\nimport { AsyncQueue, DelayedOperation, TimerId } from '../util/async_queue';\nimport { FirestoreError } from '../util/error';\nimport { logError, logDebug } from '../util/log';\n\nconst LOG_TAG = 'OnlineStateTracker';\n\n// To deal with transient failures, we allow multiple stream attempts before\n// giving up and transitioning from OnlineState.Unknown to Offline.\n// TODO(mikelehen): This used to be set to 2 as a mitigation for b/66228394.\n// @jdimond thinks that bug is sufficiently fixed so that we can set this back\n// to 1. If that works okay, we could potentially remove this logic entirely.\nconst MAX_WATCH_STREAM_FAILURES = 1;\n\n// To deal with stream attempts that don't succeed or fail in a timely manner,\n// we have a timeout for OnlineState to reach Online or Offline.\n// If the timeout is reached, we transition to Offline rather than waiting\n// indefinitely.\nconst ONLINE_STATE_TIMEOUT_MS = 10 * 1000;\n\n/**\n * A component used by the RemoteStore to track the OnlineState (that is,\n * whether or not the client as a whole should be considered to be online or\n * offline), implementing the appropriate heuristics.\n *\n * In particular, when the client is trying to connect to the backend, we\n * allow up to MAX_WATCH_STREAM_FAILURES within ONLINE_STATE_TIMEOUT_MS for\n * a connection to succeed. If we have too many failures or the timeout elapses,\n * then we set the OnlineState to Offline, and the client will behave as if\n * it is offline (get()s will return cached data, etc.).\n */\nexport class OnlineStateTracker {\n /** The current OnlineState. */\n private state = OnlineState.Unknown;\n\n /**\n * A count of consecutive failures to open the stream. If it reaches the\n * maximum defined by MAX_WATCH_STREAM_FAILURES, we'll set the OnlineState to\n * Offline.\n */\n private watchStreamFailures = 0;\n\n /**\n * A timer that elapses after ONLINE_STATE_TIMEOUT_MS, at which point we\n * transition from OnlineState.Unknown to OnlineState.Offline without waiting\n * for the stream to actually fail (MAX_WATCH_STREAM_FAILURES times).\n */\n private onlineStateTimer: DelayedOperation<void> | null = null;\n\n /**\n * Whether the client should log a warning message if it fails to connect to\n * the backend (initially true, cleared after a successful stream, or if we've\n * logged the message already).\n */\n private shouldWarnClientIsOffline = true;\n\n constructor(\n private asyncQueue: AsyncQueue,\n private onlineStateHandler: (onlineState: OnlineState) => void\n ) {}\n\n /**\n * Called by RemoteStore when a watch stream is started (including on each\n * backoff attempt).\n *\n * If this is the first attempt, it sets the OnlineState to Unknown and starts\n * the onlineStateTimer.\n */\n handleWatchStreamStart(): void {\n if (this.watchStreamFailures === 0) {\n this.setAndBroadcast(OnlineState.Unknown);\n\n debugAssert(\n this.onlineStateTimer === null,\n `onlineStateTimer shouldn't be started yet`\n );\n this.onlineStateTimer = this.asyncQueue.enqueueAfterDelay(\n TimerId.OnlineStateTimeout,\n ONLINE_STATE_TIMEOUT_MS,\n () => {\n this.onlineStateTimer = null;\n debugAssert(\n this.state === OnlineState.Unknown,\n 'Timer should be canceled if we transitioned to a different state.'\n );\n this.logClientOfflineWarningIfNecessary(\n `Backend didn't respond within ${ONLINE_STATE_TIMEOUT_MS / 1000} ` +\n `seconds.`\n );\n this.setAndBroadcast(OnlineState.Offline);\n\n // NOTE: handleWatchStreamFailure() will continue to increment\n // watchStreamFailures even though we are already marked Offline,\n // but this is non-harmful.\n\n return Promise.resolve();\n }\n );\n }\n }\n\n /**\n * Updates our OnlineState as appropriate after the watch stream reports a\n * failure. The first failure moves us to the 'Unknown' state. We then may\n * allow multiple failures (based on MAX_WATCH_STREAM_FAILURES) before we\n * actually transition to the 'Offline' state.\n */\n handleWatchStreamFailure(error: FirestoreError): void {\n if (this.state === OnlineState.Online) {\n this.setAndBroadcast(OnlineState.Unknown);\n\n // To get to OnlineState.Online, set() must have been called which would\n // have reset our heuristics.\n debugAssert(\n this.watchStreamFailures === 0,\n 'watchStreamFailures must be 0'\n );\n debugAssert(\n this.onlineStateTimer === null,\n 'onlineStateTimer must be null'\n );\n } else {\n this.watchStreamFailures++;\n if (this.watchStreamFailures >= MAX_WATCH_STREAM_FAILURES) {\n this.clearOnlineStateTimer();\n\n this.logClientOfflineWarningIfNecessary(\n `Connection failed ${MAX_WATCH_STREAM_FAILURES} ` +\n `times. Most recent error: ${error.toString()}`\n );\n\n this.setAndBroadcast(OnlineState.Offline);\n }\n }\n }\n\n /**\n * Explicitly sets the OnlineState to the specified state.\n *\n * Note that this resets our timers / failure counters, etc. used by our\n * Offline heuristics, so must not be used in place of\n * handleWatchStreamStart() and handleWatchStreamFailure().\n */\n set(newState: OnlineState): void {\n this.clearOnlineStateTimer();\n this.watchStreamFailures = 0;\n\n if (newState === OnlineState.Online) {\n // We've connected to watch at least once. Don't warn the developer\n // about being offline going forward.\n this.shouldWarnClientIsOffline = false;\n }\n\n this.setAndBroadcast(newState);\n }\n\n private setAndBroadcast(newState: OnlineState): void {\n if (newState !== this.state) {\n this.state = newState;\n this.onlineStateHandler(newState);\n }\n }\n\n private logClientOfflineWarningIfNecessary(details: string): void {\n const message =\n `Could not reach Cloud Firestore backend. ${details}\\n` +\n `This typically indicates that your device does not have a healthy ` +\n `Internet connection at the moment. The client will operate in offline ` +\n `mode until it is able to successfully connect to the backend.`;\n if (this.shouldWarnClientIsOffline) {\n logError(message);\n this.shouldWarnClientIsOffline = false;\n } else {\n logDebug(LOG_TAG, message);\n }\n }\n\n private clearOnlineStateTimer(): void {\n if (this.onlineStateTimer !== null) {\n this.onlineStateTimer.cancel();\n this.onlineStateTimer = null;\n }\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { Transaction } from '../core/transaction';\nimport { OnlineState, TargetId } from '../core/types';\nimport { ignoreIfPrimaryLeaseLoss, LocalStore } from '../local/local_store';\nimport { TargetData, TargetPurpose } from '../local/target_data';\nimport { MutationResult } from '../model/mutation';\nimport {\n BATCHID_UNKNOWN,\n MutationBatch,\n MutationBatchResult\n} from '../model/mutation_batch';\nimport { debugAssert } from '../util/assert';\nimport { FirestoreError } from '../util/error';\nimport { logDebug } from '../util/log';\nimport { DocumentKeySet } from '../model/collections';\nimport { AsyncQueue } from '../util/async_queue';\nimport { ConnectivityMonitor, NetworkStatus } from './connectivity_monitor';\nimport {\n Datastore,\n newPersistentWatchStream,\n newPersistentWriteStream\n} from './datastore';\nimport { OnlineStateTracker } from './online_state_tracker';\nimport {\n PersistentListenStream,\n PersistentWriteStream\n} from './persistent_stream';\nimport { RemoteSyncer } from './remote_syncer';\nimport { isPermanentError, isPermanentWriteError } from './rpc_error';\nimport {\n DocumentWatchChange,\n ExistenceFilterChange,\n TargetMetadataProvider,\n WatchChange,\n WatchChangeAggregator,\n WatchTargetChange,\n WatchTargetChangeState\n} from './watch_change';\nimport { ByteString } from '../util/byte_string';\nimport { isIndexedDbTransactionError } from '../local/simple_db';\n\nconst LOG_TAG = 'RemoteStore';\n\n// TODO(b/35853402): Negotiate this with the stream.\nconst MAX_PENDING_WRITES = 10;\n\n/**\n * RemoteStore - An interface to remotely stored data, basically providing a\n * wrapper around the Datastore that is more reliable for the rest of the\n * system.\n *\n * RemoteStore is responsible for maintaining the connection to the server.\n * - maintaining a list of active listens.\n * - reconnecting when the connection is dropped.\n * - resuming all the active listens on reconnect.\n *\n * RemoteStore handles all incoming events from the Datastore.\n * - listening to the watch stream and repackaging the events as RemoteEvents\n * - notifying SyncEngine of any changes to the active listens.\n *\n * RemoteStore takes writes from other components and handles them reliably.\n * - pulling pending mutations from LocalStore and sending them to Datastore.\n * - retrying mutations that failed because of network problems.\n * - acking mutations to the SyncEngine once they are accepted or rejected.\n */\nexport class RemoteStore implements TargetMetadataProvider {\n /**\n * A list of up to MAX_PENDING_WRITES writes that we have fetched from the\n * LocalStore via fillWritePipeline() and have or will send to the write\n * stream.\n *\n * Whenever writePipeline.length > 0 the RemoteStore will attempt to start or\n * restart the write stream. When the stream is established the writes in the\n * pipeline will be sent in order.\n *\n * Writes remain in writePipeline until they are acknowledged by the backend\n * and thus will automatically be re-sent if the stream is interrupted /\n * restarted before they're acknowledged.\n *\n * Write responses from the backend are linked to their originating request\n * purely based on order, and so we can just shift() writes from the front of\n * the writePipeline as we receive responses.\n */\n private writePipeline: MutationBatch[] = [];\n\n /**\n * A mapping of watched targets that the client cares about tracking and the\n * user has explicitly called a 'listen' for this target.\n *\n * These targets may or may not have been sent to or acknowledged by the\n * server. On re-establishing the listen stream, these targets should be sent\n * to the server. The targets removed with unlistens are removed eagerly\n * without waiting for confirmation from the listen stream.\n */\n private listenTargets = new Map<TargetId, TargetData>();\n\n private connectivityMonitor: ConnectivityMonitor;\n private watchStream: PersistentListenStream;\n private writeStream: PersistentWriteStream;\n private watchChangeAggregator: WatchChangeAggregator | null = null;\n\n /**\n * Set to true by enableNetwork() and false by disableNetwork() and indicates\n * the user-preferred network state.\n */\n private networkEnabled = false;\n\n private isPrimary = false;\n\n /**\n * When set to `true`, the network was taken offline due to an IndexedDB\n * failure. The state is flipped to `false` when access becomes available\n * again.\n */\n private indexedDbFailed = false;\n\n private onlineStateTracker: OnlineStateTracker;\n\n constructor(\n /**\n * The local store, used to fill the write pipeline with outbound mutations.\n */\n private localStore: LocalStore,\n /** The client-side proxy for interacting with the backend. */\n private datastore: Datastore,\n private asyncQueue: AsyncQueue,\n onlineStateHandler: (onlineState: OnlineState) => void,\n connectivityMonitor: ConnectivityMonitor\n ) {\n this.connectivityMonitor = connectivityMonitor;\n this.connectivityMonitor.addCallback((status: NetworkStatus) => {\n asyncQueue.enqueueAndForget(async () => {\n if (this.canUseNetwork()) {\n logDebug(\n LOG_TAG,\n 'Restarting streams for network reachability change.'\n );\n await this.restartNetwork();\n }\n });\n });\n\n this.onlineStateTracker = new OnlineStateTracker(\n asyncQueue,\n onlineStateHandler\n );\n\n // Create streams (but note they're not started yet).\n this.watchStream = newPersistentWatchStream(this.datastore, asyncQueue, {\n onOpen: this.onWatchStreamOpen.bind(this),\n onClose: this.onWatchStreamClose.bind(this),\n onWatchChange: this.onWatchStreamChange.bind(this)\n });\n\n this.writeStream = newPersistentWriteStream(this.datastore, asyncQueue, {\n onOpen: this.onWriteStreamOpen.bind(this),\n onClose: this.onWriteStreamClose.bind(this),\n onHandshakeComplete: this.onWriteHandshakeComplete.bind(this),\n onMutationResult: this.onMutationResult.bind(this)\n });\n }\n\n /**\n * SyncEngine to notify of watch and write events. This must be set\n * immediately after construction.\n */\n syncEngine!: RemoteSyncer;\n\n /**\n * Starts up the remote store, creating streams, restoring state from\n * LocalStore, etc.\n */\n start(): Promise<void> {\n return this.enableNetwork();\n }\n\n /** Re-enables the network. Idempotent. */\n enableNetwork(): Promise<void> {\n this.networkEnabled = true;\n return this.enableNetworkInternal();\n }\n\n private async enableNetworkInternal(): Promise<void> {\n if (this.canUseNetwork()) {\n this.writeStream.lastStreamToken = await this.localStore.getLastStreamToken();\n\n if (this.shouldStartWatchStream()) {\n this.startWatchStream();\n } else {\n this.onlineStateTracker.set(OnlineState.Unknown);\n }\n\n // This will start the write stream if necessary.\n await this.fillWritePipeline();\n }\n }\n\n /**\n * Temporarily disables the network. The network can be re-enabled using\n * enableNetwork().\n */\n async disableNetwork(): Promise<void> {\n this.networkEnabled = false;\n await this.disableNetworkInternal();\n\n // Set the OnlineState to Offline so get()s return from cache, etc.\n this.onlineStateTracker.set(OnlineState.Offline);\n }\n\n private async disableNetworkInternal(): Promise<void> {\n await this.writeStream.stop();\n await this.watchStream.stop();\n\n if (this.writePipeline.length > 0) {\n logDebug(\n LOG_TAG,\n `Stopping write stream with ${this.writePipeline.length} pending writes`\n );\n this.writePipeline = [];\n }\n\n this.cleanUpWatchStreamState();\n }\n\n async shutdown(): Promise<void> {\n logDebug(LOG_TAG, 'RemoteStore shutting down.');\n this.networkEnabled = false;\n await this.disableNetworkInternal();\n this.connectivityMonitor.shutdown();\n\n // Set the OnlineState to Unknown (rather than Offline) to avoid potentially\n // triggering spurious listener events with cached data, etc.\n this.onlineStateTracker.set(OnlineState.Unknown);\n }\n\n /**\n * Starts new listen for the given target. Uses resume token if provided. It\n * is a no-op if the target of given `TargetData` is already being listened to.\n */\n listen(targetData: TargetData): void {\n if (this.listenTargets.has(targetData.targetId)) {\n return;\n }\n\n // Mark this as something the client is currently listening for.\n this.listenTargets.set(targetData.targetId, targetData);\n\n if (this.shouldStartWatchStream()) {\n // The listen will be sent in onWatchStreamOpen\n this.startWatchStream();\n } else if (this.watchStream.isOpen()) {\n this.sendWatchRequest(targetData);\n }\n }\n\n /**\n * Removes the listen from server. It is a no-op if the given target id is\n * not being listened to.\n */\n unlisten(targetId: TargetId): void {\n debugAssert(\n this.listenTargets.has(targetId),\n `unlisten called on target no currently watched: ${targetId}`\n );\n\n this.listenTargets.delete(targetId);\n if (this.watchStream.isOpen()) {\n this.sendUnwatchRequest(targetId);\n }\n\n if (this.listenTargets.size === 0) {\n if (this.watchStream.isOpen()) {\n this.watchStream.markIdle();\n } else if (this.canUseNetwork()) {\n // Revert to OnlineState.Unknown if the watch stream is not open and we\n // have no listeners, since without any listens to send we cannot\n // confirm if the stream is healthy and upgrade to OnlineState.Online.\n this.onlineStateTracker.set(OnlineState.Unknown);\n }\n }\n }\n\n /** {@link TargetMetadataProvider.getTargetDataForTarget} */\n getTargetDataForTarget(targetId: TargetId): TargetData | null {\n return this.listenTargets.get(targetId) || null;\n }\n\n /** {@link TargetMetadataProvider.getRemoteKeysForTarget} */\n getRemoteKeysForTarget(targetId: TargetId): DocumentKeySet {\n return this.syncEngine.getRemoteKeysForTarget(targetId);\n }\n\n /**\n * We need to increment the the expected number of pending responses we're due\n * from watch so we wait for the ack to process any messages from this target.\n */\n private sendWatchRequest(targetData: TargetData): void {\n this.watchChangeAggregator!.recordPendingTargetRequest(targetData.targetId);\n this.watchStream.watch(targetData);\n }\n\n /**\n * We need to increment the expected number of pending responses we're due\n * from watch so we wait for the removal on the server before we process any\n * messages from this target.\n */\n private sendUnwatchRequest(targetId: TargetId): void {\n this.watchChangeAggregator!.recordPendingTargetRequest(targetId);\n this.watchStream.unwatch(targetId);\n }\n\n private startWatchStream(): void {\n debugAssert(\n this.shouldStartWatchStream(),\n 'startWatchStream() called when shouldStartWatchStream() is false.'\n );\n\n this.watchChangeAggregator = new WatchChangeAggregator(this);\n this.watchStream.start();\n this.onlineStateTracker.handleWatchStreamStart();\n }\n\n /**\n * Returns whether the watch stream should be started because it's necessary\n * and has not yet been started.\n */\n private shouldStartWatchStream(): boolean {\n return (\n this.canUseNetwork() &&\n !this.watchStream.isStarted() &&\n this.listenTargets.size > 0\n );\n }\n\n canUseNetwork(): boolean {\n return !this.indexedDbFailed && this.isPrimary && this.networkEnabled;\n }\n\n private cleanUpWatchStreamState(): void {\n this.watchChangeAggregator = null;\n }\n\n private async onWatchStreamOpen(): Promise<void> {\n this.listenTargets.forEach((targetData, targetId) => {\n this.sendWatchRequest(targetData);\n });\n }\n\n private async onWatchStreamClose(error?: FirestoreError): Promise<void> {\n if (error === undefined) {\n // Graceful stop (due to stop() or idle timeout). Make sure that's\n // desirable.\n debugAssert(\n !this.shouldStartWatchStream(),\n 'Watch stream was stopped gracefully while still needed.'\n );\n }\n\n this.cleanUpWatchStreamState();\n\n // If we still need the watch stream, retry the connection.\n if (this.shouldStartWatchStream()) {\n this.onlineStateTracker.handleWatchStreamFailure(error!);\n\n this.startWatchStream();\n } else {\n // No need to restart watch stream because there are no active targets.\n // The online state is set to unknown because there is no active attempt\n // at establishing a connection\n this.onlineStateTracker.set(OnlineState.Unknown);\n }\n }\n\n private async onWatchStreamChange(\n watchChange: WatchChange,\n snapshotVersion: SnapshotVersion\n ): Promise<void> {\n // Mark the client as online since we got a message from the server\n this.onlineStateTracker.set(OnlineState.Online);\n\n if (\n watchChange instanceof WatchTargetChange &&\n watchChange.state === WatchTargetChangeState.Removed &&\n watchChange.cause\n ) {\n // There was an error on a target, don't wait for a consistent snapshot\n // to raise events\n try {\n await this.handleTargetError(watchChange);\n } catch (e) {\n logDebug(\n LOG_TAG,\n 'Failed to remove targets %s: %s ',\n watchChange.targetIds.join(','),\n e\n );\n await this.disableNetworkUntilRecovery(e);\n }\n return;\n }\n\n if (watchChange instanceof DocumentWatchChange) {\n this.watchChangeAggregator!.handleDocumentChange(watchChange);\n } else if (watchChange instanceof ExistenceFilterChange) {\n this.watchChangeAggregator!.handleExistenceFilter(watchChange);\n } else {\n debugAssert(\n watchChange instanceof WatchTargetChange,\n 'Expected watchChange to be an instance of WatchTargetChange'\n );\n this.watchChangeAggregator!.handleTargetChange(watchChange);\n }\n\n if (!snapshotVersion.isEqual(SnapshotVersion.min())) {\n try {\n const lastRemoteSnapshotVersion = await this.localStore.getLastRemoteSnapshotVersion();\n if (snapshotVersion.compareTo(lastRemoteSnapshotVersion) >= 0) {\n // We have received a target change with a global snapshot if the snapshot\n // version is not equal to SnapshotVersion.min().\n await this.raiseWatchSnapshot(snapshotVersion);\n }\n } catch (e) {\n logDebug(LOG_TAG, 'Failed to raise snapshot:', e);\n await this.disableNetworkUntilRecovery(e);\n }\n }\n }\n\n /**\n * Recovery logic for IndexedDB errors that takes the network offline until\n * IndexedDb probing succeeds. Retries are scheduled with backoff using\n * `enqueueRetryable()`.\n */\n private async disableNetworkUntilRecovery(e: FirestoreError): Promise<void> {\n if (isIndexedDbTransactionError(e)) {\n debugAssert(\n !this.indexedDbFailed,\n 'Unexpected network event when IndexedDB was marked failed.'\n );\n this.indexedDbFailed = true;\n\n // Disable network and raise offline snapshots\n await this.disableNetworkInternal();\n this.onlineStateTracker.set(OnlineState.Offline);\n\n // Probe IndexedDB periodically and re-enable network\n this.asyncQueue.enqueueRetryable(async () => {\n logDebug(LOG_TAG, 'Retrying IndexedDB access');\n // Issue a simple read operation to determine if IndexedDB recovered.\n // Ideally, we would expose a health check directly on SimpleDb, but\n // RemoteStore only has access to persistence through LocalStore.\n await this.localStore.getLastRemoteSnapshotVersion();\n this.indexedDbFailed = false;\n await this.enableNetworkInternal();\n });\n } else {\n throw e;\n }\n }\n\n /**\n * Takes a batch of changes from the Datastore, repackages them as a\n * RemoteEvent, and passes that on to the listener, which is typically the\n * SyncEngine.\n */\n private raiseWatchSnapshot(snapshotVersion: SnapshotVersion): Promise<void> {\n debugAssert(\n !snapshotVersion.isEqual(SnapshotVersion.min()),\n \"Can't raise event for unknown SnapshotVersion\"\n );\n const remoteEvent = this.watchChangeAggregator!.createRemoteEvent(\n snapshotVersion\n );\n\n // Update in-memory resume tokens. LocalStore will update the\n // persistent view of these when applying the completed RemoteEvent.\n remoteEvent.targetChanges.forEach((change, targetId) => {\n if (change.resumeToken.approximateByteSize() > 0) {\n const targetData = this.listenTargets.get(targetId);\n // A watched target might have been removed already.\n if (targetData) {\n this.listenTargets.set(\n targetId,\n targetData.withResumeToken(change.resumeToken, snapshotVersion)\n );\n }\n }\n });\n\n // Re-establish listens for the targets that have been invalidated by\n // existence filter mismatches.\n remoteEvent.targetMismatches.forEach(targetId => {\n const targetData = this.listenTargets.get(targetId);\n if (!targetData) {\n // A watched target might have been removed already.\n return;\n }\n\n // Clear the resume token for the target, since we're in a known mismatch\n // state.\n this.listenTargets.set(\n targetId,\n targetData.withResumeToken(\n ByteString.EMPTY_BYTE_STRING,\n targetData.snapshotVersion\n )\n );\n\n // Cause a hard reset by unwatching and rewatching immediately, but\n // deliberately don't send a resume token so that we get a full update.\n this.sendUnwatchRequest(targetId);\n\n // Mark the target we send as being on behalf of an existence filter\n // mismatch, but don't actually retain that in listenTargets. This ensures\n // that we flag the first re-listen this way without impacting future\n // listens of this target (that might happen e.g. on reconnect).\n const requestTargetData = new TargetData(\n targetData.target,\n targetId,\n TargetPurpose.ExistenceFilterMismatch,\n targetData.sequenceNumber\n );\n this.sendWatchRequest(requestTargetData);\n });\n\n // Finally raise remote event\n return this.syncEngine.applyRemoteEvent(remoteEvent);\n }\n\n /** Handles an error on a target */\n private async handleTargetError(\n watchChange: WatchTargetChange\n ): Promise<void> {\n debugAssert(!!watchChange.cause, 'Handling target error without a cause');\n const error = watchChange.cause!;\n for (const targetId of watchChange.targetIds) {\n // A watched target might have been removed already.\n if (this.listenTargets.has(targetId)) {\n await this.syncEngine.rejectListen(targetId, error);\n this.listenTargets.delete(targetId);\n this.watchChangeAggregator!.removeTarget(targetId);\n }\n }\n }\n\n /**\n * Attempts to fill our write pipeline with writes from the LocalStore.\n *\n * Called internally to bootstrap or refill the write pipeline and by\n * SyncEngine whenever there are new mutations to process.\n *\n * Starts the write stream if necessary.\n */\n async fillWritePipeline(): Promise<void> {\n if (this.canAddToWritePipeline()) {\n const lastBatchIdRetrieved =\n this.writePipeline.length > 0\n ? this.writePipeline[this.writePipeline.length - 1].batchId\n : BATCHID_UNKNOWN;\n const batch = await this.localStore.nextMutationBatch(\n lastBatchIdRetrieved\n );\n\n if (batch === null) {\n if (this.writePipeline.length === 0) {\n this.writeStream.markIdle();\n }\n } else {\n this.addToWritePipeline(batch);\n await this.fillWritePipeline();\n }\n }\n\n if (this.shouldStartWriteStream()) {\n this.startWriteStream();\n }\n }\n\n /**\n * Returns true if we can add to the write pipeline (i.e. the network is\n * enabled and the write pipeline is not full).\n */\n private canAddToWritePipeline(): boolean {\n return (\n this.canUseNetwork() && this.writePipeline.length < MAX_PENDING_WRITES\n );\n }\n\n // For testing\n outstandingWrites(): number {\n return this.writePipeline.length;\n }\n\n /**\n * Queues additional writes to be sent to the write stream, sending them\n * immediately if the write stream is established.\n */\n private addToWritePipeline(batch: MutationBatch): void {\n debugAssert(\n this.canAddToWritePipeline(),\n 'addToWritePipeline called when pipeline is full'\n );\n this.writePipeline.push(batch);\n\n if (this.writeStream.isOpen() && this.writeStream.handshakeComplete) {\n this.writeStream.writeMutations(batch.mutations);\n }\n }\n\n private shouldStartWriteStream(): boolean {\n return (\n this.canUseNetwork() &&\n !this.writeStream.isStarted() &&\n this.writePipeline.length > 0\n );\n }\n\n private startWriteStream(): void {\n debugAssert(\n this.shouldStartWriteStream(),\n 'startWriteStream() called when shouldStartWriteStream() is false.'\n );\n this.writeStream.start();\n }\n\n private async onWriteStreamOpen(): Promise<void> {\n this.writeStream.writeHandshake();\n }\n\n private onWriteHandshakeComplete(): Promise<void> {\n // Record the stream token.\n return this.localStore\n .setLastStreamToken(this.writeStream.lastStreamToken)\n .then(() => {\n // Send the write pipeline now that the stream is established.\n for (const batch of this.writePipeline) {\n this.writeStream.writeMutations(batch.mutations);\n }\n })\n .catch(ignoreIfPrimaryLeaseLoss);\n }\n\n private onMutationResult(\n commitVersion: SnapshotVersion,\n results: MutationResult[]\n ): Promise<void> {\n // This is a response to a write containing mutations and should be\n // correlated to the first write in our write pipeline.\n debugAssert(\n this.writePipeline.length > 0,\n 'Got result for empty write pipeline'\n );\n const batch = this.writePipeline.shift()!;\n const success = MutationBatchResult.from(\n batch,\n commitVersion,\n results,\n this.writeStream.lastStreamToken\n );\n return this.syncEngine.applySuccessfulWrite(success).then(() => {\n // It's possible that with the completion of this mutation another\n // slot has freed up.\n return this.fillWritePipeline();\n });\n }\n\n private async onWriteStreamClose(error?: FirestoreError): Promise<void> {\n if (error === undefined) {\n // Graceful stop (due to stop() or idle timeout). Make sure that's\n // desirable.\n debugAssert(\n !this.shouldStartWriteStream(),\n 'Write stream was stopped gracefully while still needed.'\n );\n }\n\n // If the write stream closed due to an error, invoke the error callbacks if\n // there are pending writes.\n if (error && this.writePipeline.length > 0) {\n if (this.writeStream.handshakeComplete) {\n // This error affects the actual write.\n await this.handleWriteError(error!);\n } else {\n // If there was an error before the handshake has finished, it's\n // possible that the server is unable to process the stream token\n // we're sending. (Perhaps it's too old?)\n await this.handleHandshakeError(error!);\n }\n\n // The write stream might have been started by refilling the write\n // pipeline for failed writes\n if (this.shouldStartWriteStream()) {\n this.startWriteStream();\n }\n }\n // No pending writes, nothing to do\n }\n\n private async handleHandshakeError(error: FirestoreError): Promise<void> {\n // Reset the token if it's a permanent error, signaling the write stream is\n // no longer valid. Note that the handshake does not count as a write: see\n // comments on isPermanentWriteError for details.\n if (isPermanentError(error.code)) {\n logDebug(\n LOG_TAG,\n 'RemoteStore error before completed handshake; resetting stream token: ',\n this.writeStream.lastStreamToken\n );\n this.writeStream.lastStreamToken = ByteString.EMPTY_BYTE_STRING;\n\n return this.localStore\n .setLastStreamToken(ByteString.EMPTY_BYTE_STRING)\n .catch(ignoreIfPrimaryLeaseLoss);\n } else {\n // Some other error, don't reset stream token. Our stream logic will\n // just retry with exponential backoff.\n }\n }\n\n private async handleWriteError(error: FirestoreError): Promise<void> {\n // Only handle permanent errors here. If it's transient, just let the retry\n // logic kick in.\n if (isPermanentWriteError(error.code)) {\n // This was a permanent error, the request itself was the problem\n // so it's not going to succeed if we resend it.\n const batch = this.writePipeline.shift()!;\n\n // In this case it's also unlikely that the server itself is melting\n // down -- this was just a bad request so inhibit backoff on the next\n // restart.\n this.writeStream.inhibitBackoff();\n\n return this.syncEngine\n .rejectFailedWrite(batch.batchId, error)\n .then(() => {\n // It's possible that with the completion of this mutation\n // another slot has freed up.\n return this.fillWritePipeline();\n });\n } else {\n // Transient error, just let the retry logic kick in.\n }\n }\n\n createTransaction(): Transaction {\n return new Transaction(this.datastore);\n }\n\n private async restartNetwork(): Promise<void> {\n this.networkEnabled = false;\n await this.disableNetworkInternal();\n this.onlineStateTracker.set(OnlineState.Unknown);\n await this.enableNetwork();\n }\n\n async handleCredentialChange(): Promise<void> {\n if (this.canUseNetwork()) {\n // Tear down and re-create our network streams. This will ensure we get a fresh auth token\n // for the new user and re-fill the write pipeline with new mutations from the LocalStore\n // (since mutations are per-user).\n logDebug(LOG_TAG, 'RemoteStore restarting streams for new credential');\n await this.restartNetwork();\n }\n }\n\n /**\n * Toggles the network state when the client gains or loses its primary lease.\n */\n async applyPrimaryState(isPrimary: boolean): Promise<void> {\n this.isPrimary = isPrimary;\n\n if (isPrimary && this.networkEnabled) {\n await this.enableNetwork();\n } else if (!isPrimary) {\n await this.disableNetworkInternal();\n this.onlineStateTracker.set(OnlineState.Unknown);\n }\n }\n}\n","/**\n * @license\n * Copyright 2018 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { User } from '../auth/user';\nimport { ListenSequence } from '../core/listen_sequence';\nimport {\n BatchId,\n ListenSequenceNumber,\n MutationBatchState,\n OnlineState,\n TargetId\n} from '../core/types';\nimport { TargetIdSet, targetIdSet } from '../model/collections';\nimport { Platform } from '../platform/platform';\nimport { hardAssert, debugAssert } from '../util/assert';\nimport { AsyncQueue } from '../util/async_queue';\nimport { Code, FirestoreError } from '../util/error';\nimport { logError, logDebug } from '../util/log';\nimport { SortedSet } from '../util/sorted_set';\nimport { SortedMap } from '../util/sorted_map';\nimport { primitiveComparator } from '../util/misc';\nimport { isSafeInteger } from '../util/types';\nimport {\n QueryTargetState,\n SharedClientStateSyncer\n} from './shared_client_state_syncer';\nimport {\n CLIENT_STATE_KEY_PREFIX,\n ClientStateSchema,\n createWebStorageClientStateKey,\n createWebStorageMutationBatchKey,\n createWebStorageOnlineStateKey,\n createWebStorageQueryTargetMetadataKey,\n createWebStorageSequenceNumberKey,\n MUTATION_BATCH_KEY_PREFIX,\n MutationMetadataSchema,\n QUERY_TARGET_KEY_PREFIX,\n QueryTargetStateSchema,\n SharedOnlineStateSchema\n} from './shared_client_state_schema';\n\nconst LOG_TAG = 'SharedClientState';\n\n/**\n * A randomly-generated key assigned to each Firestore instance at startup.\n */\nexport type ClientId = string;\n\n/**\n * A `SharedClientState` keeps track of the global state of the mutations\n * and query targets for all active clients with the same persistence key (i.e.\n * project ID and FirebaseApp name). It relays local changes to other clients\n * and updates its local state as new state is observed.\n *\n * `SharedClientState` is primarily used for synchronization in Multi-Tab\n * environments. Each tab is responsible for registering its active query\n * targets and mutations. `SharedClientState` will then notify the listener\n * assigned to `.syncEngine` for updates to mutations and queries that\n * originated in other clients.\n *\n * To receive notifications, `.syncEngine` and `.onlineStateHandler` has to be\n * assigned before calling `start()`.\n */\nexport interface SharedClientState {\n syncEngine: SharedClientStateSyncer | null;\n onlineStateHandler: ((onlineState: OnlineState) => void) | null;\n sequenceNumberHandler:\n | ((sequenceNumber: ListenSequenceNumber) => void)\n | null;\n\n /** Registers the Mutation Batch ID of a newly pending mutation. */\n addPendingMutation(batchId: BatchId): void;\n\n /**\n * Records that a pending mutation has been acknowledged or rejected.\n * Called by the primary client to notify secondary clients of mutation\n * results as they come back from the backend.\n */\n updateMutationState(\n batchId: BatchId,\n state: 'acknowledged' | 'rejected',\n error?: FirestoreError\n ): void;\n\n /**\n * Associates a new Query Target ID with the local Firestore client. Returns\n * the new query state for the query (which can be 'current' if the query is\n * already associated with another tab).\n *\n * If the target id is already associated with local client, the method simply\n * returns its `QueryTargetState`.\n */\n addLocalQueryTarget(targetId: TargetId): QueryTargetState;\n\n /** Removes the Query Target ID association from the local client. */\n removeLocalQueryTarget(targetId: TargetId): void;\n\n /** Checks whether the target is associated with the local client. */\n isLocalQueryTarget(targetId: TargetId): boolean;\n\n /**\n * Processes an update to a query target.\n *\n * Called by the primary client to notify secondary clients of document\n * changes or state transitions that affect the provided query target.\n */\n updateQueryState(\n targetId: TargetId,\n state: QueryTargetState,\n error?: FirestoreError\n ): void;\n\n /**\n * Removes the target's metadata entry.\n *\n * Called by the primary client when all clients stopped listening to a query\n * target.\n */\n clearQueryState(targetId: TargetId): void;\n\n /**\n * Gets the active Query Targets IDs for all active clients.\n *\n * The implementation for this may require O(n) runtime, where 'n' is the size\n * of the result set.\n */\n // Visible for testing\n getAllActiveQueryTargets(): SortedSet<TargetId>;\n\n /**\n * Checks whether the provided target ID is currently being listened to by\n * any of the active clients.\n *\n * The implementation may require O(n*log m) runtime, where 'n' is the number\n * of clients and 'm' the number of targets.\n */\n isActiveQueryTarget(targetId: TargetId): boolean;\n\n /**\n * Starts the SharedClientState, reads existing client data and registers\n * listeners for updates to new and existing clients.\n */\n start(): Promise<void>;\n\n /** Shuts down the `SharedClientState` and its listeners. */\n shutdown(): void;\n\n /**\n * Changes the active user and removes all existing user-specific data. The\n * user change does not call back into SyncEngine (for example, no mutations\n * will be marked as removed).\n */\n handleUserChange(\n user: User,\n removedBatchIds: BatchId[],\n addedBatchIds: BatchId[]\n ): void;\n\n /** Changes the shared online state of all clients. */\n setOnlineState(onlineState: OnlineState): void;\n\n writeSequenceNumber(sequenceNumber: ListenSequenceNumber): void;\n}\n\n/**\n * Holds the state of a mutation batch, including its user ID, batch ID and\n * whether the batch is 'pending', 'acknowledged' or 'rejected'.\n */\n// Visible for testing\nexport class MutationMetadata {\n constructor(\n readonly user: User,\n readonly batchId: BatchId,\n readonly state: MutationBatchState,\n readonly error?: FirestoreError\n ) {\n debugAssert(\n (error !== undefined) === (state === 'rejected'),\n `MutationMetadata must contain an error iff state is 'rejected'`\n );\n }\n\n /**\n * Parses a MutationMetadata from its JSON representation in WebStorage.\n * Logs a warning and returns null if the format of the data is not valid.\n */\n static fromWebStorageEntry(\n user: User,\n batchId: BatchId,\n value: string\n ): MutationMetadata | null {\n const mutationBatch = JSON.parse(value) as MutationMetadataSchema;\n\n let validData =\n typeof mutationBatch === 'object' &&\n ['pending', 'acknowledged', 'rejected'].indexOf(mutationBatch.state) !==\n -1 &&\n (mutationBatch.error === undefined ||\n typeof mutationBatch.error === 'object');\n\n let firestoreError: FirestoreError | undefined = undefined;\n\n if (validData && mutationBatch.error) {\n validData =\n typeof mutationBatch.error.message === 'string' &&\n typeof mutationBatch.error.code === 'string';\n if (validData) {\n firestoreError = new FirestoreError(\n mutationBatch.error.code as Code,\n mutationBatch.error.message\n );\n }\n }\n\n if (validData) {\n return new MutationMetadata(\n user,\n batchId,\n mutationBatch.state,\n firestoreError\n );\n } else {\n logError(\n LOG_TAG,\n `Failed to parse mutation state for ID '${batchId}': ${value}`\n );\n return null;\n }\n }\n\n toWebStorageJSON(): string {\n const batchMetadata: MutationMetadataSchema = {\n state: this.state,\n updateTimeMs: Date.now() // Modify the existing value to trigger update.\n };\n\n if (this.error) {\n batchMetadata.error = {\n code: this.error.code,\n message: this.error.message\n };\n }\n\n return JSON.stringify(batchMetadata);\n }\n}\n\n/**\n * Holds the state of a query target, including its target ID and whether the\n * target is 'not-current', 'current' or 'rejected'.\n */\n// Visible for testing\nexport class QueryTargetMetadata {\n constructor(\n readonly targetId: TargetId,\n readonly state: QueryTargetState,\n readonly error?: FirestoreError\n ) {\n debugAssert(\n (error !== undefined) === (state === 'rejected'),\n `QueryTargetMetadata must contain an error iff state is 'rejected'`\n );\n }\n\n /**\n * Parses a QueryTargetMetadata from its JSON representation in WebStorage.\n * Logs a warning and returns null if the format of the data is not valid.\n */\n static fromWebStorageEntry(\n targetId: TargetId,\n value: string\n ): QueryTargetMetadata | null {\n const targetState = JSON.parse(value) as QueryTargetStateSchema;\n\n let validData =\n typeof targetState === 'object' &&\n ['not-current', 'current', 'rejected'].indexOf(targetState.state) !==\n -1 &&\n (targetState.error === undefined ||\n typeof targetState.error === 'object');\n\n let firestoreError: FirestoreError | undefined = undefined;\n\n if (validData && targetState.error) {\n validData =\n typeof targetState.error.message === 'string' &&\n typeof targetState.error.code === 'string';\n if (validData) {\n firestoreError = new FirestoreError(\n targetState.error.code as Code,\n targetState.error.message\n );\n }\n }\n\n if (validData) {\n return new QueryTargetMetadata(\n targetId,\n targetState.state,\n firestoreError\n );\n } else {\n logError(\n LOG_TAG,\n `Failed to parse target state for ID '${targetId}': ${value}`\n );\n return null;\n }\n }\n\n toWebStorageJSON(): string {\n const targetState: QueryTargetStateSchema = {\n state: this.state,\n updateTimeMs: Date.now() // Modify the existing value to trigger update.\n };\n\n if (this.error) {\n targetState.error = {\n code: this.error.code,\n message: this.error.message\n };\n }\n\n return JSON.stringify(targetState);\n }\n}\n\n/**\n * Metadata state of a single client denoting the query targets it is actively\n * listening to.\n */\n// Visible for testing.\nexport interface ClientState {\n readonly activeTargetIds: TargetIdSet;\n}\n\n/**\n * This class represents the immutable ClientState for a client read from\n * WebStorage, containing the list of active query targets.\n */\nclass RemoteClientState implements ClientState {\n private constructor(\n readonly clientId: ClientId,\n readonly activeTargetIds: TargetIdSet\n ) {}\n\n /**\n * Parses a RemoteClientState from the JSON representation in WebStorage.\n * Logs a warning and returns null if the format of the data is not valid.\n */\n static fromWebStorageEntry(\n clientId: ClientId,\n value: string\n ): RemoteClientState | null {\n const clientState = JSON.parse(value) as ClientStateSchema;\n\n let validData =\n typeof clientState === 'object' &&\n clientState.activeTargetIds instanceof Array;\n\n let activeTargetIdsSet = targetIdSet();\n\n for (let i = 0; validData && i < clientState.activeTargetIds.length; ++i) {\n validData = isSafeInteger(clientState.activeTargetIds[i]);\n activeTargetIdsSet = activeTargetIdsSet.add(\n clientState.activeTargetIds[i]\n );\n }\n\n if (validData) {\n return new RemoteClientState(clientId, activeTargetIdsSet);\n } else {\n logError(\n LOG_TAG,\n `Failed to parse client data for instance '${clientId}': ${value}`\n );\n return null;\n }\n }\n}\n\n/**\n * This class represents the online state for all clients participating in\n * multi-tab. The online state is only written to by the primary client, and\n * used in secondary clients to update their query views.\n */\nexport class SharedOnlineState {\n constructor(readonly clientId: string, readonly onlineState: OnlineState) {}\n\n /**\n * Parses a SharedOnlineState from its JSON representation in WebStorage.\n * Logs a warning and returns null if the format of the data is not valid.\n */\n static fromWebStorageEntry(value: string): SharedOnlineState | null {\n const onlineState = JSON.parse(value) as SharedOnlineStateSchema;\n\n const validData =\n typeof onlineState === 'object' &&\n ['Unknown', 'Online', 'Offline'].indexOf(onlineState.onlineState) !==\n -1 &&\n typeof onlineState.clientId === 'string';\n\n if (validData) {\n return new SharedOnlineState(\n onlineState.clientId,\n onlineState.onlineState as OnlineState\n );\n } else {\n logError(LOG_TAG, `Failed to parse online state: ${value}`);\n return null;\n }\n }\n}\n\n/**\n * Metadata state of the local client. Unlike `RemoteClientState`, this class is\n * mutable and keeps track of all pending mutations, which allows us to\n * update the range of pending mutation batch IDs as new mutations are added or\n * removed.\n *\n * The data in `LocalClientState` is not read from WebStorage and instead\n * updated via its instance methods. The updated state can be serialized via\n * `toWebStorageJSON()`.\n */\n// Visible for testing.\nexport class LocalClientState implements ClientState {\n activeTargetIds = targetIdSet();\n\n addQueryTarget(targetId: TargetId): void {\n this.activeTargetIds = this.activeTargetIds.add(targetId);\n }\n\n removeQueryTarget(targetId: TargetId): void {\n this.activeTargetIds = this.activeTargetIds.delete(targetId);\n }\n\n /**\n * Converts this entry into a JSON-encoded format we can use for WebStorage.\n * Does not encode `clientId` as it is part of the key in WebStorage.\n */\n toWebStorageJSON(): string {\n const data: ClientStateSchema = {\n activeTargetIds: this.activeTargetIds.toArray(),\n updateTimeMs: Date.now() // Modify the existing value to trigger update.\n };\n return JSON.stringify(data);\n }\n}\n\n/**\n * `WebStorageSharedClientState` uses WebStorage (window.localStorage) as the\n * backing store for the SharedClientState. It keeps track of all active\n * clients and supports modifications of the local client's data.\n */\nexport class WebStorageSharedClientState implements SharedClientState {\n syncEngine: SharedClientStateSyncer | null = null;\n onlineStateHandler: ((onlineState: OnlineState) => void) | null = null;\n sequenceNumberHandler:\n | ((sequenceNumber: ListenSequenceNumber) => void)\n | null = null;\n\n private readonly storage: Storage;\n private readonly localClientStorageKey: string;\n private readonly sequenceNumberKey: string;\n private readonly storageListener = this.handleWebStorageEvent.bind(this);\n private readonly onlineStateKey: string;\n private readonly clientStateKeyRe: RegExp;\n private readonly mutationBatchKeyRe: RegExp;\n private readonly queryTargetKeyRe: RegExp;\n private activeClients = new SortedMap<string, ClientState>(\n primitiveComparator\n );\n private started = false;\n private currentUser: User;\n\n /**\n * Captures WebStorage events that occur before `start()` is called. These\n * events are replayed once `WebStorageSharedClientState` is started.\n */\n private earlyEvents: StorageEvent[] = [];\n\n constructor(\n private readonly queue: AsyncQueue,\n private readonly platform: Platform,\n private readonly persistenceKey: string,\n private readonly localClientId: ClientId,\n initialUser: User\n ) {\n if (!WebStorageSharedClientState.isAvailable(this.platform)) {\n throw new FirestoreError(\n Code.UNIMPLEMENTED,\n 'LocalStorage is not available on this platform.'\n );\n }\n // Escape the special characters mentioned here:\n // https://developer.mozilla.org/en-US/docs/Web/JavaScript/Guide/Regular_Expressions\n const escapedPersistenceKey = persistenceKey.replace(\n /[.*+?^${}()|[\\]\\\\]/g,\n '\\\\$&'\n );\n\n this.storage = this.platform.window!.localStorage;\n this.currentUser = initialUser;\n this.localClientStorageKey = createWebStorageClientStateKey(\n this.persistenceKey,\n this.localClientId\n );\n this.sequenceNumberKey = createWebStorageSequenceNumberKey(\n this.persistenceKey\n );\n this.activeClients = this.activeClients.insert(\n this.localClientId,\n new LocalClientState()\n );\n\n this.clientStateKeyRe = new RegExp(\n `^${CLIENT_STATE_KEY_PREFIX}_${escapedPersistenceKey}_([^_]*)$`\n );\n this.mutationBatchKeyRe = new RegExp(\n `^${MUTATION_BATCH_KEY_PREFIX}_${escapedPersistenceKey}_(\\\\d+)(?:_(.*))?$`\n );\n this.queryTargetKeyRe = new RegExp(\n `^${QUERY_TARGET_KEY_PREFIX}_${escapedPersistenceKey}_(\\\\d+)$`\n );\n\n this.onlineStateKey = createWebStorageOnlineStateKey(this.persistenceKey);\n\n // Rather than adding the storage observer during start(), we add the\n // storage observer during initialization. This ensures that we collect\n // events before other components populate their initial state (during their\n // respective start() calls). Otherwise, we might for example miss a\n // mutation that is added after LocalStore's start() processed the existing\n // mutations but before we observe WebStorage events.\n this.platform.window!.addEventListener('storage', this.storageListener);\n }\n\n /** Returns 'true' if WebStorage is available in the current environment. */\n static isAvailable(platform: Platform): boolean {\n return !!(platform.window && platform.window.localStorage != null);\n }\n\n async start(): Promise<void> {\n debugAssert(!this.started, 'WebStorageSharedClientState already started');\n debugAssert(\n this.syncEngine !== null,\n 'syncEngine property must be set before calling start()'\n );\n debugAssert(\n this.onlineStateHandler !== null,\n 'onlineStateHandler property must be set before calling start()'\n );\n\n // Retrieve the list of existing clients to backfill the data in\n // SharedClientState.\n const existingClients = await this.syncEngine!.getActiveClients();\n\n for (const clientId of existingClients) {\n if (clientId === this.localClientId) {\n continue;\n }\n\n const storageItem = this.getItem(\n createWebStorageClientStateKey(this.persistenceKey, clientId)\n );\n if (storageItem) {\n const clientState = RemoteClientState.fromWebStorageEntry(\n clientId,\n storageItem\n );\n if (clientState) {\n this.activeClients = this.activeClients.insert(\n clientState.clientId,\n clientState\n );\n }\n }\n }\n\n this.persistClientState();\n\n // Check if there is an existing online state and call the callback handler\n // if applicable.\n const onlineStateJSON = this.storage.getItem(this.onlineStateKey);\n if (onlineStateJSON) {\n const onlineState = this.fromWebStorageOnlineState(onlineStateJSON);\n if (onlineState) {\n this.handleOnlineStateEvent(onlineState);\n }\n }\n\n for (const event of this.earlyEvents) {\n this.handleWebStorageEvent(event);\n }\n\n this.earlyEvents = [];\n\n // Register a window unload hook to remove the client metadata entry from\n // WebStorage even if `shutdown()` was not called.\n this.platform.window!.addEventListener('unload', () => this.shutdown());\n\n this.started = true;\n }\n\n writeSequenceNumber(sequenceNumber: ListenSequenceNumber): void {\n this.setItem(this.sequenceNumberKey, JSON.stringify(sequenceNumber));\n }\n\n getAllActiveQueryTargets(): TargetIdSet {\n return this.extractActiveQueryTargets(this.activeClients);\n }\n\n isActiveQueryTarget(targetId: TargetId): boolean {\n let found = false;\n this.activeClients.forEach((key, value) => {\n if (value.activeTargetIds.has(targetId)) {\n found = true;\n }\n });\n return found;\n }\n\n addPendingMutation(batchId: BatchId): void {\n this.persistMutationState(batchId, 'pending');\n }\n\n updateMutationState(\n batchId: BatchId,\n state: 'acknowledged' | 'rejected',\n error?: FirestoreError\n ): void {\n this.persistMutationState(batchId, state, error);\n\n // Once a final mutation result is observed by other clients, they no longer\n // access the mutation's metadata entry. Since WebStorage replays events\n // in order, it is safe to delete the entry right after updating it.\n this.removeMutationState(batchId);\n }\n\n addLocalQueryTarget(targetId: TargetId): QueryTargetState {\n let queryState: QueryTargetState = 'not-current';\n\n // Lookup an existing query state if the target ID was already registered\n // by another tab\n if (this.isActiveQueryTarget(targetId)) {\n const storageItem = this.storage.getItem(\n createWebStorageQueryTargetMetadataKey(this.persistenceKey, targetId)\n );\n\n if (storageItem) {\n const metadata = QueryTargetMetadata.fromWebStorageEntry(\n targetId,\n storageItem\n );\n if (metadata) {\n queryState = metadata.state;\n }\n }\n }\n\n this.localClientState.addQueryTarget(targetId);\n this.persistClientState();\n\n return queryState;\n }\n\n removeLocalQueryTarget(targetId: TargetId): void {\n this.localClientState.removeQueryTarget(targetId);\n this.persistClientState();\n }\n\n isLocalQueryTarget(targetId: TargetId): boolean {\n return this.localClientState.activeTargetIds.has(targetId);\n }\n\n clearQueryState(targetId: TargetId): void {\n this.removeItem(\n createWebStorageQueryTargetMetadataKey(this.persistenceKey, targetId)\n );\n }\n\n updateQueryState(\n targetId: TargetId,\n state: QueryTargetState,\n error?: FirestoreError\n ): void {\n this.persistQueryTargetState(targetId, state, error);\n }\n\n handleUserChange(\n user: User,\n removedBatchIds: BatchId[],\n addedBatchIds: BatchId[]\n ): void {\n removedBatchIds.forEach(batchId => {\n this.removeMutationState(batchId);\n });\n this.currentUser = user;\n addedBatchIds.forEach(batchId => {\n this.addPendingMutation(batchId);\n });\n }\n\n setOnlineState(onlineState: OnlineState): void {\n this.persistOnlineState(onlineState);\n }\n\n shutdown(): void {\n if (this.started) {\n this.platform.window!.removeEventListener(\n 'storage',\n this.storageListener\n );\n this.removeItem(this.localClientStorageKey);\n this.started = false;\n }\n }\n\n private getItem(key: string): string | null {\n const value = this.storage.getItem(key);\n logDebug(LOG_TAG, 'READ', key, value);\n return value;\n }\n\n private setItem(key: string, value: string): void {\n logDebug(LOG_TAG, 'SET', key, value);\n this.storage.setItem(key, value);\n }\n\n private removeItem(key: string): void {\n logDebug(LOG_TAG, 'REMOVE', key);\n this.storage.removeItem(key);\n }\n\n private handleWebStorageEvent(event: StorageEvent): void {\n if (event.storageArea === this.storage) {\n logDebug(LOG_TAG, 'EVENT', event.key, event.newValue);\n\n if (event.key === this.localClientStorageKey) {\n logError(\n 'Received WebStorage notification for local change. Another client might have ' +\n 'garbage-collected our state'\n );\n return;\n }\n\n this.queue.enqueueRetryable(async () => {\n if (!this.started) {\n this.earlyEvents.push(event);\n return;\n }\n\n if (event.key === null) {\n return;\n }\n\n if (this.clientStateKeyRe.test(event.key)) {\n if (event.newValue != null) {\n const clientState = this.fromWebStorageClientState(\n event.key,\n event.newValue\n );\n if (clientState) {\n return this.handleClientStateEvent(\n clientState.clientId,\n clientState\n );\n }\n } else {\n const clientId = this.fromWebStorageClientStateKey(event.key)!;\n return this.handleClientStateEvent(clientId, null);\n }\n } else if (this.mutationBatchKeyRe.test(event.key)) {\n if (event.newValue !== null) {\n const mutationMetadata = this.fromWebStorageMutationMetadata(\n event.key,\n event.newValue\n );\n if (mutationMetadata) {\n return this.handleMutationBatchEvent(mutationMetadata);\n }\n }\n } else if (this.queryTargetKeyRe.test(event.key)) {\n if (event.newValue !== null) {\n const queryTargetMetadata = this.fromWebStorageQueryTargetMetadata(\n event.key,\n event.newValue\n );\n if (queryTargetMetadata) {\n return this.handleQueryTargetEvent(queryTargetMetadata);\n }\n }\n } else if (event.key === this.onlineStateKey) {\n if (event.newValue !== null) {\n const onlineState = this.fromWebStorageOnlineState(event.newValue);\n if (onlineState) {\n return this.handleOnlineStateEvent(onlineState);\n }\n }\n } else if (event.key === this.sequenceNumberKey) {\n debugAssert(\n !!this.sequenceNumberHandler,\n 'Missing sequenceNumberHandler'\n );\n const sequenceNumber = fromWebStorageSequenceNumber(event.newValue);\n if (sequenceNumber !== ListenSequence.INVALID) {\n this.sequenceNumberHandler!(sequenceNumber);\n }\n }\n });\n }\n }\n\n private get localClientState(): LocalClientState {\n return this.activeClients.get(this.localClientId) as LocalClientState;\n }\n\n private persistClientState(): void {\n this.setItem(\n this.localClientStorageKey,\n this.localClientState.toWebStorageJSON()\n );\n }\n\n private persistMutationState(\n batchId: BatchId,\n state: MutationBatchState,\n error?: FirestoreError\n ): void {\n const mutationState = new MutationMetadata(\n this.currentUser,\n batchId,\n state,\n error\n );\n const mutationKey = createWebStorageMutationBatchKey(\n this.persistenceKey,\n this.currentUser,\n batchId\n );\n this.setItem(mutationKey, mutationState.toWebStorageJSON());\n }\n\n private removeMutationState(batchId: BatchId): void {\n const mutationKey = createWebStorageMutationBatchKey(\n this.persistenceKey,\n this.currentUser,\n batchId\n );\n this.removeItem(mutationKey);\n }\n\n private persistOnlineState(onlineState: OnlineState): void {\n const entry: SharedOnlineStateSchema = {\n clientId: this.localClientId,\n onlineState\n };\n this.storage.setItem(this.onlineStateKey, JSON.stringify(entry));\n }\n\n private persistQueryTargetState(\n targetId: TargetId,\n state: QueryTargetState,\n error?: FirestoreError\n ): void {\n const targetKey = createWebStorageQueryTargetMetadataKey(\n this.persistenceKey,\n targetId\n );\n const targetMetadata = new QueryTargetMetadata(targetId, state, error);\n this.setItem(targetKey, targetMetadata.toWebStorageJSON());\n }\n\n /**\n * Parses a client state key in WebStorage. Returns null if the key does not\n * match the expected key format.\n */\n private fromWebStorageClientStateKey(key: string): ClientId | null {\n const match = this.clientStateKeyRe.exec(key);\n return match ? match[1] : null;\n }\n\n /**\n * Parses a client state in WebStorage. Returns 'null' if the value could not\n * be parsed.\n */\n private fromWebStorageClientState(\n key: string,\n value: string\n ): RemoteClientState | null {\n const clientId = this.fromWebStorageClientStateKey(key);\n debugAssert(clientId !== null, `Cannot parse client state key '${key}'`);\n return RemoteClientState.fromWebStorageEntry(clientId, value);\n }\n\n /**\n * Parses a mutation batch state in WebStorage. Returns 'null' if the value\n * could not be parsed.\n */\n private fromWebStorageMutationMetadata(\n key: string,\n value: string\n ): MutationMetadata | null {\n const match = this.mutationBatchKeyRe.exec(key);\n debugAssert(match !== null, `Cannot parse mutation batch key '${key}'`);\n\n const batchId = Number(match[1]);\n const userId = match[2] !== undefined ? match[2] : null;\n return MutationMetadata.fromWebStorageEntry(\n new User(userId),\n batchId,\n value\n );\n }\n\n /**\n * Parses a query target state from WebStorage. Returns 'null' if the value\n * could not be parsed.\n */\n private fromWebStorageQueryTargetMetadata(\n key: string,\n value: string\n ): QueryTargetMetadata | null {\n const match = this.queryTargetKeyRe.exec(key);\n debugAssert(match !== null, `Cannot parse query target key '${key}'`);\n\n const targetId = Number(match[1]);\n return QueryTargetMetadata.fromWebStorageEntry(targetId, value);\n }\n\n /**\n * Parses an online state from WebStorage. Returns 'null' if the value\n * could not be parsed.\n */\n private fromWebStorageOnlineState(value: string): SharedOnlineState | null {\n return SharedOnlineState.fromWebStorageEntry(value);\n }\n\n private async handleMutationBatchEvent(\n mutationBatch: MutationMetadata\n ): Promise<void> {\n if (mutationBatch.user.uid !== this.currentUser.uid) {\n logDebug(\n LOG_TAG,\n `Ignoring mutation for non-active user ${mutationBatch.user.uid}`\n );\n return;\n }\n\n return this.syncEngine!.applyBatchState(\n mutationBatch.batchId,\n mutationBatch.state,\n mutationBatch.error\n );\n }\n\n private handleQueryTargetEvent(\n targetMetadata: QueryTargetMetadata\n ): Promise<void> {\n return this.syncEngine!.applyTargetState(\n targetMetadata.targetId,\n targetMetadata.state,\n targetMetadata.error\n );\n }\n\n private handleClientStateEvent(\n clientId: ClientId,\n clientState: RemoteClientState | null\n ): Promise<void> {\n const updatedClients = clientState\n ? this.activeClients.insert(clientId, clientState)\n : this.activeClients.remove(clientId);\n\n const existingTargets = this.extractActiveQueryTargets(this.activeClients);\n const newTargets = this.extractActiveQueryTargets(updatedClients);\n\n const addedTargets: TargetId[] = [];\n const removedTargets: TargetId[] = [];\n\n newTargets.forEach(targetId => {\n if (!existingTargets.has(targetId)) {\n addedTargets.push(targetId);\n }\n });\n\n existingTargets.forEach(targetId => {\n if (!newTargets.has(targetId)) {\n removedTargets.push(targetId);\n }\n });\n\n return this.syncEngine!.applyActiveTargetsChange(\n addedTargets,\n removedTargets\n ).then(() => {\n this.activeClients = updatedClients;\n });\n }\n\n private handleOnlineStateEvent(onlineState: SharedOnlineState): void {\n // We check whether the client that wrote this online state is still active\n // by comparing its client ID to the list of clients kept active in\n // IndexedDb. If a client does not update their IndexedDb client state\n // within 5 seconds, it is considered inactive and we don't emit an online\n // state event.\n if (this.activeClients.get(onlineState.clientId)) {\n this.onlineStateHandler!(onlineState.onlineState);\n }\n }\n\n private extractActiveQueryTargets(\n clients: SortedMap<string, ClientState>\n ): SortedSet<TargetId> {\n let activeTargets = targetIdSet();\n clients.forEach((kev, value) => {\n activeTargets = activeTargets.unionWith(value.activeTargetIds);\n });\n return activeTargets;\n }\n}\n\nfunction fromWebStorageSequenceNumber(\n seqString: string | null\n): ListenSequenceNumber {\n let sequenceNumber = ListenSequence.INVALID;\n if (seqString != null) {\n try {\n const parsed = JSON.parse(seqString);\n hardAssert(\n typeof parsed === 'number',\n 'Found non-numeric sequence number'\n );\n sequenceNumber = parsed;\n } catch (e) {\n logError(LOG_TAG, 'Failed to read sequence number from WebStorage', e);\n }\n }\n return sequenceNumber;\n}\n\n/**\n * `MemorySharedClientState` is a simple implementation of SharedClientState for\n * clients using memory persistence. The state in this class remains fully\n * isolated and no synchronization is performed.\n */\nexport class MemorySharedClientState implements SharedClientState {\n private localState = new LocalClientState();\n private queryState: { [targetId: number]: QueryTargetState } = {};\n\n syncEngine: SharedClientStateSyncer | null = null;\n onlineStateHandler: ((onlineState: OnlineState) => void) | null = null;\n sequenceNumberHandler:\n | ((sequenceNumber: ListenSequenceNumber) => void)\n | null = null;\n\n addPendingMutation(batchId: BatchId): void {\n // No op.\n }\n\n updateMutationState(\n batchId: BatchId,\n state: 'acknowledged' | 'rejected',\n error?: FirestoreError\n ): void {\n // No op.\n }\n\n addLocalQueryTarget(targetId: TargetId): QueryTargetState {\n this.localState.addQueryTarget(targetId);\n return this.queryState[targetId] || 'not-current';\n }\n\n updateQueryState(\n targetId: TargetId,\n state: QueryTargetState,\n error?: FirestoreError\n ): void {\n this.queryState[targetId] = state;\n }\n\n removeLocalQueryTarget(targetId: TargetId): void {\n this.localState.removeQueryTarget(targetId);\n }\n\n isLocalQueryTarget(targetId: TargetId): boolean {\n return this.localState.activeTargetIds.has(targetId);\n }\n\n clearQueryState(targetId: TargetId): void {\n delete this.queryState[targetId];\n }\n\n getAllActiveQueryTargets(): TargetIdSet {\n return this.localState.activeTargetIds;\n }\n\n isActiveQueryTarget(targetId: TargetId): boolean {\n return this.localState.activeTargetIds.has(targetId);\n }\n\n start(): Promise<void> {\n this.localState = new LocalClientState();\n return Promise.resolve();\n }\n\n handleUserChange(\n user: User,\n removedBatchIds: BatchId[],\n addedBatchIds: BatchId[]\n ): void {\n // No op.\n }\n\n setOnlineState(onlineState: OnlineState): void {\n // No op.\n }\n\n shutdown(): void {}\n\n writeSequenceNumber(sequenceNumber: ListenSequenceNumber): void {}\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { QueryResult } from '../local/local_store';\nimport {\n documentKeySet,\n DocumentKeySet,\n MaybeDocumentMap\n} from '../model/collections';\nimport { Document, MaybeDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { DocumentSet } from '../model/document_set';\nimport { TargetChange } from '../remote/remote_event';\nimport { debugAssert, fail } from '../util/assert';\n\nimport { Query } from './query';\nimport { OnlineState } from './types';\nimport {\n ChangeType,\n DocumentChangeSet,\n SyncState,\n ViewSnapshot\n} from './view_snapshot';\n\nexport type LimboDocumentChange = AddedLimboDocument | RemovedLimboDocument;\nexport class AddedLimboDocument {\n constructor(public key: DocumentKey) {}\n}\nexport class RemovedLimboDocument {\n constructor(public key: DocumentKey) {}\n}\n\n/** The result of applying a set of doc changes to a view. */\nexport interface ViewDocumentChanges {\n /** The new set of docs that should be in the view. */\n documentSet: DocumentSet;\n /** The diff of these docs with the previous set of docs. */\n changeSet: DocumentChangeSet;\n /**\n * Whether the set of documents passed in was not sufficient to calculate the\n * new state of the view and there needs to be another pass based on the\n * local cache.\n */\n needsRefill: boolean;\n\n mutatedKeys: DocumentKeySet;\n}\n\nexport interface ViewChange {\n snapshot?: ViewSnapshot;\n limboChanges: LimboDocumentChange[];\n}\n\n/**\n * View is responsible for computing the final merged truth of what docs are in\n * a query. It gets notified of local and remote changes to docs, and applies\n * the query filters and limits to determine the most correct possible results.\n */\nexport class View {\n private syncState: SyncState | null = null;\n /**\n * A flag whether the view is current with the backend. A view is considered\n * current after it has seen the current flag from the backend and did not\n * lose consistency within the watch stream (e.g. because of an existence\n * filter mismatch).\n */\n private current = false;\n private documentSet: DocumentSet;\n /** Documents in the view but not in the remote target */\n private limboDocuments = documentKeySet();\n /** Document Keys that have local changes */\n private mutatedKeys = documentKeySet();\n\n constructor(\n private query: Query,\n /** Documents included in the remote target */\n private _syncedDocuments: DocumentKeySet\n ) {\n this.documentSet = new DocumentSet(query.docComparator.bind(query));\n }\n\n /**\n * The set of remote documents that the server has told us belongs to the target associated with\n * this view.\n */\n get syncedDocuments(): DocumentKeySet {\n return this._syncedDocuments;\n }\n\n /**\n * Iterates over a set of doc changes, applies the query limit, and computes\n * what the new results should be, what the changes were, and whether we may\n * need to go back to the local cache for more results. Does not make any\n * changes to the view.\n * @param docChanges The doc changes to apply to this view.\n * @param previousChanges If this is being called with a refill, then start\n * with this set of docs and changes instead of the current view.\n * @return a new set of docs, changes, and refill flag.\n */\n computeDocChanges(\n docChanges: MaybeDocumentMap,\n previousChanges?: ViewDocumentChanges\n ): ViewDocumentChanges {\n const changeSet = previousChanges\n ? previousChanges.changeSet\n : new DocumentChangeSet();\n const oldDocumentSet = previousChanges\n ? previousChanges.documentSet\n : this.documentSet;\n let newMutatedKeys = previousChanges\n ? previousChanges.mutatedKeys\n : this.mutatedKeys;\n let newDocumentSet = oldDocumentSet;\n let needsRefill = false;\n\n // Track the last doc in a (full) limit. This is necessary, because some\n // update (a delete, or an update moving a doc past the old limit) might\n // mean there is some other document in the local cache that either should\n // come (1) between the old last limit doc and the new last document, in the\n // case of updates, or (2) after the new last document, in the case of\n // deletes. So we keep this doc at the old limit to compare the updates to.\n //\n // Note that this should never get used in a refill (when previousChanges is\n // set), because there will only be adds -- no deletes or updates.\n const lastDocInLimit =\n this.query.hasLimitToFirst() && oldDocumentSet.size === this.query.limit\n ? oldDocumentSet.last()\n : null;\n const firstDocInLimit =\n this.query.hasLimitToLast() && oldDocumentSet.size === this.query.limit\n ? oldDocumentSet.first()\n : null;\n\n docChanges.inorderTraversal(\n (key: DocumentKey, newMaybeDoc: MaybeDocument) => {\n const oldDoc = oldDocumentSet.get(key);\n let newDoc = newMaybeDoc instanceof Document ? newMaybeDoc : null;\n if (newDoc) {\n debugAssert(\n key.isEqual(newDoc.key),\n 'Mismatching keys found in document changes: ' +\n key +\n ' != ' +\n newDoc.key\n );\n newDoc = this.query.matches(newDoc) ? newDoc : null;\n }\n\n const oldDocHadPendingMutations = oldDoc\n ? this.mutatedKeys.has(oldDoc.key)\n : false;\n const newDocHasPendingMutations = newDoc\n ? newDoc.hasLocalMutations ||\n // We only consider committed mutations for documents that were\n // mutated during the lifetime of the view.\n (this.mutatedKeys.has(newDoc.key) && newDoc.hasCommittedMutations)\n : false;\n\n let changeApplied = false;\n\n // Calculate change\n if (oldDoc && newDoc) {\n const docsEqual = oldDoc.data().isEqual(newDoc.data());\n if (!docsEqual) {\n if (!this.shouldWaitForSyncedDocument(oldDoc, newDoc)) {\n changeSet.track({\n type: ChangeType.Modified,\n doc: newDoc\n });\n changeApplied = true;\n\n if (\n (lastDocInLimit &&\n this.query.docComparator(newDoc, lastDocInLimit) > 0) ||\n (firstDocInLimit &&\n this.query.docComparator(newDoc, firstDocInLimit) < 0)\n ) {\n // This doc moved from inside the limit to outside the limit.\n // That means there may be some other doc in the local cache\n // that should be included instead.\n needsRefill = true;\n }\n }\n } else if (oldDocHadPendingMutations !== newDocHasPendingMutations) {\n changeSet.track({ type: ChangeType.Metadata, doc: newDoc });\n changeApplied = true;\n }\n } else if (!oldDoc && newDoc) {\n changeSet.track({ type: ChangeType.Added, doc: newDoc });\n changeApplied = true;\n } else if (oldDoc && !newDoc) {\n changeSet.track({ type: ChangeType.Removed, doc: oldDoc });\n changeApplied = true;\n\n if (lastDocInLimit || firstDocInLimit) {\n // A doc was removed from a full limit query. We'll need to\n // requery from the local cache to see if we know about some other\n // doc that should be in the results.\n needsRefill = true;\n }\n }\n\n if (changeApplied) {\n if (newDoc) {\n newDocumentSet = newDocumentSet.add(newDoc);\n if (newDocHasPendingMutations) {\n newMutatedKeys = newMutatedKeys.add(key);\n } else {\n newMutatedKeys = newMutatedKeys.delete(key);\n }\n } else {\n newDocumentSet = newDocumentSet.delete(key);\n newMutatedKeys = newMutatedKeys.delete(key);\n }\n }\n }\n );\n\n // Drop documents out to meet limit/limitToLast requirement.\n if (this.query.hasLimitToFirst() || this.query.hasLimitToLast()) {\n while (newDocumentSet.size > this.query.limit!) {\n const oldDoc = this.query.hasLimitToFirst()\n ? newDocumentSet.last()\n : newDocumentSet.first();\n newDocumentSet = newDocumentSet.delete(oldDoc!.key);\n newMutatedKeys = newMutatedKeys.delete(oldDoc!.key);\n changeSet.track({ type: ChangeType.Removed, doc: oldDoc! });\n }\n }\n\n debugAssert(\n !needsRefill || !previousChanges,\n 'View was refilled using docs that themselves needed refilling.'\n );\n return {\n documentSet: newDocumentSet,\n changeSet,\n needsRefill,\n mutatedKeys: newMutatedKeys\n };\n }\n\n private shouldWaitForSyncedDocument(\n oldDoc: Document,\n newDoc: Document\n ): boolean {\n // We suppress the initial change event for documents that were modified as\n // part of a write acknowledgment (e.g. when the value of a server transform\n // is applied) as Watch will send us the same document again.\n // By suppressing the event, we only raise two user visible events (one with\n // `hasPendingWrites` and the final state of the document) instead of three\n // (one with `hasPendingWrites`, the modified document with\n // `hasPendingWrites` and the final state of the document).\n return (\n oldDoc.hasLocalMutations &&\n newDoc.hasCommittedMutations &&\n !newDoc.hasLocalMutations\n );\n }\n\n /**\n * Updates the view with the given ViewDocumentChanges and optionally updates\n * limbo docs and sync state from the provided target change.\n * @param docChanges The set of changes to make to the view's docs.\n * @param updateLimboDocuments Whether to update limbo documents based on this\n * change.\n * @param targetChange A target change to apply for computing limbo docs and\n * sync state.\n * @return A new ViewChange with the given docs, changes, and sync state.\n */\n // PORTING NOTE: The iOS/Android clients always compute limbo document changes.\n applyChanges(\n docChanges: ViewDocumentChanges,\n updateLimboDocuments: boolean,\n targetChange?: TargetChange\n ): ViewChange {\n debugAssert(\n !docChanges.needsRefill,\n 'Cannot apply changes that need a refill'\n );\n const oldDocs = this.documentSet;\n this.documentSet = docChanges.documentSet;\n this.mutatedKeys = docChanges.mutatedKeys;\n // Sort changes based on type and query comparator\n const changes = docChanges.changeSet.getChanges();\n changes.sort((c1, c2) => {\n return (\n compareChangeType(c1.type, c2.type) ||\n this.query.docComparator(c1.doc, c2.doc)\n );\n });\n\n this.applyTargetChange(targetChange);\n const limboChanges = updateLimboDocuments\n ? this.updateLimboDocuments()\n : [];\n const synced = this.limboDocuments.size === 0 && this.current;\n const newSyncState = synced ? SyncState.Synced : SyncState.Local;\n const syncStateChanged = newSyncState !== this.syncState;\n this.syncState = newSyncState;\n\n if (changes.length === 0 && !syncStateChanged) {\n // no changes\n return { limboChanges };\n } else {\n const snap: ViewSnapshot = new ViewSnapshot(\n this.query,\n docChanges.documentSet,\n oldDocs,\n changes,\n docChanges.mutatedKeys,\n newSyncState === SyncState.Local,\n syncStateChanged,\n /* excludesMetadataChanges= */ false\n );\n return {\n snapshot: snap,\n limboChanges\n };\n }\n }\n\n /**\n * Applies an OnlineState change to the view, potentially generating a\n * ViewChange if the view's syncState changes as a result.\n */\n applyOnlineStateChange(onlineState: OnlineState): ViewChange {\n if (this.current && onlineState === OnlineState.Offline) {\n // If we're offline, set `current` to false and then call applyChanges()\n // to refresh our syncState and generate a ViewChange as appropriate. We\n // are guaranteed to get a new TargetChange that sets `current` back to\n // true once the client is back online.\n this.current = false;\n return this.applyChanges(\n {\n documentSet: this.documentSet,\n changeSet: new DocumentChangeSet(),\n mutatedKeys: this.mutatedKeys,\n needsRefill: false\n },\n /* updateLimboDocuments= */ false\n );\n } else {\n // No effect, just return a no-op ViewChange.\n return { limboChanges: [] };\n }\n }\n\n /**\n * Returns whether the doc for the given key should be in limbo.\n */\n private shouldBeInLimbo(key: DocumentKey): boolean {\n // If the remote end says it's part of this query, it's not in limbo.\n if (this._syncedDocuments.has(key)) {\n return false;\n }\n // The local store doesn't think it's a result, so it shouldn't be in limbo.\n if (!this.documentSet.has(key)) {\n return false;\n }\n // If there are local changes to the doc, they might explain why the server\n // doesn't know that it's part of the query. So don't put it in limbo.\n // TODO(klimt): Ideally, we would only consider changes that might actually\n // affect this specific query.\n if (this.documentSet.get(key)!.hasLocalMutations) {\n return false;\n }\n // Everything else is in limbo.\n return true;\n }\n\n /**\n * Updates syncedDocuments, current, and limbo docs based on the given change.\n * Returns the list of changes to which docs are in limbo.\n */\n private applyTargetChange(targetChange?: TargetChange): void {\n if (targetChange) {\n targetChange.addedDocuments.forEach(\n key => (this._syncedDocuments = this._syncedDocuments.add(key))\n );\n targetChange.modifiedDocuments.forEach(key => {\n debugAssert(\n this._syncedDocuments.has(key),\n `Modified document ${key} not found in view.`\n );\n });\n targetChange.removedDocuments.forEach(\n key => (this._syncedDocuments = this._syncedDocuments.delete(key))\n );\n this.current = targetChange.current;\n }\n }\n\n private updateLimboDocuments(): LimboDocumentChange[] {\n // We can only determine limbo documents when we're in-sync with the server.\n if (!this.current) {\n return [];\n }\n\n // TODO(klimt): Do this incrementally so that it's not quadratic when\n // updating many documents.\n const oldLimboDocuments = this.limboDocuments;\n this.limboDocuments = documentKeySet();\n this.documentSet.forEach(doc => {\n if (this.shouldBeInLimbo(doc.key)) {\n this.limboDocuments = this.limboDocuments.add(doc.key);\n }\n });\n\n // Diff the new limbo docs with the old limbo docs.\n const changes: LimboDocumentChange[] = [];\n oldLimboDocuments.forEach(key => {\n if (!this.limboDocuments.has(key)) {\n changes.push(new RemovedLimboDocument(key));\n }\n });\n this.limboDocuments.forEach(key => {\n if (!oldLimboDocuments.has(key)) {\n changes.push(new AddedLimboDocument(key));\n }\n });\n return changes;\n }\n\n /**\n * Update the in-memory state of the current view with the state read from\n * persistence.\n *\n * We update the query view whenever a client's primary status changes:\n * - When a client transitions from primary to secondary, it can miss\n * LocalStorage updates and its query views may temporarily not be\n * synchronized with the state on disk.\n * - For secondary to primary transitions, the client needs to update the list\n * of `syncedDocuments` since secondary clients update their query views\n * based purely on synthesized RemoteEvents.\n *\n * @param queryResult.documents - The documents that match the query according\n * to the LocalStore.\n * @param queryResult.remoteKeys - The keys of the documents that match the\n * query according to the backend.\n *\n * @return The ViewChange that resulted from this synchronization.\n */\n // PORTING NOTE: Multi-tab only.\n synchronizeWithPersistedState(queryResult: QueryResult): ViewChange {\n this._syncedDocuments = queryResult.remoteKeys;\n this.limboDocuments = documentKeySet();\n const docChanges = this.computeDocChanges(queryResult.documents);\n return this.applyChanges(docChanges, /*updateLimboDocuments=*/ true);\n }\n\n /**\n * Returns a view snapshot as if this query was just listened to. Contains\n * a document add for every existing document and the `fromCache` and\n * `hasPendingWrites` status of the already established view.\n */\n // PORTING NOTE: Multi-tab only.\n computeInitialSnapshot(): ViewSnapshot {\n return ViewSnapshot.fromInitialDocuments(\n this.query,\n this.documentSet,\n this.mutatedKeys,\n this.syncState === SyncState.Local\n );\n }\n}\n\nfunction compareChangeType(c1: ChangeType, c2: ChangeType): number {\n const order = (change: ChangeType): 0 | 1 | 2 => {\n switch (change) {\n case ChangeType.Added:\n return 1;\n case ChangeType.Modified:\n return 2;\n case ChangeType.Metadata:\n // A metadata change is converted to a modified change at the public\n // api layer. Since we sort by document key and then change type,\n // metadata and modified changes must be sorted equivalently.\n return 2;\n case ChangeType.Removed:\n return 0;\n default:\n return fail('Unknown ChangeType: ' + change);\n }\n };\n\n return order(c1) - order(c2);\n}\n","/**\n * @license\n * Copyright 2019 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Deferred } from '../util/promise';\nimport { TimerId, AsyncQueue } from '../util/async_queue';\nimport { ExponentialBackoff } from '../remote/backoff';\nimport { Transaction } from './transaction';\nimport { RemoteStore } from '../remote/remote_store';\nimport { isNullOrUndefined } from '../util/types';\nimport { isPermanentError } from '../remote/rpc_error';\nimport { FirestoreError } from '../util/error';\n\nconst RETRY_COUNT = 5;\n\n/**\n * TransactionRunner encapsulates the logic needed to run and retry transactions\n * with backoff.\n */\nexport class TransactionRunner<T> {\n private retries = RETRY_COUNT;\n private backoff: ExponentialBackoff;\n\n constructor(\n private readonly asyncQueue: AsyncQueue,\n private readonly remoteStore: RemoteStore,\n private readonly updateFunction: (transaction: Transaction) => Promise<T>,\n private readonly deferred: Deferred<T>\n ) {\n this.backoff = new ExponentialBackoff(\n this.asyncQueue,\n TimerId.TransactionRetry\n );\n }\n\n /** Runs the transaction and sets the result on deferred. */\n run(): void {\n this.runWithBackOff();\n }\n\n private runWithBackOff(): void {\n this.backoff.backoffAndRun(async () => {\n const transaction = this.remoteStore.createTransaction();\n const userPromise = this.tryRunUpdateFunction(transaction);\n if (userPromise) {\n userPromise\n .then(result => {\n this.asyncQueue.enqueueAndForget(() => {\n return transaction\n .commit()\n .then(() => {\n this.deferred.resolve(result);\n })\n .catch(commitError => {\n this.handleTransactionError(commitError);\n });\n });\n })\n .catch(userPromiseError => {\n this.handleTransactionError(userPromiseError);\n });\n }\n });\n }\n\n private tryRunUpdateFunction(transaction: Transaction): Promise<T> | null {\n try {\n const userPromise = this.updateFunction(transaction);\n if (\n isNullOrUndefined(userPromise) ||\n !userPromise.catch ||\n !userPromise.then\n ) {\n this.deferred.reject(\n Error('Transaction callback must return a Promise')\n );\n return null;\n }\n return userPromise;\n } catch (error) {\n // Do not retry errors thrown by user provided updateFunction.\n this.deferred.reject(error);\n return null;\n }\n }\n\n private handleTransactionError(error: Error): void {\n if (this.retries > 0 && this.isRetryableTransactionError(error)) {\n this.retries -= 1;\n this.asyncQueue.enqueueAndForget(() => {\n this.runWithBackOff();\n return Promise.resolve();\n });\n } else {\n this.deferred.reject(error);\n }\n }\n\n private isRetryableTransactionError(error: Error): boolean {\n if (error.name === 'FirebaseError') {\n // In transactions, the backend will fail outdated reads with FAILED_PRECONDITION and\n // non-matching document versions with ABORTED. These errors should be retried.\n const code = (error as FirestoreError).code;\n return (\n code === 'aborted' ||\n code === 'failed-precondition' ||\n !isPermanentError(code)\n );\n }\n return false;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { User } from '../auth/user';\nimport {\n ignoreIfPrimaryLeaseLoss,\n LocalStore,\n MultiTabLocalStore\n} from '../local/local_store';\nimport { LocalViewChanges } from '../local/local_view_changes';\nimport { ReferenceSet } from '../local/reference_set';\nimport { TargetData, TargetPurpose } from '../local/target_data';\nimport {\n documentKeySet,\n DocumentKeySet,\n MaybeDocumentMap\n} from '../model/collections';\nimport { MaybeDocument, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { Mutation } from '../model/mutation';\nimport { BATCHID_UNKNOWN, MutationBatchResult } from '../model/mutation_batch';\nimport { RemoteEvent, TargetChange } from '../remote/remote_event';\nimport { RemoteStore } from '../remote/remote_store';\nimport { RemoteSyncer } from '../remote/remote_syncer';\nimport { debugAssert, fail, hardAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { logDebug } from '../util/log';\nimport { primitiveComparator } from '../util/misc';\nimport { ObjectMap } from '../util/obj_map';\nimport { Deferred } from '../util/promise';\nimport { SortedMap } from '../util/sorted_map';\n\nimport { ClientId, SharedClientState } from '../local/shared_client_state';\nimport {\n QueryTargetState,\n SharedClientStateSyncer\n} from '../local/shared_client_state_syncer';\nimport { SortedSet } from '../util/sorted_set';\nimport { ListenSequence } from './listen_sequence';\nimport { LimitType, Query } from './query';\nimport { SnapshotVersion } from './snapshot_version';\nimport { Target } from './target';\nimport { TargetIdGenerator } from './target_id_generator';\nimport { Transaction } from './transaction';\nimport {\n BatchId,\n MutationBatchState,\n OnlineState,\n OnlineStateSource,\n TargetId\n} from './types';\nimport {\n AddedLimboDocument,\n LimboDocumentChange,\n RemovedLimboDocument,\n View,\n ViewChange,\n ViewDocumentChanges\n} from './view';\nimport { ViewSnapshot } from './view_snapshot';\nimport { AsyncQueue, wrapInUserErrorIfRecoverable } from '../util/async_queue';\nimport { TransactionRunner } from './transaction_runner';\n\nconst LOG_TAG = 'SyncEngine';\n\n/**\n * QueryView contains all of the data that SyncEngine needs to keep track of for\n * a particular query.\n */\nclass QueryView {\n constructor(\n /**\n * The query itself.\n */\n public query: Query,\n /**\n * The target number created by the client that is used in the watch\n * stream to identify this query.\n */\n public targetId: TargetId,\n /**\n * The view is responsible for computing the final merged truth of what\n * docs are in the query. It gets notified of local and remote changes,\n * and applies the query filters and limits to determine the most correct\n * possible results.\n */\n public view: View\n ) {}\n}\n\n/** Tracks a limbo resolution. */\nclass LimboResolution {\n constructor(public key: DocumentKey) {}\n\n /**\n * Set to true once we've received a document. This is used in\n * getRemoteKeysForTarget() and ultimately used by WatchChangeAggregator to\n * decide whether it needs to manufacture a delete event for the target once\n * the target is CURRENT.\n */\n receivedDocument: boolean = false;\n}\n\n/**\n * Interface implemented by EventManager to handle notifications from\n * SyncEngine.\n */\nexport interface SyncEngineListener {\n /** Handles new view snapshots. */\n onWatchChange(snapshots: ViewSnapshot[]): void;\n\n /** Handles the failure of a query. */\n onWatchError(query: Query, error: Error): void;\n\n /** Handles a change in online state. */\n onOnlineStateChange(onlineState: OnlineState): void;\n}\n\n/**\n * SyncEngine is the central controller in the client SDK architecture. It is\n * the glue code between the EventManager, LocalStore, and RemoteStore. Some of\n * SyncEngine's responsibilities include:\n * 1. Coordinating client requests and remote events between the EventManager\n * and the local and remote data stores.\n * 2. Managing a View object for each query, providing the unified view between\n * the local and remote data stores.\n * 3. Notifying the RemoteStore when the LocalStore has new mutations in its\n * queue that need sending to the backend.\n *\n * The SyncEngine’s methods should only ever be called by methods running in the\n * global async queue.\n */\nexport class SyncEngine implements RemoteSyncer {\n protected syncEngineListener: SyncEngineListener | null = null;\n\n protected queryViewsByQuery = new ObjectMap<Query, QueryView>(q =>\n q.canonicalId()\n );\n protected queriesByTarget = new Map<TargetId, Query[]>();\n /**\n * The keys of documents that are in limbo for which we haven't yet started a\n * limbo resolution query.\n */\n private enqueuedLimboResolutions: DocumentKey[] = [];\n /**\n * Keeps track of the target ID for each document that is in limbo with an\n * active target.\n */\n protected activeLimboTargetsByKey = new SortedMap<DocumentKey, TargetId>(\n DocumentKey.comparator\n );\n /**\n * Keeps track of the information about an active limbo resolution for each\n * active target ID that was started for the purpose of limbo resolution.\n */\n protected activeLimboResolutionsByTarget = new Map<\n TargetId,\n LimboResolution\n >();\n protected limboDocumentRefs = new ReferenceSet();\n /** Stores user completion handlers, indexed by User and BatchId. */\n private mutationUserCallbacks = {} as {\n [uidKey: string]: SortedMap<BatchId, Deferred<void>>;\n };\n /** Stores user callbacks waiting for all pending writes to be acknowledged. */\n private pendingWritesCallbacks = new Map<BatchId, Array<Deferred<void>>>();\n private limboTargetIdGenerator = TargetIdGenerator.forSyncEngine();\n\n private onlineState = OnlineState.Unknown;\n\n constructor(\n protected localStore: LocalStore,\n protected remoteStore: RemoteStore,\n // PORTING NOTE: Manages state synchronization in multi-tab environments.\n protected sharedClientState: SharedClientState,\n private currentUser: User,\n private maxConcurrentLimboResolutions: number\n ) {}\n\n get isPrimaryClient(): boolean {\n return true;\n }\n\n /** Subscribes to SyncEngine notifications. Has to be called exactly once. */\n subscribe(syncEngineListener: SyncEngineListener): void {\n debugAssert(\n syncEngineListener !== null,\n 'SyncEngine listener cannot be null'\n );\n debugAssert(\n this.syncEngineListener === null,\n 'SyncEngine already has a subscriber.'\n );\n\n this.syncEngineListener = syncEngineListener;\n }\n\n /**\n * Initiates the new listen, resolves promise when listen enqueued to the\n * server. All the subsequent view snapshots or errors are sent to the\n * subscribed handlers. Returns the initial snapshot.\n */\n async listen(query: Query): Promise<ViewSnapshot> {\n this.assertSubscribed('listen()');\n\n let targetId;\n let viewSnapshot;\n\n const queryView = this.queryViewsByQuery.get(query);\n if (queryView) {\n // PORTING NOTE: With Multi-Tab Web, it is possible that a query view\n // already exists when EventManager calls us for the first time. This\n // happens when the primary tab is already listening to this query on\n // behalf of another tab and the user of the primary also starts listening\n // to the query. EventManager will not have an assigned target ID in this\n // case and calls `listen` to obtain this ID.\n targetId = queryView.targetId;\n this.sharedClientState.addLocalQueryTarget(targetId);\n viewSnapshot = queryView.view.computeInitialSnapshot();\n } else {\n const targetData = await this.localStore.allocateTarget(query.toTarget());\n\n const status = this.sharedClientState.addLocalQueryTarget(\n targetData.targetId\n );\n targetId = targetData.targetId;\n viewSnapshot = await this.initializeViewAndComputeSnapshot(\n query,\n targetId,\n status === 'current'\n );\n if (this.isPrimaryClient) {\n this.remoteStore.listen(targetData);\n }\n }\n\n return viewSnapshot;\n }\n\n /**\n * Registers a view for a previously unknown query and computes its initial\n * snapshot.\n */\n protected async initializeViewAndComputeSnapshot(\n query: Query,\n targetId: TargetId,\n current: boolean\n ): Promise<ViewSnapshot> {\n const queryResult = await this.localStore.executeQuery(\n query,\n /* usePreviousResults= */ true\n );\n const view = new View(query, queryResult.remoteKeys);\n const viewDocChanges = view.computeDocChanges(queryResult.documents);\n const synthesizedTargetChange = TargetChange.createSynthesizedTargetChangeForCurrentChange(\n targetId,\n current && this.onlineState !== OnlineState.Offline\n );\n const viewChange = view.applyChanges(\n viewDocChanges,\n /* updateLimboDocuments= */ this.isPrimaryClient,\n synthesizedTargetChange\n );\n this.updateTrackedLimbos(targetId, viewChange.limboChanges);\n\n debugAssert(\n !!viewChange.snapshot,\n 'applyChanges for new view should always return a snapshot'\n );\n\n const data = new QueryView(query, targetId, view);\n this.queryViewsByQuery.set(query, data);\n if (this.queriesByTarget.has(targetId)) {\n this.queriesByTarget.get(targetId)!.push(query);\n } else {\n this.queriesByTarget.set(targetId, [query]);\n }\n return viewChange.snapshot!;\n }\n\n /** Stops listening to the query. */\n async unlisten(query: Query): Promise<void> {\n this.assertSubscribed('unlisten()');\n\n const queryView = this.queryViewsByQuery.get(query)!;\n debugAssert(!!queryView, 'Trying to unlisten on query not found:' + query);\n\n // Only clean up the query view and target if this is the only query mapped\n // to the target.\n const queries = this.queriesByTarget.get(queryView.targetId)!;\n if (queries.length > 1) {\n this.queriesByTarget.set(\n queryView.targetId,\n queries.filter(q => !q.isEqual(query))\n );\n this.queryViewsByQuery.delete(query);\n return;\n }\n\n // No other queries are mapped to the target, clean up the query and the target.\n if (this.isPrimaryClient) {\n // We need to remove the local query target first to allow us to verify\n // whether any other client is still interested in this target.\n this.sharedClientState.removeLocalQueryTarget(queryView.targetId);\n const targetRemainsActive = this.sharedClientState.isActiveQueryTarget(\n queryView.targetId\n );\n\n if (!targetRemainsActive) {\n await this.localStore\n .releaseTarget(queryView.targetId, /*keepPersistedTargetData=*/ false)\n .then(() => {\n this.sharedClientState.clearQueryState(queryView.targetId);\n this.remoteStore.unlisten(queryView.targetId);\n this.removeAndCleanupTarget(queryView.targetId);\n })\n .catch(ignoreIfPrimaryLeaseLoss);\n }\n } else {\n this.removeAndCleanupTarget(queryView.targetId);\n await this.localStore.releaseTarget(\n queryView.targetId,\n /*keepPersistedTargetData=*/ true\n );\n }\n }\n\n /**\n * Initiates the write of local mutation batch which involves adding the\n * writes to the mutation queue, notifying the remote store about new\n * mutations and raising events for any changes this write caused.\n *\n * The promise returned by this call is resolved when the above steps\n * have completed, *not* when the write was acked by the backend. The\n * userCallback is resolved once the write was acked/rejected by the\n * backend (or failed locally for any other reason).\n */\n async write(batch: Mutation[], userCallback: Deferred<void>): Promise<void> {\n this.assertSubscribed('write()');\n\n try {\n const result = await this.localStore.localWrite(batch);\n this.sharedClientState.addPendingMutation(result.batchId);\n this.addMutationCallback(result.batchId, userCallback);\n await this.emitNewSnapsAndNotifyLocalStore(result.changes);\n await this.remoteStore.fillWritePipeline();\n } catch (e) {\n // If we can't persist the mutation, we reject the user callback and\n // don't send the mutation. The user can then retry the write.\n const error = wrapInUserErrorIfRecoverable(e, `Failed to persist write`);\n userCallback.reject(error);\n }\n }\n\n /**\n * Takes an updateFunction in which a set of reads and writes can be performed\n * atomically. In the updateFunction, the client can read and write values\n * using the supplied transaction object. After the updateFunction, all\n * changes will be committed. If a retryable error occurs (ex: some other\n * client has changed any of the data referenced), then the updateFunction\n * will be called again after a backoff. If the updateFunction still fails\n * after all retries, then the transaction will be rejected.\n *\n * The transaction object passed to the updateFunction contains methods for\n * accessing documents and collections. Unlike other datastore access, data\n * accessed with the transaction will not reflect local changes that have not\n * been committed. For this reason, it is required that all reads are\n * performed before any writes. Transactions must be performed while online.\n *\n * The Deferred input is resolved when the transaction is fully committed.\n */\n runTransaction<T>(\n asyncQueue: AsyncQueue,\n updateFunction: (transaction: Transaction) => Promise<T>,\n deferred: Deferred<T>\n ): void {\n new TransactionRunner<T>(\n asyncQueue,\n this.remoteStore,\n updateFunction,\n deferred\n ).run();\n }\n\n async applyRemoteEvent(remoteEvent: RemoteEvent): Promise<void> {\n this.assertSubscribed('applyRemoteEvent()');\n try {\n const changes = await this.localStore.applyRemoteEvent(remoteEvent);\n // Update `receivedDocument` as appropriate for any limbo targets.\n remoteEvent.targetChanges.forEach((targetChange, targetId) => {\n const limboResolution = this.activeLimboResolutionsByTarget.get(\n targetId\n );\n if (limboResolution) {\n // Since this is a limbo resolution lookup, it's for a single document\n // and it could be added, modified, or removed, but not a combination.\n hardAssert(\n targetChange.addedDocuments.size +\n targetChange.modifiedDocuments.size +\n targetChange.removedDocuments.size <=\n 1,\n 'Limbo resolution for single document contains multiple changes.'\n );\n if (targetChange.addedDocuments.size > 0) {\n limboResolution.receivedDocument = true;\n } else if (targetChange.modifiedDocuments.size > 0) {\n hardAssert(\n limboResolution.receivedDocument,\n 'Received change for limbo target document without add.'\n );\n } else if (targetChange.removedDocuments.size > 0) {\n hardAssert(\n limboResolution.receivedDocument,\n 'Received remove for limbo target document without add.'\n );\n limboResolution.receivedDocument = false;\n } else {\n // This was probably just a CURRENT targetChange or similar.\n }\n }\n });\n await this.emitNewSnapsAndNotifyLocalStore(changes, remoteEvent);\n } catch (error) {\n await ignoreIfPrimaryLeaseLoss(error);\n }\n }\n\n /**\n * Applies an OnlineState change to the sync engine and notifies any views of\n * the change.\n */\n applyOnlineStateChange(\n onlineState: OnlineState,\n source: OnlineStateSource\n ): void {\n this.assertSubscribed('applyOnlineStateChange()');\n const newViewSnapshots = [] as ViewSnapshot[];\n this.queryViewsByQuery.forEach((query, queryView) => {\n const viewChange = queryView.view.applyOnlineStateChange(onlineState);\n debugAssert(\n viewChange.limboChanges.length === 0,\n 'OnlineState should not affect limbo documents.'\n );\n if (viewChange.snapshot) {\n newViewSnapshots.push(viewChange.snapshot);\n }\n });\n this.syncEngineListener!.onOnlineStateChange(onlineState);\n this.syncEngineListener!.onWatchChange(newViewSnapshots);\n this.onlineState = onlineState;\n }\n\n async rejectListen(targetId: TargetId, err: FirestoreError): Promise<void> {\n this.assertSubscribed('rejectListens()');\n\n // PORTING NOTE: Multi-tab only.\n this.sharedClientState.updateQueryState(targetId, 'rejected', err);\n\n const limboResolution = this.activeLimboResolutionsByTarget.get(targetId);\n const limboKey = limboResolution && limboResolution.key;\n if (limboKey) {\n // TODO(klimt): We really only should do the following on permission\n // denied errors, but we don't have the cause code here.\n\n // It's a limbo doc. Create a synthetic event saying it was deleted.\n // This is kind of a hack. Ideally, we would have a method in the local\n // store to purge a document. However, it would be tricky to keep all of\n // the local store's invariants with another method.\n let documentUpdates = new SortedMap<DocumentKey, MaybeDocument>(\n DocumentKey.comparator\n );\n documentUpdates = documentUpdates.insert(\n limboKey,\n new NoDocument(limboKey, SnapshotVersion.min())\n );\n const resolvedLimboDocuments = documentKeySet().add(limboKey);\n const event = new RemoteEvent(\n SnapshotVersion.min(),\n /* targetChanges= */ new Map<TargetId, TargetChange>(),\n /* targetMismatches= */ new SortedSet<TargetId>(primitiveComparator),\n documentUpdates,\n resolvedLimboDocuments\n );\n\n await this.applyRemoteEvent(event);\n\n // Since this query failed, we won't want to manually unlisten to it.\n // We only remove it from bookkeeping after we successfully applied the\n // RemoteEvent. If `applyRemoteEvent()` throws, we want to re-listen to\n // this query when the RemoteStore restarts the Watch stream, which should\n // re-trigger the target failure.\n this.activeLimboTargetsByKey = this.activeLimboTargetsByKey.remove(\n limboKey\n );\n this.activeLimboResolutionsByTarget.delete(targetId);\n this.pumpEnqueuedLimboResolutions();\n } else {\n await this.localStore\n .releaseTarget(targetId, /* keepPersistedTargetData */ false)\n .then(() => this.removeAndCleanupTarget(targetId, err))\n .catch(ignoreIfPrimaryLeaseLoss);\n }\n }\n\n async applySuccessfulWrite(\n mutationBatchResult: MutationBatchResult\n ): Promise<void> {\n this.assertSubscribed('applySuccessfulWrite()');\n\n const batchId = mutationBatchResult.batch.batchId;\n\n // The local store may or may not be able to apply the write result and\n // raise events immediately (depending on whether the watcher is caught\n // up), so we raise user callbacks first so that they consistently happen\n // before listen events.\n this.processUserCallback(batchId, /*error=*/ null);\n\n this.triggerPendingWritesCallbacks(batchId);\n\n try {\n const changes = await this.localStore.acknowledgeBatch(\n mutationBatchResult\n );\n this.sharedClientState.updateMutationState(batchId, 'acknowledged');\n await this.emitNewSnapsAndNotifyLocalStore(changes);\n } catch (error) {\n await ignoreIfPrimaryLeaseLoss(error);\n }\n }\n\n async rejectFailedWrite(\n batchId: BatchId,\n error: FirestoreError\n ): Promise<void> {\n this.assertSubscribed('rejectFailedWrite()');\n\n // The local store may or may not be able to apply the write result and\n // raise events immediately (depending on whether the watcher is caught up),\n // so we raise user callbacks first so that they consistently happen before\n // listen events.\n this.processUserCallback(batchId, error);\n\n this.triggerPendingWritesCallbacks(batchId);\n\n try {\n const changes = await this.localStore.rejectBatch(batchId);\n this.sharedClientState.updateMutationState(batchId, 'rejected', error);\n await this.emitNewSnapsAndNotifyLocalStore(changes);\n } catch (error) {\n await ignoreIfPrimaryLeaseLoss(error);\n }\n }\n\n /**\n * Registers a user callback that resolves when all pending mutations at the moment of calling\n * are acknowledged .\n */\n async registerPendingWritesCallback(callback: Deferred<void>): Promise<void> {\n if (!this.remoteStore.canUseNetwork()) {\n logDebug(\n LOG_TAG,\n 'The network is disabled. The task returned by ' +\n \"'awaitPendingWrites()' will not complete until the network is enabled.\"\n );\n }\n\n try {\n const highestBatchId = await this.localStore.getHighestUnacknowledgedBatchId();\n if (highestBatchId === BATCHID_UNKNOWN) {\n // Trigger the callback right away if there is no pending writes at the moment.\n callback.resolve();\n return;\n }\n\n const callbacks = this.pendingWritesCallbacks.get(highestBatchId) || [];\n callbacks.push(callback);\n this.pendingWritesCallbacks.set(highestBatchId, callbacks);\n } catch (e) {\n const firestoreError = wrapInUserErrorIfRecoverable(\n e,\n 'Initialization of waitForPendingWrites() operation failed'\n );\n callback.reject(firestoreError);\n }\n }\n\n /**\n * Triggers the callbacks that are waiting for this batch id to get acknowledged by server,\n * if there are any.\n */\n private triggerPendingWritesCallbacks(batchId: BatchId): void {\n (this.pendingWritesCallbacks.get(batchId) || []).forEach(callback => {\n callback.resolve();\n });\n\n this.pendingWritesCallbacks.delete(batchId);\n }\n\n /** Reject all outstanding callbacks waiting for pending writes to complete. */\n private rejectOutstandingPendingWritesCallbacks(errorMessage: string): void {\n this.pendingWritesCallbacks.forEach(callbacks => {\n callbacks.forEach(callback => {\n callback.reject(new FirestoreError(Code.CANCELLED, errorMessage));\n });\n });\n\n this.pendingWritesCallbacks.clear();\n }\n\n private addMutationCallback(\n batchId: BatchId,\n callback: Deferred<void>\n ): void {\n let newCallbacks = this.mutationUserCallbacks[this.currentUser.toKey()];\n if (!newCallbacks) {\n newCallbacks = new SortedMap<BatchId, Deferred<void>>(\n primitiveComparator\n );\n }\n newCallbacks = newCallbacks.insert(batchId, callback);\n this.mutationUserCallbacks[this.currentUser.toKey()] = newCallbacks;\n }\n\n /**\n * Resolves or rejects the user callback for the given batch and then discards\n * it.\n */\n protected processUserCallback(batchId: BatchId, error: Error | null): void {\n let newCallbacks = this.mutationUserCallbacks[this.currentUser.toKey()];\n\n // NOTE: Mutations restored from persistence won't have callbacks, so it's\n // okay for there to be no callback for this ID.\n if (newCallbacks) {\n const callback = newCallbacks.get(batchId);\n if (callback) {\n debugAssert(\n batchId === newCallbacks.minKey(),\n 'Mutation callbacks processed out-of-order?'\n );\n if (error) {\n callback.reject(error);\n } else {\n callback.resolve();\n }\n newCallbacks = newCallbacks.remove(batchId);\n }\n this.mutationUserCallbacks[this.currentUser.toKey()] = newCallbacks;\n }\n }\n\n protected removeAndCleanupTarget(\n targetId: number,\n error: Error | null = null\n ): void {\n this.sharedClientState.removeLocalQueryTarget(targetId);\n\n debugAssert(\n this.queriesByTarget.has(targetId) &&\n this.queriesByTarget.get(targetId)!.length !== 0,\n `There are no queries mapped to target id ${targetId}`\n );\n\n for (const query of this.queriesByTarget.get(targetId)!) {\n this.queryViewsByQuery.delete(query);\n if (error) {\n this.syncEngineListener!.onWatchError(query, error);\n }\n }\n\n this.queriesByTarget.delete(targetId);\n\n if (this.isPrimaryClient) {\n const limboKeys = this.limboDocumentRefs.removeReferencesForId(targetId);\n limboKeys.forEach(limboKey => {\n const isReferenced = this.limboDocumentRefs.containsKey(limboKey);\n if (!isReferenced) {\n // We removed the last reference for this key\n this.removeLimboTarget(limboKey);\n }\n });\n }\n }\n\n private removeLimboTarget(key: DocumentKey): void {\n // It's possible that the target already got removed because the query failed. In that case,\n // the key won't exist in `limboTargetsByKey`. Only do the cleanup if we still have the target.\n const limboTargetId = this.activeLimboTargetsByKey.get(key);\n if (limboTargetId === null) {\n // This target already got removed, because the query failed.\n return;\n }\n\n this.remoteStore.unlisten(limboTargetId);\n this.activeLimboTargetsByKey = this.activeLimboTargetsByKey.remove(key);\n this.activeLimboResolutionsByTarget.delete(limboTargetId);\n this.pumpEnqueuedLimboResolutions();\n }\n\n protected updateTrackedLimbos(\n targetId: TargetId,\n limboChanges: LimboDocumentChange[]\n ): void {\n for (const limboChange of limboChanges) {\n if (limboChange instanceof AddedLimboDocument) {\n this.limboDocumentRefs.addReference(limboChange.key, targetId);\n this.trackLimboChange(limboChange);\n } else if (limboChange instanceof RemovedLimboDocument) {\n logDebug(LOG_TAG, 'Document no longer in limbo: ' + limboChange.key);\n this.limboDocumentRefs.removeReference(limboChange.key, targetId);\n const isReferenced = this.limboDocumentRefs.containsKey(\n limboChange.key\n );\n if (!isReferenced) {\n // We removed the last reference for this key\n this.removeLimboTarget(limboChange.key);\n }\n } else {\n fail('Unknown limbo change: ' + JSON.stringify(limboChange));\n }\n }\n }\n\n private trackLimboChange(limboChange: AddedLimboDocument): void {\n const key = limboChange.key;\n if (!this.activeLimboTargetsByKey.get(key)) {\n logDebug(LOG_TAG, 'New document in limbo: ' + key);\n this.enqueuedLimboResolutions.push(key);\n this.pumpEnqueuedLimboResolutions();\n }\n }\n\n /**\n * Starts listens for documents in limbo that are enqueued for resolution,\n * subject to a maximum number of concurrent resolutions.\n *\n * Without bounding the number of concurrent resolutions, the server can fail\n * with \"resource exhausted\" errors which can lead to pathological client\n * behavior as seen in https://github.com/firebase/firebase-js-sdk/issues/2683.\n */\n private pumpEnqueuedLimboResolutions(): void {\n while (\n this.enqueuedLimboResolutions.length > 0 &&\n this.activeLimboTargetsByKey.size < this.maxConcurrentLimboResolutions\n ) {\n const key = this.enqueuedLimboResolutions.shift()!;\n const limboTargetId = this.limboTargetIdGenerator.next();\n this.activeLimboResolutionsByTarget.set(\n limboTargetId,\n new LimboResolution(key)\n );\n this.activeLimboTargetsByKey = this.activeLimboTargetsByKey.insert(\n key,\n limboTargetId\n );\n this.remoteStore.listen(\n new TargetData(\n Query.atPath(key.path).toTarget(),\n limboTargetId,\n TargetPurpose.LimboResolution,\n ListenSequence.INVALID\n )\n );\n }\n }\n\n // Visible for testing\n activeLimboDocumentResolutions(): SortedMap<DocumentKey, TargetId> {\n return this.activeLimboTargetsByKey;\n }\n\n // Visible for testing\n enqueuedLimboDocumentResolutions(): DocumentKey[] {\n return this.enqueuedLimboResolutions;\n }\n\n protected async emitNewSnapsAndNotifyLocalStore(\n changes: MaybeDocumentMap,\n remoteEvent?: RemoteEvent\n ): Promise<void> {\n const newSnaps: ViewSnapshot[] = [];\n const docChangesInAllViews: LocalViewChanges[] = [];\n const queriesProcessed: Array<Promise<void>> = [];\n\n this.queryViewsByQuery.forEach((_, queryView) => {\n queriesProcessed.push(\n Promise.resolve()\n .then(() => {\n const viewDocChanges = queryView.view.computeDocChanges(changes);\n if (!viewDocChanges.needsRefill) {\n return viewDocChanges;\n }\n // The query has a limit and some docs were removed, so we need\n // to re-run the query against the local store to make sure we\n // didn't lose any good docs that had been past the limit.\n return this.localStore\n .executeQuery(queryView.query, /* usePreviousResults= */ false)\n .then(({ documents }) => {\n return queryView.view.computeDocChanges(\n documents,\n viewDocChanges\n );\n });\n })\n .then((viewDocChanges: ViewDocumentChanges) => {\n const targetChange =\n remoteEvent && remoteEvent.targetChanges.get(queryView.targetId);\n const viewChange = queryView.view.applyChanges(\n viewDocChanges,\n /* updateLimboDocuments= */ this.isPrimaryClient,\n targetChange\n );\n this.updateTrackedLimbos(\n queryView.targetId,\n viewChange.limboChanges\n );\n if (viewChange.snapshot) {\n if (this.isPrimaryClient) {\n this.sharedClientState.updateQueryState(\n queryView.targetId,\n viewChange.snapshot.fromCache ? 'not-current' : 'current'\n );\n }\n\n newSnaps.push(viewChange.snapshot);\n const docChanges = LocalViewChanges.fromSnapshot(\n queryView.targetId,\n viewChange.snapshot\n );\n docChangesInAllViews.push(docChanges);\n }\n })\n );\n });\n\n await Promise.all(queriesProcessed);\n this.syncEngineListener!.onWatchChange(newSnaps);\n await this.localStore.notifyLocalViewChanges(docChangesInAllViews);\n }\n\n protected assertSubscribed(fnName: string): void {\n debugAssert(\n this.syncEngineListener !== null,\n 'Trying to call ' + fnName + ' before calling subscribe().'\n );\n }\n\n async handleCredentialChange(user: User): Promise<void> {\n const userChanged = !this.currentUser.isEqual(user);\n\n if (userChanged) {\n const result = await this.localStore.handleUserChange(user);\n this.currentUser = user;\n\n // Fails tasks waiting for pending writes requested by previous user.\n this.rejectOutstandingPendingWritesCallbacks(\n \"'waitForPendingWrites' promise is rejected due to a user change.\"\n );\n // TODO(b/114226417): Consider calling this only in the primary tab.\n this.sharedClientState.handleUserChange(\n user,\n result.removedBatchIds,\n result.addedBatchIds\n );\n await this.emitNewSnapsAndNotifyLocalStore(result.affectedDocuments);\n }\n\n await this.remoteStore.handleCredentialChange();\n }\n\n enableNetwork(): Promise<void> {\n return this.remoteStore.enableNetwork();\n }\n\n disableNetwork(): Promise<void> {\n return this.remoteStore.disableNetwork();\n }\n\n getRemoteKeysForTarget(targetId: TargetId): DocumentKeySet {\n const limboResolution = this.activeLimboResolutionsByTarget.get(targetId);\n if (limboResolution && limboResolution.receivedDocument) {\n return documentKeySet().add(limboResolution.key);\n } else {\n let keySet = documentKeySet();\n const queries = this.queriesByTarget.get(targetId);\n if (!queries) {\n return keySet;\n }\n for (const query of queries) {\n const queryView = this.queryViewsByQuery.get(query);\n debugAssert(!!queryView, `No query view found for ${query}`);\n keySet = keySet.unionWith(queryView.view.syncedDocuments);\n }\n return keySet;\n }\n }\n}\n\n/**\n * An impplementation of SyncEngine that implement SharedClientStateSyncer for\n * Multi-Tab synchronization.\n */\n// PORTING NOTE: Web only\nexport class MultiTabSyncEngine extends SyncEngine\n implements SharedClientStateSyncer {\n // The primary state is set to `true` or `false` immediately after Firestore\n // startup. In the interim, a client should only be considered primary if\n // `isPrimary` is true.\n private _isPrimaryClient: undefined | boolean = undefined;\n\n constructor(\n protected localStore: MultiTabLocalStore,\n remoteStore: RemoteStore,\n sharedClientState: SharedClientState,\n currentUser: User,\n maxConcurrentLimboResolutions: number\n ) {\n super(\n localStore,\n remoteStore,\n sharedClientState,\n currentUser,\n maxConcurrentLimboResolutions\n );\n }\n\n get isPrimaryClient(): boolean {\n return this._isPrimaryClient === true;\n }\n\n enableNetwork(): Promise<void> {\n this.localStore.setNetworkEnabled(true);\n return super.enableNetwork();\n }\n\n disableNetwork(): Promise<void> {\n this.localStore.setNetworkEnabled(false);\n return super.disableNetwork();\n }\n\n /**\n * Reconcile the list of synced documents in an existing view with those\n * from persistence.\n */\n private async synchronizeViewAndComputeSnapshot(\n queryView: QueryView\n ): Promise<ViewChange> {\n const queryResult = await this.localStore.executeQuery(\n queryView.query,\n /* usePreviousResults= */ true\n );\n const viewSnapshot = queryView.view.synchronizeWithPersistedState(\n queryResult\n );\n if (this._isPrimaryClient) {\n this.updateTrackedLimbos(queryView.targetId, viewSnapshot.limboChanges);\n }\n return viewSnapshot;\n }\n\n applyOnlineStateChange(\n onlineState: OnlineState,\n source: OnlineStateSource\n ): void {\n // If we are the primary client, the online state of all clients only\n // depends on the online state of the local RemoteStore.\n if (this.isPrimaryClient && source === OnlineStateSource.RemoteStore) {\n super.applyOnlineStateChange(onlineState, source);\n this.sharedClientState.setOnlineState(onlineState);\n }\n\n // If we are the secondary client, we explicitly ignore the remote store's\n // online state (the local client may go offline, even though the primary\n // tab remains online) and only apply the primary tab's online state from\n // SharedClientState.\n if (\n !this.isPrimaryClient &&\n source === OnlineStateSource.SharedClientState\n ) {\n super.applyOnlineStateChange(onlineState, source);\n }\n }\n\n async applyBatchState(\n batchId: BatchId,\n batchState: MutationBatchState,\n error?: FirestoreError\n ): Promise<void> {\n this.assertSubscribed('applyBatchState()');\n const documents = await this.localStore.lookupMutationDocuments(batchId);\n\n if (documents === null) {\n // A throttled tab may not have seen the mutation before it was completed\n // and removed from the mutation queue, in which case we won't have cached\n // the affected documents. In this case we can safely ignore the update\n // since that means we didn't apply the mutation locally at all (if we\n // had, we would have cached the affected documents), and so we will just\n // see any resulting document changes via normal remote document updates\n // as applicable.\n logDebug(LOG_TAG, 'Cannot apply mutation batch with id: ' + batchId);\n return;\n }\n\n if (batchState === 'pending') {\n // If we are the primary client, we need to send this write to the\n // backend. Secondary clients will ignore these writes since their remote\n // connection is disabled.\n await this.remoteStore.fillWritePipeline();\n } else if (batchState === 'acknowledged' || batchState === 'rejected') {\n // NOTE: Both these methods are no-ops for batches that originated from\n // other clients.\n this.processUserCallback(batchId, error ? error : null);\n this.localStore.removeCachedMutationBatchMetadata(batchId);\n } else {\n fail(`Unknown batchState: ${batchState}`);\n }\n\n await this.emitNewSnapsAndNotifyLocalStore(documents);\n }\n\n async applyPrimaryState(isPrimary: boolean): Promise<void> {\n if (isPrimary === true && this._isPrimaryClient !== true) {\n // Secondary tabs only maintain Views for their local listeners and the\n // Views internal state may not be 100% populated (in particular\n // secondary tabs don't track syncedDocuments, the set of documents the\n // server considers to be in the target). So when a secondary becomes\n // primary, we need to need to make sure that all views for all targets\n // match the state on disk.\n const activeTargets = this.sharedClientState.getAllActiveQueryTargets();\n const activeQueries = await this.synchronizeQueryViewsAndRaiseSnapshots(\n activeTargets.toArray(),\n /*transitionToPrimary=*/ true\n );\n this._isPrimaryClient = true;\n await this.remoteStore.applyPrimaryState(true);\n for (const targetData of activeQueries) {\n this.remoteStore.listen(targetData);\n }\n } else if (isPrimary === false && this._isPrimaryClient !== false) {\n const activeTargets: TargetId[] = [];\n\n let p = Promise.resolve();\n this.queriesByTarget.forEach((_, targetId) => {\n if (this.sharedClientState.isLocalQueryTarget(targetId)) {\n activeTargets.push(targetId);\n } else {\n p = p.then(() => {\n this.removeAndCleanupTarget(targetId);\n return this.localStore.releaseTarget(\n targetId,\n /*keepPersistedTargetData=*/ true\n );\n });\n }\n this.remoteStore.unlisten(targetId);\n });\n await p;\n\n await this.synchronizeQueryViewsAndRaiseSnapshots(\n activeTargets,\n /*transitionToPrimary=*/ false\n );\n this.resetLimboDocuments();\n this._isPrimaryClient = false;\n await this.remoteStore.applyPrimaryState(false);\n }\n }\n\n private resetLimboDocuments(): void {\n this.activeLimboResolutionsByTarget.forEach((_, targetId) => {\n this.remoteStore.unlisten(targetId);\n });\n this.limboDocumentRefs.removeAllReferences();\n this.activeLimboResolutionsByTarget = new Map<TargetId, LimboResolution>();\n this.activeLimboTargetsByKey = new SortedMap<DocumentKey, TargetId>(\n DocumentKey.comparator\n );\n }\n\n /**\n * Reconcile the query views of the provided query targets with the state from\n * persistence. Raises snapshots for any changes that affect the local\n * client and returns the updated state of all target's query data.\n *\n * @param targets the list of targets with views that need to be recomputed\n * @param transitionToPrimary `true` iff the tab transitions from a secondary\n * tab to a primary tab\n */\n private async synchronizeQueryViewsAndRaiseSnapshots(\n targets: TargetId[],\n transitionToPrimary: boolean\n ): Promise<TargetData[]> {\n const activeQueries: TargetData[] = [];\n const newViewSnapshots: ViewSnapshot[] = [];\n for (const targetId of targets) {\n let targetData: TargetData;\n const queries = this.queriesByTarget.get(targetId);\n\n if (queries && queries.length !== 0) {\n // For queries that have a local View, we need to update their state\n // in LocalStore (as the resume token and the snapshot version\n // might have changed) and reconcile their views with the persisted\n // state (the list of syncedDocuments may have gotten out of sync).\n await this.localStore.releaseTarget(\n targetId,\n /*keepPersistedTargetData=*/ true\n );\n targetData = await this.localStore.allocateTarget(\n queries[0].toTarget()\n );\n\n for (const query of queries) {\n const queryView = this.queryViewsByQuery.get(query);\n debugAssert(!!queryView, `No query view found for ${query}`);\n\n const viewChange = await this.synchronizeViewAndComputeSnapshot(\n queryView\n );\n if (viewChange.snapshot) {\n newViewSnapshots.push(viewChange.snapshot);\n }\n }\n } else {\n debugAssert(\n transitionToPrimary,\n 'A secondary tab should never have an active target without an active query.'\n );\n // For queries that never executed on this client, we need to\n // allocate the target in LocalStore and initialize a new View.\n const target = await this.localStore.getTarget(targetId);\n debugAssert(!!target, `Target for id ${targetId} not found`);\n targetData = await this.localStore.allocateTarget(target);\n await this.initializeViewAndComputeSnapshot(\n this.synthesizeTargetToQuery(target!),\n targetId,\n /*current=*/ false\n );\n }\n\n activeQueries.push(targetData!);\n }\n\n this.syncEngineListener!.onWatchChange(newViewSnapshots);\n return activeQueries;\n }\n\n /**\n * Creates a `Query` object from the specified `Target`. There is no way to\n * obtain the original `Query`, so we synthesize a `Query` from the `Target`\n * object.\n *\n * The synthesized result might be different from the original `Query`, but\n * since the synthesized `Query` should return the same results as the\n * original one (only the presentation of results might differ), the potential\n * difference will not cause issues.\n */\n private synthesizeTargetToQuery(target: Target): Query {\n return new Query(\n target.path,\n target.collectionGroup,\n target.orderBy,\n target.filters,\n target.limit,\n LimitType.First,\n target.startAt,\n target.endAt\n );\n }\n\n getActiveClients(): Promise<ClientId[]> {\n return this.localStore.getActiveClients();\n }\n\n async applyTargetState(\n targetId: TargetId,\n state: QueryTargetState,\n error?: FirestoreError\n ): Promise<void> {\n if (this._isPrimaryClient) {\n // If we receive a target state notification via WebStorage, we are\n // either already secondary or another tab has taken the primary lease.\n logDebug(LOG_TAG, 'Ignoring unexpected query state notification.');\n return;\n }\n\n if (this.queriesByTarget.has(targetId)) {\n switch (state) {\n case 'current':\n case 'not-current': {\n const changes = await this.localStore.getNewDocumentChanges();\n const synthesizedRemoteEvent = RemoteEvent.createSynthesizedRemoteEventForCurrentChange(\n targetId,\n state === 'current'\n );\n await this.emitNewSnapsAndNotifyLocalStore(\n changes,\n synthesizedRemoteEvent\n );\n break;\n }\n case 'rejected': {\n await this.localStore.releaseTarget(\n targetId,\n /* keepPersistedTargetData */ true\n );\n this.removeAndCleanupTarget(targetId, error);\n break;\n }\n default:\n fail('Unexpected target state: ' + state);\n }\n }\n }\n\n async applyActiveTargetsChange(\n added: TargetId[],\n removed: TargetId[]\n ): Promise<void> {\n if (!this._isPrimaryClient) {\n return;\n }\n\n for (const targetId of added) {\n if (this.queriesByTarget.has(targetId)) {\n // A target might have been added in a previous attempt\n logDebug(LOG_TAG, 'Adding an already active target ' + targetId);\n continue;\n }\n\n const target = await this.localStore.getTarget(targetId);\n debugAssert(\n !!target,\n `Query data for active target ${targetId} not found`\n );\n const targetData = await this.localStore.allocateTarget(target);\n await this.initializeViewAndComputeSnapshot(\n this.synthesizeTargetToQuery(target),\n targetData.targetId,\n /*current=*/ false\n );\n this.remoteStore.listen(targetData);\n }\n\n for (const targetId of removed) {\n // Check that the target is still active since the target might have been\n // removed if it has been rejected by the backend.\n if (!this.queriesByTarget.has(targetId)) {\n continue;\n }\n\n // Release queries that are still active.\n await this.localStore\n .releaseTarget(targetId, /* keepPersistedTargetData */ false)\n .then(() => {\n this.remoteStore.unlisten(targetId);\n this.removeAndCleanupTarget(targetId);\n })\n .catch(ignoreIfPrimaryLeaseLoss);\n }\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert } from '../util/assert';\nimport { EventHandler } from '../util/misc';\nimport { ObjectMap } from '../util/obj_map';\nimport { Query } from './query';\nimport { SyncEngine, SyncEngineListener } from './sync_engine';\nimport { OnlineState } from './types';\nimport { ChangeType, DocumentViewChange, ViewSnapshot } from './view_snapshot';\nimport { wrapInUserErrorIfRecoverable } from '../util/async_queue';\n\n/**\n * Holds the listeners and the last received ViewSnapshot for a query being\n * tracked by EventManager.\n */\nclass QueryListenersInfo {\n viewSnap: ViewSnapshot | undefined = undefined;\n listeners: QueryListener[] = [];\n}\n\n/**\n * Interface for handling events from the EventManager.\n */\nexport interface Observer<T> {\n next: EventHandler<T>;\n error: EventHandler<Error>;\n}\n\n/**\n * EventManager is responsible for mapping queries to query event emitters.\n * It handles \"fan-out\". -- Identical queries will re-use the same watch on the\n * backend.\n */\nexport class EventManager implements SyncEngineListener {\n private queries = new ObjectMap<Query, QueryListenersInfo>(q =>\n q.canonicalId()\n );\n\n private onlineState = OnlineState.Unknown;\n\n private snapshotsInSyncListeners: Set<Observer<void>> = new Set();\n\n constructor(private syncEngine: SyncEngine) {\n this.syncEngine.subscribe(this);\n }\n\n async listen(listener: QueryListener): Promise<void> {\n const query = listener.query;\n let firstListen = false;\n\n let queryInfo = this.queries.get(query);\n if (!queryInfo) {\n firstListen = true;\n queryInfo = new QueryListenersInfo();\n }\n\n if (firstListen) {\n try {\n queryInfo.viewSnap = await this.syncEngine.listen(query);\n } catch (e) {\n const firestoreError = wrapInUserErrorIfRecoverable(\n e,\n `Initialization of query '${listener.query}' failed`\n );\n listener.onError(firestoreError);\n return;\n }\n }\n\n this.queries.set(query, queryInfo);\n queryInfo.listeners.push(listener);\n\n // Run global snapshot listeners if a consistent snapshot has been emitted.\n const raisedEvent = listener.applyOnlineStateChange(this.onlineState);\n debugAssert(\n !raisedEvent,\n \"applyOnlineStateChange() shouldn't raise an event for brand-new listeners.\"\n );\n\n if (queryInfo.viewSnap) {\n const raisedEvent = listener.onViewSnapshot(queryInfo.viewSnap);\n if (raisedEvent) {\n this.raiseSnapshotsInSyncEvent();\n }\n }\n }\n\n async unlisten(listener: QueryListener): Promise<void> {\n const query = listener.query;\n let lastListen = false;\n\n const queryInfo = this.queries.get(query);\n if (queryInfo) {\n const i = queryInfo.listeners.indexOf(listener);\n if (i >= 0) {\n queryInfo.listeners.splice(i, 1);\n lastListen = queryInfo.listeners.length === 0;\n }\n }\n\n if (lastListen) {\n this.queries.delete(query);\n return this.syncEngine.unlisten(query);\n }\n }\n\n onWatchChange(viewSnaps: ViewSnapshot[]): void {\n let raisedEvent = false;\n for (const viewSnap of viewSnaps) {\n const query = viewSnap.query;\n const queryInfo = this.queries.get(query);\n if (queryInfo) {\n for (const listener of queryInfo.listeners) {\n if (listener.onViewSnapshot(viewSnap)) {\n raisedEvent = true;\n }\n }\n queryInfo.viewSnap = viewSnap;\n }\n }\n if (raisedEvent) {\n this.raiseSnapshotsInSyncEvent();\n }\n }\n\n onWatchError(query: Query, error: Error): void {\n const queryInfo = this.queries.get(query);\n if (queryInfo) {\n for (const listener of queryInfo.listeners) {\n listener.onError(error);\n }\n }\n\n // Remove all listeners. NOTE: We don't need to call syncEngine.unlisten()\n // after an error.\n this.queries.delete(query);\n }\n\n onOnlineStateChange(onlineState: OnlineState): void {\n this.onlineState = onlineState;\n let raisedEvent = false;\n this.queries.forEach((_, queryInfo) => {\n for (const listener of queryInfo.listeners) {\n // Run global snapshot listeners if a consistent snapshot has been emitted.\n if (listener.applyOnlineStateChange(onlineState)) {\n raisedEvent = true;\n }\n }\n });\n if (raisedEvent) {\n this.raiseSnapshotsInSyncEvent();\n }\n }\n\n addSnapshotsInSyncListener(observer: Observer<void>): void {\n this.snapshotsInSyncListeners.add(observer);\n // Immediately fire an initial event, indicating all existing listeners\n // are in-sync.\n observer.next();\n }\n\n removeSnapshotsInSyncListener(observer: Observer<void>): void {\n this.snapshotsInSyncListeners.delete(observer);\n }\n\n // Call all global snapshot listeners that have been set.\n private raiseSnapshotsInSyncEvent(): void {\n this.snapshotsInSyncListeners.forEach(observer => {\n observer.next();\n });\n }\n}\n\nexport interface ListenOptions {\n /** Raise events even when only the metadata changes */\n readonly includeMetadataChanges?: boolean;\n\n /**\n * Wait for a sync with the server when online, but still raise events while\n * offline.\n */\n readonly waitForSyncWhenOnline?: boolean;\n}\n\n/**\n * QueryListener takes a series of internal view snapshots and determines\n * when to raise the event.\n *\n * It uses an Observer to dispatch events.\n */\nexport class QueryListener {\n /**\n * Initial snapshots (e.g. from cache) may not be propagated to the wrapped\n * observer. This flag is set to true once we've actually raised an event.\n */\n private raisedInitialEvent = false;\n\n private options: ListenOptions;\n\n private snap: ViewSnapshot | null = null;\n\n private onlineState = OnlineState.Unknown;\n\n constructor(\n readonly query: Query,\n private queryObserver: Observer<ViewSnapshot>,\n options?: ListenOptions\n ) {\n this.options = options || {};\n }\n\n /**\n * Applies the new ViewSnapshot to this listener, raising a user-facing event\n * if applicable (depending on what changed, whether the user has opted into\n * metadata-only changes, etc.). Returns true if a user-facing event was\n * indeed raised.\n */\n onViewSnapshot(snap: ViewSnapshot): boolean {\n debugAssert(\n snap.docChanges.length > 0 || snap.syncStateChanged,\n 'We got a new snapshot with no changes?'\n );\n\n if (!this.options.includeMetadataChanges) {\n // Remove the metadata only changes.\n const docChanges: DocumentViewChange[] = [];\n for (const docChange of snap.docChanges) {\n if (docChange.type !== ChangeType.Metadata) {\n docChanges.push(docChange);\n }\n }\n snap = new ViewSnapshot(\n snap.query,\n snap.docs,\n snap.oldDocs,\n docChanges,\n snap.mutatedKeys,\n snap.fromCache,\n snap.syncStateChanged,\n /* excludesMetadataChanges= */ true\n );\n }\n let raisedEvent = false;\n if (!this.raisedInitialEvent) {\n if (this.shouldRaiseInitialEvent(snap, this.onlineState)) {\n this.raiseInitialEvent(snap);\n raisedEvent = true;\n }\n } else if (this.shouldRaiseEvent(snap)) {\n this.queryObserver.next(snap);\n raisedEvent = true;\n }\n\n this.snap = snap;\n return raisedEvent;\n }\n\n onError(error: Error): void {\n this.queryObserver.error(error);\n }\n\n /** Returns whether a snapshot was raised. */\n applyOnlineStateChange(onlineState: OnlineState): boolean {\n this.onlineState = onlineState;\n let raisedEvent = false;\n if (\n this.snap &&\n !this.raisedInitialEvent &&\n this.shouldRaiseInitialEvent(this.snap, onlineState)\n ) {\n this.raiseInitialEvent(this.snap);\n raisedEvent = true;\n }\n return raisedEvent;\n }\n\n private shouldRaiseInitialEvent(\n snap: ViewSnapshot,\n onlineState: OnlineState\n ): boolean {\n debugAssert(\n !this.raisedInitialEvent,\n 'Determining whether to raise first event but already had first event'\n );\n\n // Always raise the first event when we're synced\n if (!snap.fromCache) {\n return true;\n }\n\n // NOTE: We consider OnlineState.Unknown as online (it should become Offline\n // or Online if we wait long enough).\n const maybeOnline = onlineState !== OnlineState.Offline;\n // Don't raise the event if we're online, aren't synced yet (checked\n // above) and are waiting for a sync.\n if (this.options.waitForSyncWhenOnline && maybeOnline) {\n debugAssert(\n snap.fromCache,\n 'Waiting for sync, but snapshot is not from cache'\n );\n return false;\n }\n\n // Raise data from cache if we have any documents or we are offline\n return !snap.docs.isEmpty() || onlineState === OnlineState.Offline;\n }\n\n private shouldRaiseEvent(snap: ViewSnapshot): boolean {\n // We don't need to handle includeDocumentMetadataChanges here because\n // the Metadata only changes have already been stripped out if needed.\n // At this point the only changes we will see are the ones we should\n // propagate.\n if (snap.docChanges.length > 0) {\n return true;\n }\n\n const hasPendingWritesChanged =\n this.snap && this.snap.hasPendingWrites !== snap.hasPendingWrites;\n if (snap.syncStateChanged || hasPendingWritesChanged) {\n return this.options.includeMetadataChanges === true;\n }\n\n // Generally we should have hit one of the cases above, but it's possible\n // to get here if there were only metadata docChanges and they got\n // stripped out.\n return false;\n }\n\n private raiseInitialEvent(snap: ViewSnapshot): void {\n debugAssert(\n !this.raisedInitialEvent,\n 'Trying to raise initial events for second time'\n );\n snap = ViewSnapshot.fromInitialDocuments(\n snap.query,\n snap.docs,\n snap.mutatedKeys,\n snap.fromCache\n );\n this.raisedInitialEvent = true;\n this.queryObserver.next(snap);\n }\n}\n","/**\n * @license\n * Copyright 2019 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { QueryEngine } from './query_engine';\nimport { LocalDocumentsView } from './local_documents_view';\nimport { PersistenceTransaction } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { LimitType, Query } from '../core/query';\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport {\n DocumentKeySet,\n DocumentMap,\n MaybeDocumentMap\n} from '../model/collections';\nimport { Document } from '../model/document';\nimport { debugAssert } from '../util/assert';\nimport { getLogLevel, LogLevel, logDebug } from '../util/log';\nimport { SortedSet } from '../util/sorted_set';\n\n// TOOD(b/140938512): Drop SimpleQueryEngine and rename IndexFreeQueryEngine.\n\n/**\n * A query engine that takes advantage of the target document mapping in the\n * QueryCache. The IndexFreeQueryEngine optimizes query execution by only\n * reading the documents that previously matched a query plus any documents that were\n * edited after the query was last listened to.\n *\n * There are some cases where Index-Free queries are not guaranteed to produce\n * the same results as full collection scans. In these cases, the\n * IndexFreeQueryEngine falls back to full query processing. These cases are:\n *\n * - Limit queries where a document that matched the query previously no longer\n * matches the query.\n *\n * - Limit queries where a document edit may cause the document to sort below\n * another document that is in the local cache.\n *\n * - Queries that have never been CURRENT or free of Limbo documents.\n */\nexport class IndexFreeQueryEngine implements QueryEngine {\n private localDocumentsView: LocalDocumentsView | undefined;\n\n setLocalDocumentsView(localDocuments: LocalDocumentsView): void {\n this.localDocumentsView = localDocuments;\n }\n\n getDocumentsMatchingQuery(\n transaction: PersistenceTransaction,\n query: Query,\n lastLimboFreeSnapshotVersion: SnapshotVersion,\n remoteKeys: DocumentKeySet\n ): PersistencePromise<DocumentMap> {\n debugAssert(\n this.localDocumentsView !== undefined,\n 'setLocalDocumentsView() not called'\n );\n\n // Queries that match all documents don't benefit from using\n // IndexFreeQueries. It is more efficient to scan all documents in a\n // collection, rather than to perform individual lookups.\n if (query.matchesAllDocuments()) {\n return this.executeFullCollectionScan(transaction, query);\n }\n\n // Queries that have never seen a snapshot without limbo free documents\n // should also be run as a full collection scan.\n if (lastLimboFreeSnapshotVersion.isEqual(SnapshotVersion.min())) {\n return this.executeFullCollectionScan(transaction, query);\n }\n\n return this.localDocumentsView!.getDocuments(transaction, remoteKeys).next(\n documents => {\n const previousResults = this.applyQuery(query, documents);\n\n if (\n (query.hasLimitToFirst() || query.hasLimitToLast()) &&\n this.needsRefill(\n query.limitType,\n previousResults,\n remoteKeys,\n lastLimboFreeSnapshotVersion\n )\n ) {\n return this.executeFullCollectionScan(transaction, query);\n }\n\n if (getLogLevel() <= LogLevel.DEBUG) {\n logDebug(\n 'IndexFreeQueryEngine',\n 'Re-using previous result from %s to execute query: %s',\n lastLimboFreeSnapshotVersion.toString(),\n query.toString()\n );\n }\n\n // Retrieve all results for documents that were updated since the last\n // limbo-document free remote snapshot.\n return this.localDocumentsView!.getDocumentsMatchingQuery(\n transaction,\n query,\n lastLimboFreeSnapshotVersion\n ).next(updatedResults => {\n // We merge `previousResults` into `updateResults`, since\n // `updateResults` is already a DocumentMap. If a document is\n // contained in both lists, then its contents are the same.\n previousResults.forEach(doc => {\n updatedResults = updatedResults.insert(doc.key, doc);\n });\n return updatedResults;\n });\n }\n );\n }\n\n /** Applies the query filter and sorting to the provided documents. */\n private applyQuery(\n query: Query,\n documents: MaybeDocumentMap\n ): SortedSet<Document> {\n // Sort the documents and re-apply the query filter since previously\n // matching documents do not necessarily still match the query.\n let queryResults = new SortedSet<Document>((d1, d2) =>\n query.docComparator(d1, d2)\n );\n documents.forEach((_, maybeDoc) => {\n if (maybeDoc instanceof Document && query.matches(maybeDoc)) {\n queryResults = queryResults.add(maybeDoc);\n }\n });\n return queryResults;\n }\n\n /**\n * Determines if a limit query needs to be refilled from cache, making it\n * ineligible for index-free execution.\n *\n * @param sortedPreviousResults The documents that matched the query when it\n * was last synchronized, sorted by the query's comparator.\n * @param remoteKeys The document keys that matched the query at the last\n * snapshot.\n * @param limboFreeSnapshotVersion The version of the snapshot when the query\n * was last synchronized.\n */\n private needsRefill(\n limitType: LimitType,\n sortedPreviousResults: SortedSet<Document>,\n remoteKeys: DocumentKeySet,\n limboFreeSnapshotVersion: SnapshotVersion\n ): boolean {\n // The query needs to be refilled if a previously matching document no\n // longer matches.\n if (remoteKeys.size !== sortedPreviousResults.size) {\n return true;\n }\n\n // Limit queries are not eligible for index-free query execution if there is\n // a potential that an older document from cache now sorts before a document\n // that was previously part of the limit. This, however, can only happen if\n // the document at the edge of the limit goes out of limit.\n // If a document that is not the limit boundary sorts differently,\n // the boundary of the limit itself did not change and documents from cache\n // will continue to be \"rejected\" by this boundary. Therefore, we can ignore\n // any modifications that don't affect the last document.\n const docAtLimitEdge =\n limitType === LimitType.First\n ? sortedPreviousResults.last()\n : sortedPreviousResults.first();\n if (!docAtLimitEdge) {\n // We don't need to refill the query if there were already no documents.\n return false;\n }\n return (\n docAtLimitEdge.hasPendingWrites ||\n docAtLimitEdge.version.compareTo(limboFreeSnapshotVersion) > 0\n );\n }\n\n private executeFullCollectionScan(\n transaction: PersistenceTransaction,\n query: Query\n ): PersistencePromise<DocumentMap> {\n if (getLogLevel() <= LogLevel.DEBUG) {\n logDebug(\n 'IndexFreeQueryEngine',\n 'Using full collection scan to execute query:',\n query.toString()\n );\n }\n\n return this.localDocumentsView!.getDocumentsMatchingQuery(\n transaction,\n query,\n SnapshotVersion.min()\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Timestamp } from '../api/timestamp';\nimport { Query } from '../core/query';\nimport { BatchId } from '../core/types';\nimport { DocumentKey } from '../model/document_key';\nimport { Mutation } from '../model/mutation';\nimport { MutationBatch, BATCHID_UNKNOWN } from '../model/mutation_batch';\nimport { debugAssert, hardAssert } from '../util/assert';\nimport { primitiveComparator } from '../util/misc';\nimport { ByteString } from '../util/byte_string';\nimport { SortedMap } from '../util/sorted_map';\nimport { SortedSet } from '../util/sorted_set';\n\nimport { IndexManager } from './index_manager';\nimport { MutationQueue } from './mutation_queue';\nimport { PersistenceTransaction, ReferenceDelegate } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { DocReference } from './reference_set';\n\nexport class MemoryMutationQueue implements MutationQueue {\n /**\n * The set of all mutations that have been sent but not yet been applied to\n * the backend.\n */\n private mutationQueue: MutationBatch[] = [];\n\n /** Next value to use when assigning sequential IDs to each mutation batch. */\n private nextBatchId: BatchId = 1;\n\n /** The last received stream token from the server, used to acknowledge which\n * responses the client has processed. Stream tokens are opaque checkpoint\n * markers whose only real value is their inclusion in the next request.\n */\n private lastStreamToken: ByteString = ByteString.EMPTY_BYTE_STRING;\n\n /** An ordered mapping between documents and the mutations batch IDs. */\n private batchesByDocumentKey = new SortedSet(DocReference.compareByKey);\n\n constructor(\n private readonly indexManager: IndexManager,\n private readonly referenceDelegate: ReferenceDelegate\n ) {}\n\n checkEmpty(transaction: PersistenceTransaction): PersistencePromise<boolean> {\n return PersistencePromise.resolve(this.mutationQueue.length === 0);\n }\n\n acknowledgeBatch(\n transaction: PersistenceTransaction,\n batch: MutationBatch,\n streamToken: ByteString\n ): PersistencePromise<void> {\n const batchId = batch.batchId;\n const batchIndex = this.indexOfExistingBatchId(batchId, 'acknowledged');\n hardAssert(\n batchIndex === 0,\n 'Can only acknowledge the first batch in the mutation queue'\n );\n\n // Verify that the batch in the queue is the one to be acknowledged.\n const check = this.mutationQueue[batchIndex];\n debugAssert(\n batchId === check.batchId,\n 'Queue ordering failure: expected batch ' +\n batchId +\n ', got batch ' +\n check.batchId\n );\n\n this.lastStreamToken = streamToken;\n return PersistencePromise.resolve();\n }\n\n getLastStreamToken(\n transaction: PersistenceTransaction\n ): PersistencePromise<ByteString> {\n return PersistencePromise.resolve(this.lastStreamToken);\n }\n\n setLastStreamToken(\n transaction: PersistenceTransaction,\n streamToken: ByteString\n ): PersistencePromise<void> {\n this.lastStreamToken = streamToken;\n return PersistencePromise.resolve();\n }\n\n addMutationBatch(\n transaction: PersistenceTransaction,\n localWriteTime: Timestamp,\n baseMutations: Mutation[],\n mutations: Mutation[]\n ): PersistencePromise<MutationBatch> {\n debugAssert(mutations.length !== 0, 'Mutation batches should not be empty');\n\n const batchId = this.nextBatchId;\n this.nextBatchId++;\n\n if (this.mutationQueue.length > 0) {\n const prior = this.mutationQueue[this.mutationQueue.length - 1];\n debugAssert(\n prior.batchId < batchId,\n 'Mutation batchIDs must be monotonically increasing order'\n );\n }\n\n const batch = new MutationBatch(\n batchId,\n localWriteTime,\n baseMutations,\n mutations\n );\n this.mutationQueue.push(batch);\n\n // Track references by document key and index collection parents.\n for (const mutation of mutations) {\n this.batchesByDocumentKey = this.batchesByDocumentKey.add(\n new DocReference(mutation.key, batchId)\n );\n\n this.indexManager.addToCollectionParentIndex(\n transaction,\n mutation.key.path.popLast()\n );\n }\n\n return PersistencePromise.resolve(batch);\n }\n\n lookupMutationBatch(\n transaction: PersistenceTransaction,\n batchId: BatchId\n ): PersistencePromise<MutationBatch | null> {\n return PersistencePromise.resolve(this.findMutationBatch(batchId));\n }\n\n getNextMutationBatchAfterBatchId(\n transaction: PersistenceTransaction,\n batchId: BatchId\n ): PersistencePromise<MutationBatch | null> {\n const nextBatchId = batchId + 1;\n\n // The requested batchId may still be out of range so normalize it to the\n // start of the queue.\n const rawIndex = this.indexOfBatchId(nextBatchId);\n const index = rawIndex < 0 ? 0 : rawIndex;\n return PersistencePromise.resolve(\n this.mutationQueue.length > index ? this.mutationQueue[index] : null\n );\n }\n\n getHighestUnacknowledgedBatchId(): PersistencePromise<BatchId> {\n return PersistencePromise.resolve(\n this.mutationQueue.length === 0 ? BATCHID_UNKNOWN : this.nextBatchId - 1\n );\n }\n\n getAllMutationBatches(\n transaction: PersistenceTransaction\n ): PersistencePromise<MutationBatch[]> {\n return PersistencePromise.resolve(this.mutationQueue.slice());\n }\n\n getAllMutationBatchesAffectingDocumentKey(\n transaction: PersistenceTransaction,\n documentKey: DocumentKey\n ): PersistencePromise<MutationBatch[]> {\n const start = new DocReference(documentKey, 0);\n const end = new DocReference(documentKey, Number.POSITIVE_INFINITY);\n const result: MutationBatch[] = [];\n this.batchesByDocumentKey.forEachInRange([start, end], ref => {\n debugAssert(\n documentKey.isEqual(ref.key),\n \"Should only iterate over a single key's batches\"\n );\n const batch = this.findMutationBatch(ref.targetOrBatchId);\n debugAssert(\n batch !== null,\n 'Batches in the index must exist in the main table'\n );\n result.push(batch!);\n });\n\n return PersistencePromise.resolve(result);\n }\n\n getAllMutationBatchesAffectingDocumentKeys(\n transaction: PersistenceTransaction,\n documentKeys: SortedMap<DocumentKey, unknown>\n ): PersistencePromise<MutationBatch[]> {\n let uniqueBatchIDs = new SortedSet<number>(primitiveComparator);\n\n documentKeys.forEach(documentKey => {\n const start = new DocReference(documentKey, 0);\n const end = new DocReference(documentKey, Number.POSITIVE_INFINITY);\n this.batchesByDocumentKey.forEachInRange([start, end], ref => {\n debugAssert(\n documentKey.isEqual(ref.key),\n \"For each key, should only iterate over a single key's batches\"\n );\n\n uniqueBatchIDs = uniqueBatchIDs.add(ref.targetOrBatchId);\n });\n });\n\n return PersistencePromise.resolve(this.findMutationBatches(uniqueBatchIDs));\n }\n\n getAllMutationBatchesAffectingQuery(\n transaction: PersistenceTransaction,\n query: Query\n ): PersistencePromise<MutationBatch[]> {\n debugAssert(\n !query.isCollectionGroupQuery(),\n 'CollectionGroup queries should be handled in LocalDocumentsView'\n );\n // Use the query path as a prefix for testing if a document matches the\n // query.\n const prefix = query.path;\n const immediateChildrenPathLength = prefix.length + 1;\n\n // Construct a document reference for actually scanning the index. Unlike\n // the prefix the document key in this reference must have an even number of\n // segments. The empty segment can be used a suffix of the query path\n // because it precedes all other segments in an ordered traversal.\n let startPath = prefix;\n if (!DocumentKey.isDocumentKey(startPath)) {\n startPath = startPath.child('');\n }\n\n const start = new DocReference(new DocumentKey(startPath), 0);\n\n // Find unique batchIDs referenced by all documents potentially matching the\n // query.\n let uniqueBatchIDs = new SortedSet<number>(primitiveComparator);\n\n this.batchesByDocumentKey.forEachWhile(ref => {\n const rowKeyPath = ref.key.path;\n if (!prefix.isPrefixOf(rowKeyPath)) {\n return false;\n } else {\n // Rows with document keys more than one segment longer than the query\n // path can't be matches. For example, a query on 'rooms' can't match\n // the document /rooms/abc/messages/xyx.\n // TODO(mcg): we'll need a different scanner when we implement\n // ancestor queries.\n if (rowKeyPath.length === immediateChildrenPathLength) {\n uniqueBatchIDs = uniqueBatchIDs.add(ref.targetOrBatchId);\n }\n return true;\n }\n }, start);\n\n return PersistencePromise.resolve(this.findMutationBatches(uniqueBatchIDs));\n }\n\n private findMutationBatches(batchIDs: SortedSet<number>): MutationBatch[] {\n // Construct an array of matching batches, sorted by batchID to ensure that\n // multiple mutations affecting the same document key are applied in order.\n const result: MutationBatch[] = [];\n batchIDs.forEach(batchId => {\n const batch = this.findMutationBatch(batchId);\n if (batch !== null) {\n result.push(batch);\n }\n });\n return result;\n }\n\n removeMutationBatch(\n transaction: PersistenceTransaction,\n batch: MutationBatch\n ): PersistencePromise<void> {\n // Find the position of the first batch for removal.\n const batchIndex = this.indexOfExistingBatchId(batch.batchId, 'removed');\n hardAssert(\n batchIndex === 0,\n 'Can only remove the first entry of the mutation queue'\n );\n this.mutationQueue.shift();\n\n let references = this.batchesByDocumentKey;\n return PersistencePromise.forEach(batch.mutations, (mutation: Mutation) => {\n const ref = new DocReference(mutation.key, batch.batchId);\n references = references.delete(ref);\n return this.referenceDelegate.markPotentiallyOrphaned(\n transaction,\n mutation.key\n );\n }).next(() => {\n this.batchesByDocumentKey = references;\n });\n }\n\n removeCachedMutationKeys(batchId: BatchId): void {\n // No-op since the memory mutation queue does not maintain a separate cache.\n }\n\n containsKey(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<boolean> {\n const ref = new DocReference(key, 0);\n const firstRef = this.batchesByDocumentKey.firstAfterOrEqual(ref);\n return PersistencePromise.resolve(key.isEqual(firstRef && firstRef.key));\n }\n\n performConsistencyCheck(\n txn: PersistenceTransaction\n ): PersistencePromise<void> {\n if (this.mutationQueue.length === 0) {\n debugAssert(\n this.batchesByDocumentKey.isEmpty(),\n 'Document leak -- detected dangling mutation references when queue is empty.'\n );\n }\n return PersistencePromise.resolve();\n }\n\n /**\n * Finds the index of the given batchId in the mutation queue and asserts that\n * the resulting index is within the bounds of the queue.\n *\n * @param batchId The batchId to search for\n * @param action A description of what the caller is doing, phrased in passive\n * form (e.g. \"acknowledged\" in a routine that acknowledges batches).\n */\n private indexOfExistingBatchId(batchId: BatchId, action: string): number {\n const index = this.indexOfBatchId(batchId);\n debugAssert(\n index >= 0 && index < this.mutationQueue.length,\n 'Batches must exist to be ' + action\n );\n return index;\n }\n\n /**\n * Finds the index of the given batchId in the mutation queue. This operation\n * is O(1).\n *\n * @return The computed index of the batch with the given batchId, based on\n * the state of the queue. Note this index can be negative if the requested\n * batchId has already been remvoed from the queue or past the end of the\n * queue if the batchId is larger than the last added batch.\n */\n private indexOfBatchId(batchId: BatchId): number {\n if (this.mutationQueue.length === 0) {\n // As an index this is past the end of the queue\n return 0;\n }\n\n // Examine the front of the queue to figure out the difference between the\n // batchId and indexes in the array. Note that since the queue is ordered\n // by batchId, if the first batch has a larger batchId then the requested\n // batchId doesn't exist in the queue.\n const firstBatchId = this.mutationQueue[0].batchId;\n return batchId - firstBatchId;\n }\n\n /**\n * A version of lookupMutationBatch that doesn't return a promise, this makes\n * other functions that uses this code easier to read and more efficent.\n */\n private findMutationBatch(batchId: BatchId): MutationBatch | null {\n const index = this.indexOfBatchId(batchId);\n if (index < 0 || index >= this.mutationQueue.length) {\n return null;\n }\n\n const batch = this.mutationQueue[index];\n debugAssert(batch.batchId === batchId, 'If found batch must match');\n return batch;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Query } from '../core/query';\nimport {\n DocumentKeySet,\n DocumentMap,\n documentMap,\n DocumentSizeEntry,\n NullableMaybeDocumentMap,\n nullableMaybeDocumentMap\n} from '../model/collections';\nimport { Document, MaybeDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { debugAssert } from '../util/assert';\nimport { SortedMap } from '../util/sorted_map';\nimport { IndexManager } from './index_manager';\nimport { PersistenceTransaction } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { RemoteDocumentCache } from './remote_document_cache';\nimport { RemoteDocumentChangeBuffer } from './remote_document_change_buffer';\n\nexport type DocumentSizer = (doc: MaybeDocument) => number;\n\n/** Miscellaneous collection types / constants. */\ninterface MemoryRemoteDocumentCacheEntry extends DocumentSizeEntry {\n readTime: SnapshotVersion;\n}\n\ntype DocumentEntryMap = SortedMap<DocumentKey, MemoryRemoteDocumentCacheEntry>;\nfunction documentEntryMap(): DocumentEntryMap {\n return new SortedMap<DocumentKey, MemoryRemoteDocumentCacheEntry>(\n DocumentKey.comparator\n );\n}\n\nexport class MemoryRemoteDocumentCache implements RemoteDocumentCache {\n /** Underlying cache of documents and their read times. */\n private docs = documentEntryMap();\n\n /** Size of all cached documents. */\n private size = 0;\n\n /**\n * @param sizer Used to assess the size of a document. For eager GC, this is expected to just\n * return 0 to avoid unnecessarily doing the work of calculating the size.\n */\n constructor(\n private readonly indexManager: IndexManager,\n private readonly sizer: DocumentSizer\n ) {}\n\n /**\n * Adds the supplied entry to the cache and updates the cache size as appropriate.\n *\n * All calls of `addEntry` are required to go through the RemoteDocumentChangeBuffer\n * returned by `newChangeBuffer()`.\n */\n private addEntry(\n transaction: PersistenceTransaction,\n doc: MaybeDocument,\n readTime: SnapshotVersion\n ): PersistencePromise<void> {\n debugAssert(\n !readTime.isEqual(SnapshotVersion.min()),\n 'Cannot add a document with a read time of zero'\n );\n\n const key = doc.key;\n const entry = this.docs.get(key);\n const previousSize = entry ? entry.size : 0;\n const currentSize = this.sizer(doc);\n\n this.docs = this.docs.insert(key, {\n maybeDocument: doc,\n size: currentSize,\n readTime\n });\n\n this.size += currentSize - previousSize;\n\n return this.indexManager.addToCollectionParentIndex(\n transaction,\n key.path.popLast()\n );\n }\n\n /**\n * Removes the specified entry from the cache and updates the cache size as appropriate.\n *\n * All calls of `removeEntry` are required to go through the RemoteDocumentChangeBuffer\n * returned by `newChangeBuffer()`.\n */\n private removeEntry(documentKey: DocumentKey): void {\n const entry = this.docs.get(documentKey);\n if (entry) {\n this.docs = this.docs.remove(documentKey);\n this.size -= entry.size;\n }\n }\n\n getEntry(\n transaction: PersistenceTransaction,\n documentKey: DocumentKey\n ): PersistencePromise<MaybeDocument | null> {\n const entry = this.docs.get(documentKey);\n return PersistencePromise.resolve(entry ? entry.maybeDocument : null);\n }\n\n getEntries(\n transaction: PersistenceTransaction,\n documentKeys: DocumentKeySet\n ): PersistencePromise<NullableMaybeDocumentMap> {\n let results = nullableMaybeDocumentMap();\n documentKeys.forEach(documentKey => {\n const entry = this.docs.get(documentKey);\n results = results.insert(documentKey, entry ? entry.maybeDocument : null);\n });\n return PersistencePromise.resolve(results);\n }\n\n getDocumentsMatchingQuery(\n transaction: PersistenceTransaction,\n query: Query,\n sinceReadTime: SnapshotVersion\n ): PersistencePromise<DocumentMap> {\n debugAssert(\n !query.isCollectionGroupQuery(),\n 'CollectionGroup queries should be handled in LocalDocumentsView'\n );\n let results = documentMap();\n\n // Documents are ordered by key, so we can use a prefix scan to narrow down\n // the documents we need to match the query against.\n const prefix = new DocumentKey(query.path.child(''));\n const iterator = this.docs.getIteratorFrom(prefix);\n while (iterator.hasNext()) {\n const {\n key,\n value: { maybeDocument, readTime }\n } = iterator.getNext();\n if (!query.path.isPrefixOf(key.path)) {\n break;\n }\n if (readTime.compareTo(sinceReadTime) <= 0) {\n continue;\n }\n if (maybeDocument instanceof Document && query.matches(maybeDocument)) {\n results = results.insert(maybeDocument.key, maybeDocument);\n }\n }\n return PersistencePromise.resolve(results);\n }\n\n forEachDocumentKey(\n transaction: PersistenceTransaction,\n f: (key: DocumentKey) => PersistencePromise<void>\n ): PersistencePromise<void> {\n return PersistencePromise.forEach(this.docs, (key: DocumentKey) => f(key));\n }\n\n newChangeBuffer(options?: {\n trackRemovals: boolean;\n }): RemoteDocumentChangeBuffer {\n // `trackRemovals` is ignores since the MemoryRemoteDocumentCache keeps\n // a separate changelog and does not need special handling for removals.\n return new MemoryRemoteDocumentCache.RemoteDocumentChangeBuffer(this);\n }\n\n getSize(txn: PersistenceTransaction): PersistencePromise<number> {\n return PersistencePromise.resolve(this.size);\n }\n\n /**\n * Handles the details of adding and updating documents in the MemoryRemoteDocumentCache.\n */\n private static RemoteDocumentChangeBuffer = class extends RemoteDocumentChangeBuffer {\n constructor(private readonly documentCache: MemoryRemoteDocumentCache) {\n super();\n }\n\n protected applyChanges(\n transaction: PersistenceTransaction\n ): PersistencePromise<void> {\n const promises: Array<PersistencePromise<void>> = [];\n this.changes.forEach((key, doc) => {\n if (doc) {\n promises.push(\n this.documentCache.addEntry(transaction, doc, this.readTime)\n );\n } else {\n this.documentCache.removeEntry(key);\n }\n });\n return PersistencePromise.waitFor(promises);\n }\n\n protected getFromCache(\n transaction: PersistenceTransaction,\n documentKey: DocumentKey\n ): PersistencePromise<MaybeDocument | null> {\n return this.documentCache.getEntry(transaction, documentKey);\n }\n\n protected getAllFromCache(\n transaction: PersistenceTransaction,\n documentKeys: DocumentKeySet\n ): PersistencePromise<NullableMaybeDocumentMap> {\n return this.documentCache.getEntries(transaction, documentKeys);\n }\n };\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { DocumentKeySet, NullableMaybeDocumentMap } from '../model/collections';\nimport { MaybeDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { debugAssert } from '../util/assert';\nimport { ObjectMap } from '../util/obj_map';\n\nimport { PersistenceTransaction } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { SnapshotVersion } from '../core/snapshot_version';\n\n/**\n * An in-memory buffer of entries to be written to a RemoteDocumentCache.\n * It can be used to batch up a set of changes to be written to the cache, but\n * additionally supports reading entries back with the `getEntry()` method,\n * falling back to the underlying RemoteDocumentCache if no entry is\n * buffered.\n *\n * Entries added to the cache *must* be read first. This is to facilitate\n * calculating the size delta of the pending changes.\n *\n * PORTING NOTE: This class was implemented then removed from other platforms.\n * If byte-counting ends up being needed on the other platforms, consider\n * porting this class as part of that implementation work.\n */\nexport abstract class RemoteDocumentChangeBuffer {\n // A mapping of document key to the new cache entry that should be written (or null if any\n // existing cache entry should be removed).\n protected changes: ObjectMap<\n DocumentKey,\n MaybeDocument | null\n > = new ObjectMap(key => key.toString());\n\n // The read time to use for all added documents in this change buffer.\n private _readTime: SnapshotVersion | undefined;\n\n private changesApplied = false;\n\n protected abstract getFromCache(\n transaction: PersistenceTransaction,\n documentKey: DocumentKey\n ): PersistencePromise<MaybeDocument | null>;\n\n protected abstract getAllFromCache(\n transaction: PersistenceTransaction,\n documentKeys: DocumentKeySet\n ): PersistencePromise<NullableMaybeDocumentMap>;\n\n protected abstract applyChanges(\n transaction: PersistenceTransaction\n ): PersistencePromise<void>;\n\n protected set readTime(value: SnapshotVersion) {\n // Right now (for simplicity) we just track a single readTime for all the\n // added entries since we expect them to all be the same, but we could\n // rework to store per-entry readTimes if necessary.\n debugAssert(\n this._readTime === undefined || this._readTime.isEqual(value),\n 'All changes in a RemoteDocumentChangeBuffer must have the same read time'\n );\n this._readTime = value;\n }\n\n protected get readTime(): SnapshotVersion {\n debugAssert(\n this._readTime !== undefined,\n 'Read time is not set. All removeEntry() calls must include a readTime if `trackRemovals` is used.'\n );\n return this._readTime;\n }\n\n /**\n * Buffers a `RemoteDocumentCache.addEntry()` call.\n *\n * You can only modify documents that have already been retrieved via\n * `getEntry()/getEntries()` (enforced via IndexedDbs `apply()`).\n */\n addEntry(maybeDocument: MaybeDocument, readTime: SnapshotVersion): void {\n this.assertNotApplied();\n this.readTime = readTime;\n this.changes.set(maybeDocument.key, maybeDocument);\n }\n\n /**\n * Buffers a `RemoteDocumentCache.removeEntry()` call.\n *\n * You can only remove documents that have already been retrieved via\n * `getEntry()/getEntries()` (enforced via IndexedDbs `apply()`).\n */\n removeEntry(key: DocumentKey, readTime?: SnapshotVersion): void {\n this.assertNotApplied();\n if (readTime) {\n this.readTime = readTime;\n }\n this.changes.set(key, null);\n }\n\n /**\n * Looks up an entry in the cache. The buffered changes will first be checked,\n * and if no buffered change applies, this will forward to\n * `RemoteDocumentCache.getEntry()`.\n *\n * @param transaction The transaction in which to perform any persistence\n * operations.\n * @param documentKey The key of the entry to look up.\n * @return The cached Document or NoDocument entry, or null if we have nothing\n * cached.\n */\n getEntry(\n transaction: PersistenceTransaction,\n documentKey: DocumentKey\n ): PersistencePromise<MaybeDocument | null> {\n this.assertNotApplied();\n const bufferedEntry = this.changes.get(documentKey);\n if (bufferedEntry !== undefined) {\n return PersistencePromise.resolve<MaybeDocument | null>(bufferedEntry);\n } else {\n return this.getFromCache(transaction, documentKey);\n }\n }\n\n /**\n * Looks up several entries in the cache, forwarding to\n * `RemoteDocumentCache.getEntry()`.\n *\n * @param transaction The transaction in which to perform any persistence\n * operations.\n * @param documentKeys The keys of the entries to look up.\n * @return A map of cached `Document`s or `NoDocument`s, indexed by key. If an\n * entry cannot be found, the corresponding key will be mapped to a null\n * value.\n */\n getEntries(\n transaction: PersistenceTransaction,\n documentKeys: DocumentKeySet\n ): PersistencePromise<NullableMaybeDocumentMap> {\n return this.getAllFromCache(transaction, documentKeys);\n }\n\n /**\n * Applies buffered changes to the underlying RemoteDocumentCache, using\n * the provided transaction.\n */\n apply(transaction: PersistenceTransaction): PersistencePromise<void> {\n this.assertNotApplied();\n this.changesApplied = true;\n return this.applyChanges(transaction);\n }\n\n /** Helper to assert this.changes is not null */\n protected assertNotApplied(): void {\n debugAssert(!this.changesApplied, 'Changes have already been applied.');\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { SnapshotVersion } from '../core/snapshot_version';\nimport { TargetIdGenerator } from '../core/target_id_generator';\nimport { ListenSequenceNumber, TargetId } from '../core/types';\nimport { DocumentKeySet } from '../model/collections';\nimport { DocumentKey } from '../model/document_key';\nimport { debugAssert } from '../util/assert';\nimport { ObjectMap } from '../util/obj_map';\n\nimport { ActiveTargets } from './lru_garbage_collector';\nimport { MemoryPersistence } from './memory_persistence';\nimport { PersistenceTransaction } from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { ReferenceSet } from './reference_set';\nimport { TargetCache } from './target_cache';\nimport { TargetData } from './target_data';\nimport { Target } from '../core/target';\n\nexport class MemoryTargetCache implements TargetCache {\n /**\n * Maps a target to the data about that target\n */\n private targets = new ObjectMap<Target, TargetData>(t => t.canonicalId());\n\n /** The last received snapshot version. */\n private lastRemoteSnapshotVersion = SnapshotVersion.min();\n /** The highest numbered target ID encountered. */\n private highestTargetId: TargetId = 0;\n /** The highest sequence number encountered. */\n private highestSequenceNumber: ListenSequenceNumber = 0;\n /**\n * A ordered bidirectional mapping between documents and the remote target\n * IDs.\n */\n private references = new ReferenceSet();\n\n private targetCount = 0;\n\n private targetIdGenerator = TargetIdGenerator.forTargetCache();\n\n constructor(private readonly persistence: MemoryPersistence) {}\n\n forEachTarget(\n txn: PersistenceTransaction,\n f: (q: TargetData) => void\n ): PersistencePromise<void> {\n this.targets.forEach((_, targetData) => f(targetData));\n return PersistencePromise.resolve();\n }\n\n getLastRemoteSnapshotVersion(\n transaction: PersistenceTransaction\n ): PersistencePromise<SnapshotVersion> {\n return PersistencePromise.resolve(this.lastRemoteSnapshotVersion);\n }\n\n getHighestSequenceNumber(\n transaction: PersistenceTransaction\n ): PersistencePromise<ListenSequenceNumber> {\n return PersistencePromise.resolve(this.highestSequenceNumber);\n }\n\n allocateTargetId(\n transaction: PersistenceTransaction\n ): PersistencePromise<TargetId> {\n this.highestTargetId = this.targetIdGenerator.next();\n return PersistencePromise.resolve(this.highestTargetId);\n }\n\n setTargetsMetadata(\n transaction: PersistenceTransaction,\n highestListenSequenceNumber: number,\n lastRemoteSnapshotVersion?: SnapshotVersion\n ): PersistencePromise<void> {\n if (lastRemoteSnapshotVersion) {\n this.lastRemoteSnapshotVersion = lastRemoteSnapshotVersion;\n }\n if (highestListenSequenceNumber > this.highestSequenceNumber) {\n this.highestSequenceNumber = highestListenSequenceNumber;\n }\n return PersistencePromise.resolve();\n }\n\n private saveTargetData(targetData: TargetData): void {\n this.targets.set(targetData.target, targetData);\n const targetId = targetData.targetId;\n if (targetId > this.highestTargetId) {\n this.targetIdGenerator = new TargetIdGenerator(targetId);\n this.highestTargetId = targetId;\n }\n if (targetData.sequenceNumber > this.highestSequenceNumber) {\n this.highestSequenceNumber = targetData.sequenceNumber;\n }\n }\n\n addTargetData(\n transaction: PersistenceTransaction,\n targetData: TargetData\n ): PersistencePromise<void> {\n debugAssert(\n !this.targets.has(targetData.target),\n 'Adding a target that already exists'\n );\n this.saveTargetData(targetData);\n this.targetCount += 1;\n return PersistencePromise.resolve();\n }\n\n updateTargetData(\n transaction: PersistenceTransaction,\n targetData: TargetData\n ): PersistencePromise<void> {\n debugAssert(\n this.targets.has(targetData.target),\n 'Updating a non-existent target'\n );\n this.saveTargetData(targetData);\n return PersistencePromise.resolve();\n }\n\n removeTargetData(\n transaction: PersistenceTransaction,\n targetData: TargetData\n ): PersistencePromise<void> {\n debugAssert(this.targetCount > 0, 'Removing a target from an empty cache');\n debugAssert(\n this.targets.has(targetData.target),\n 'Removing a non-existent target from the cache'\n );\n this.targets.delete(targetData.target);\n this.references.removeReferencesForId(targetData.targetId);\n this.targetCount -= 1;\n return PersistencePromise.resolve();\n }\n\n removeTargets(\n transaction: PersistenceTransaction,\n upperBound: ListenSequenceNumber,\n activeTargetIds: ActiveTargets\n ): PersistencePromise<number> {\n let count = 0;\n const removals: Array<PersistencePromise<void>> = [];\n this.targets.forEach((key, targetData) => {\n if (\n targetData.sequenceNumber <= upperBound &&\n activeTargetIds.get(targetData.targetId) === null\n ) {\n this.targets.delete(key);\n removals.push(\n this.removeMatchingKeysForTargetId(transaction, targetData.targetId)\n );\n count++;\n }\n });\n return PersistencePromise.waitFor(removals).next(() => count);\n }\n\n getTargetCount(\n transaction: PersistenceTransaction\n ): PersistencePromise<number> {\n return PersistencePromise.resolve(this.targetCount);\n }\n\n getTargetData(\n transaction: PersistenceTransaction,\n target: Target\n ): PersistencePromise<TargetData | null> {\n const targetData = this.targets.get(target) || null;\n return PersistencePromise.resolve(targetData);\n }\n\n addMatchingKeys(\n txn: PersistenceTransaction,\n keys: DocumentKeySet,\n targetId: TargetId\n ): PersistencePromise<void> {\n this.references.addReferences(keys, targetId);\n return PersistencePromise.resolve();\n }\n\n removeMatchingKeys(\n txn: PersistenceTransaction,\n keys: DocumentKeySet,\n targetId: TargetId\n ): PersistencePromise<void> {\n this.references.removeReferences(keys, targetId);\n const referenceDelegate = this.persistence.referenceDelegate;\n const promises: Array<PersistencePromise<void>> = [];\n if (referenceDelegate) {\n keys.forEach(key => {\n promises.push(referenceDelegate.markPotentiallyOrphaned(txn, key));\n });\n }\n return PersistencePromise.waitFor(promises);\n }\n\n removeMatchingKeysForTargetId(\n txn: PersistenceTransaction,\n targetId: TargetId\n ): PersistencePromise<void> {\n this.references.removeReferencesForId(targetId);\n return PersistencePromise.resolve();\n }\n\n getMatchingKeysForTargetId(\n txn: PersistenceTransaction,\n targetId: TargetId\n ): PersistencePromise<DocumentKeySet> {\n const matchingKeys = this.references.referencesForId(targetId);\n return PersistencePromise.resolve(matchingKeys);\n }\n\n containsKey(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<boolean> {\n return PersistencePromise.resolve(this.references.containsKey(key));\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { User } from '../auth/user';\nimport { Document, MaybeDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { fail } from '../util/assert';\nimport { logDebug } from '../util/log';\nimport { ObjectMap } from '../util/obj_map';\nimport { encodeResourcePath } from './encoded_resource_path';\nimport {\n ActiveTargets,\n LruDelegate,\n LruGarbageCollector,\n LruParams\n} from './lru_garbage_collector';\nimport { ListenSequence } from '../core/listen_sequence';\nimport { ListenSequenceNumber, TargetId } from '../core/types';\nimport { estimateByteSize } from '../model/values';\nimport { MemoryIndexManager } from './memory_index_manager';\nimport { MemoryMutationQueue } from './memory_mutation_queue';\nimport { MemoryRemoteDocumentCache } from './memory_remote_document_cache';\nimport { MemoryTargetCache } from './memory_target_cache';\nimport { MutationQueue } from './mutation_queue';\nimport {\n Persistence,\n PersistenceTransaction,\n PersistenceTransactionMode,\n ReferenceDelegate\n} from './persistence';\nimport { PersistencePromise } from './persistence_promise';\nimport { ReferenceSet } from './reference_set';\nimport { TargetData } from './target_data';\n\nconst LOG_TAG = 'MemoryPersistence';\n/**\n * A memory-backed instance of Persistence. Data is stored only in RAM and\n * not persisted across sessions.\n */\nexport class MemoryPersistence implements Persistence {\n /**\n * Note that these are retained here to make it easier to write tests\n * affecting both the in-memory and IndexedDB-backed persistence layers. Tests\n * can create a new LocalStore wrapping this Persistence instance and this\n * will make the in-memory persistence layer behave as if it were actually\n * persisting values.\n */\n private readonly indexManager: MemoryIndexManager;\n private mutationQueues: { [user: string]: MemoryMutationQueue } = {};\n private readonly remoteDocumentCache: MemoryRemoteDocumentCache;\n private readonly targetCache: MemoryTargetCache;\n private readonly listenSequence = new ListenSequence(0);\n\n private _started = false;\n\n readonly referenceDelegate: MemoryReferenceDelegate;\n\n /**\n * The constructor accepts a factory for creating a reference delegate. This\n * allows both the delegate and this instance to have strong references to\n * each other without having nullable fields that would then need to be\n * checked or asserted on every access.\n */\n constructor(\n referenceDelegateFactory: (p: MemoryPersistence) => MemoryReferenceDelegate\n ) {\n this._started = true;\n this.referenceDelegate = referenceDelegateFactory(this);\n this.targetCache = new MemoryTargetCache(this);\n const sizer = (doc: MaybeDocument): number =>\n this.referenceDelegate.documentSize(doc);\n this.indexManager = new MemoryIndexManager();\n this.remoteDocumentCache = new MemoryRemoteDocumentCache(\n this.indexManager,\n sizer\n );\n }\n\n start(): Promise<void> {\n return Promise.resolve();\n }\n\n shutdown(): Promise<void> {\n // No durable state to ensure is closed on shutdown.\n this._started = false;\n return Promise.resolve();\n }\n\n get started(): boolean {\n return this._started;\n }\n\n setDatabaseDeletedListener(): void {\n // No op.\n }\n\n getIndexManager(): MemoryIndexManager {\n return this.indexManager;\n }\n\n getMutationQueue(user: User): MutationQueue {\n let queue = this.mutationQueues[user.toKey()];\n if (!queue) {\n queue = new MemoryMutationQueue(\n this.indexManager,\n this.referenceDelegate\n );\n this.mutationQueues[user.toKey()] = queue;\n }\n return queue;\n }\n\n getTargetCache(): MemoryTargetCache {\n return this.targetCache;\n }\n\n getRemoteDocumentCache(): MemoryRemoteDocumentCache {\n return this.remoteDocumentCache;\n }\n\n runTransaction<T>(\n action: string,\n mode: PersistenceTransactionMode,\n transactionOperation: (\n transaction: PersistenceTransaction\n ) => PersistencePromise<T>\n ): Promise<T> {\n logDebug(LOG_TAG, 'Starting transaction:', action);\n const txn = new MemoryTransaction(this.listenSequence.next());\n this.referenceDelegate.onTransactionStarted();\n return transactionOperation(txn)\n .next(result => {\n return this.referenceDelegate\n .onTransactionCommitted(txn)\n .next(() => result);\n })\n .toPromise()\n .then(result => {\n txn.raiseOnCommittedEvent();\n return result;\n });\n }\n\n mutationQueuesContainKey(\n transaction: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<boolean> {\n return PersistencePromise.or(\n Object.values(this.mutationQueues).map(queue => () =>\n queue.containsKey(transaction, key)\n )\n );\n }\n}\n\n/**\n * Memory persistence is not actually transactional, but future implementations\n * may have transaction-scoped state.\n */\nexport class MemoryTransaction extends PersistenceTransaction {\n constructor(readonly currentSequenceNumber: ListenSequenceNumber) {\n super();\n }\n}\n\nexport interface MemoryReferenceDelegate extends ReferenceDelegate {\n documentSize(doc: MaybeDocument): number;\n onTransactionStarted(): void;\n onTransactionCommitted(txn: PersistenceTransaction): PersistencePromise<void>;\n}\n\nexport class MemoryEagerDelegate implements MemoryReferenceDelegate {\n /** Tracks all documents that are active in Query views. */\n private localViewReferences: ReferenceSet = new ReferenceSet();\n /** The list of documents that are potentially GCed after each transaction. */\n private _orphanedDocuments: Set<DocumentKey> | null = null;\n\n private constructor(private readonly persistence: MemoryPersistence) {}\n\n static factory(persistence: MemoryPersistence): MemoryEagerDelegate {\n return new MemoryEagerDelegate(persistence);\n }\n\n private get orphanedDocuments(): Set<DocumentKey> {\n if (!this._orphanedDocuments) {\n throw fail('orphanedDocuments is only valid during a transaction.');\n } else {\n return this._orphanedDocuments;\n }\n }\n\n addReference(\n txn: PersistenceTransaction,\n targetId: TargetId,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.localViewReferences.addReference(key, targetId);\n this.orphanedDocuments.delete(key);\n return PersistencePromise.resolve();\n }\n\n removeReference(\n txn: PersistenceTransaction,\n targetId: TargetId,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.localViewReferences.removeReference(key, targetId);\n this.orphanedDocuments.add(key);\n return PersistencePromise.resolve();\n }\n\n markPotentiallyOrphaned(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.orphanedDocuments.add(key);\n return PersistencePromise.resolve();\n }\n\n removeTarget(\n txn: PersistenceTransaction,\n targetData: TargetData\n ): PersistencePromise<void> {\n const orphaned = this.localViewReferences.removeReferencesForId(\n targetData.targetId\n );\n orphaned.forEach(key => this.orphanedDocuments.add(key));\n const cache = this.persistence.getTargetCache();\n return cache\n .getMatchingKeysForTargetId(txn, targetData.targetId)\n .next(keys => {\n keys.forEach(key => this.orphanedDocuments.add(key));\n })\n .next(() => cache.removeTargetData(txn, targetData));\n }\n\n onTransactionStarted(): void {\n this._orphanedDocuments = new Set<DocumentKey>();\n }\n\n onTransactionCommitted(\n txn: PersistenceTransaction\n ): PersistencePromise<void> {\n // Remove newly orphaned documents.\n const cache = this.persistence.getRemoteDocumentCache();\n const changeBuffer = cache.newChangeBuffer();\n return PersistencePromise.forEach(\n this.orphanedDocuments,\n (key: DocumentKey) => {\n return this.isReferenced(txn, key).next(isReferenced => {\n if (!isReferenced) {\n changeBuffer.removeEntry(key);\n }\n });\n }\n ).next(() => {\n this._orphanedDocuments = null;\n return changeBuffer.apply(txn);\n });\n }\n\n updateLimboDocument(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<void> {\n return this.isReferenced(txn, key).next(isReferenced => {\n if (isReferenced) {\n this.orphanedDocuments.delete(key);\n } else {\n this.orphanedDocuments.add(key);\n }\n });\n }\n\n documentSize(doc: MaybeDocument): number {\n // For eager GC, we don't care about the document size, there are no size thresholds.\n return 0;\n }\n\n private isReferenced(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<boolean> {\n return PersistencePromise.or([\n () =>\n PersistencePromise.resolve(this.localViewReferences.containsKey(key)),\n () => this.persistence.getTargetCache().containsKey(txn, key),\n () => this.persistence.mutationQueuesContainKey(txn, key)\n ]);\n }\n}\n\nexport class MemoryLruDelegate implements ReferenceDelegate, LruDelegate {\n private orphanedSequenceNumbers: ObjectMap<\n DocumentKey,\n ListenSequenceNumber\n > = new ObjectMap(k => encodeResourcePath(k.path));\n\n readonly garbageCollector: LruGarbageCollector;\n\n constructor(\n private readonly persistence: MemoryPersistence,\n lruParams: LruParams\n ) {\n this.garbageCollector = new LruGarbageCollector(this, lruParams);\n }\n\n // No-ops, present so memory persistence doesn't have to care which delegate\n // it has.\n onTransactionStarted(): void {}\n\n onTransactionCommitted(\n txn: PersistenceTransaction\n ): PersistencePromise<void> {\n return PersistencePromise.resolve();\n }\n\n forEachTarget(\n txn: PersistenceTransaction,\n f: (q: TargetData) => void\n ): PersistencePromise<void> {\n return this.persistence.getTargetCache().forEachTarget(txn, f);\n }\n\n getSequenceNumberCount(\n txn: PersistenceTransaction\n ): PersistencePromise<number> {\n const docCountPromise = this.orphanedDocumentCount(txn);\n const targetCountPromise = this.persistence\n .getTargetCache()\n .getTargetCount(txn);\n return targetCountPromise.next(targetCount =>\n docCountPromise.next(docCount => targetCount + docCount)\n );\n }\n\n private orphanedDocumentCount(\n txn: PersistenceTransaction\n ): PersistencePromise<number> {\n let orphanedCount = 0;\n return this.forEachOrphanedDocumentSequenceNumber(txn, _ => {\n orphanedCount++;\n }).next(() => orphanedCount);\n }\n\n forEachOrphanedDocumentSequenceNumber(\n txn: PersistenceTransaction,\n f: (sequenceNumber: ListenSequenceNumber) => void\n ): PersistencePromise<void> {\n return PersistencePromise.forEach(\n this.orphanedSequenceNumbers,\n (key, sequenceNumber) => {\n // Pass in the exact sequence number as the upper bound so we know it won't be pinned by\n // being too recent.\n return this.isPinned(txn, key, sequenceNumber).next(isPinned => {\n if (!isPinned) {\n return f(sequenceNumber);\n } else {\n return PersistencePromise.resolve();\n }\n });\n }\n );\n }\n\n removeTargets(\n txn: PersistenceTransaction,\n upperBound: ListenSequenceNumber,\n activeTargetIds: ActiveTargets\n ): PersistencePromise<number> {\n return this.persistence\n .getTargetCache()\n .removeTargets(txn, upperBound, activeTargetIds);\n }\n\n removeOrphanedDocuments(\n txn: PersistenceTransaction,\n upperBound: ListenSequenceNumber\n ): PersistencePromise<number> {\n let count = 0;\n const cache = this.persistence.getRemoteDocumentCache();\n const changeBuffer = cache.newChangeBuffer();\n const p = cache.forEachDocumentKey(txn, key => {\n return this.isPinned(txn, key, upperBound).next(isPinned => {\n if (!isPinned) {\n count++;\n changeBuffer.removeEntry(key);\n }\n });\n });\n return p.next(() => changeBuffer.apply(txn)).next(() => count);\n }\n\n markPotentiallyOrphaned(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.orphanedSequenceNumbers.set(key, txn.currentSequenceNumber);\n return PersistencePromise.resolve();\n }\n\n removeTarget(\n txn: PersistenceTransaction,\n targetData: TargetData\n ): PersistencePromise<void> {\n const updated = targetData.withSequenceNumber(txn.currentSequenceNumber);\n return this.persistence.getTargetCache().updateTargetData(txn, updated);\n }\n\n addReference(\n txn: PersistenceTransaction,\n targetId: TargetId,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.orphanedSequenceNumbers.set(key, txn.currentSequenceNumber);\n return PersistencePromise.resolve();\n }\n\n removeReference(\n txn: PersistenceTransaction,\n targetId: TargetId,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.orphanedSequenceNumbers.set(key, txn.currentSequenceNumber);\n return PersistencePromise.resolve();\n }\n\n updateLimboDocument(\n txn: PersistenceTransaction,\n key: DocumentKey\n ): PersistencePromise<void> {\n this.orphanedSequenceNumbers.set(key, txn.currentSequenceNumber);\n return PersistencePromise.resolve();\n }\n\n documentSize(maybeDoc: MaybeDocument): number {\n let documentSize = maybeDoc.key.toString().length;\n if (maybeDoc instanceof Document) {\n documentSize += estimateByteSize(maybeDoc.toProto());\n }\n return documentSize;\n }\n\n private isPinned(\n txn: PersistenceTransaction,\n key: DocumentKey,\n upperBound: ListenSequenceNumber\n ): PersistencePromise<boolean> {\n return PersistencePromise.or([\n () => this.persistence.mutationQueuesContainKey(txn, key),\n () => this.persistence.getTargetCache().containsKey(txn, key),\n () => {\n const orphanedAt = this.orphanedSequenceNumbers.get(key);\n return PersistencePromise.resolve(\n orphanedAt !== undefined && orphanedAt > upperBound\n );\n }\n ]);\n }\n\n getCacheSize(txn: PersistenceTransaction): PersistencePromise<number> {\n return this.persistence.getRemoteDocumentCache().getSize(txn);\n }\n}\n","/**\n * @license\n * Copyright 2020 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport {\n ClientId,\n MemorySharedClientState,\n SharedClientState,\n WebStorageSharedClientState\n} from '../local/shared_client_state';\nimport { LocalStore, MultiTabLocalStore } from '../local/local_store';\nimport { MultiTabSyncEngine, SyncEngine } from './sync_engine';\nimport { RemoteStore } from '../remote/remote_store';\nimport { EventManager } from './event_manager';\nimport { AsyncQueue } from '../util/async_queue';\nimport { DatabaseInfo } from './database_info';\nimport { Platform } from '../platform/platform';\nimport { Datastore } from '../remote/datastore';\nimport { User } from '../auth/user';\nimport { PersistenceSettings } from './firestore_client';\nimport { debugAssert } from '../util/assert';\nimport { GarbageCollectionScheduler, Persistence } from '../local/persistence';\nimport { Code, FirestoreError } from '../util/error';\nimport { OnlineStateSource } from './types';\nimport { LruParams, LruScheduler } from '../local/lru_garbage_collector';\nimport { IndexFreeQueryEngine } from '../local/index_free_query_engine';\nimport { IndexedDbPersistence } from '../local/indexeddb_persistence';\nimport {\n MemoryEagerDelegate,\n MemoryPersistence\n} from '../local/memory_persistence';\n\nconst MEMORY_ONLY_PERSISTENCE_ERROR_MESSAGE =\n 'You are using the memory-only build of Firestore. Persistence support is ' +\n 'only available via the @firebase/firestore bundle or the ' +\n 'firebase-firestore.js build.';\n\nexport interface ComponentConfiguration {\n asyncQueue: AsyncQueue;\n databaseInfo: DatabaseInfo;\n platform: Platform;\n datastore: Datastore;\n clientId: ClientId;\n initialUser: User;\n maxConcurrentLimboResolutions: number;\n persistenceSettings: PersistenceSettings;\n}\n\n/**\n * Initializes and wires up all core components for Firestore. Implementations\n * override `initialize()` to provide all components.\n */\nexport interface ComponentProvider {\n persistence: Persistence;\n sharedClientState: SharedClientState;\n localStore: LocalStore;\n syncEngine: SyncEngine;\n gcScheduler: GarbageCollectionScheduler | null;\n remoteStore: RemoteStore;\n eventManager: EventManager;\n\n initialize(cfg: ComponentConfiguration): Promise<void>;\n\n clearPersistence(databaseId: DatabaseInfo): Promise<void>;\n}\n\n/**\n * Provides all components needed for Firestore with in-memory persistence.\n * Uses EagerGC garbage collection.\n */\nexport class MemoryComponentProvider implements ComponentProvider {\n persistence!: Persistence;\n sharedClientState!: SharedClientState;\n localStore!: LocalStore;\n syncEngine!: SyncEngine;\n gcScheduler!: GarbageCollectionScheduler | null;\n remoteStore!: RemoteStore;\n eventManager!: EventManager;\n\n async initialize(cfg: ComponentConfiguration): Promise<void> {\n this.sharedClientState = this.createSharedClientState(cfg);\n this.persistence = this.createPersistence(cfg);\n await this.persistence.start();\n this.gcScheduler = this.createGarbageCollectionScheduler(cfg);\n this.localStore = this.createLocalStore(cfg);\n this.remoteStore = this.createRemoteStore(cfg);\n this.syncEngine = this.createSyncEngine(cfg);\n this.eventManager = this.createEventManager(cfg);\n\n this.sharedClientState.onlineStateHandler = onlineState =>\n this.syncEngine.applyOnlineStateChange(\n onlineState,\n OnlineStateSource.SharedClientState\n );\n this.remoteStore.syncEngine = this.syncEngine;\n\n await this.localStore.start();\n await this.sharedClientState.start();\n await this.remoteStore.start();\n\n await this.remoteStore.applyPrimaryState(this.syncEngine.isPrimaryClient);\n }\n\n createEventManager(cfg: ComponentConfiguration): EventManager {\n return new EventManager(this.syncEngine);\n }\n\n createGarbageCollectionScheduler(\n cfg: ComponentConfiguration\n ): GarbageCollectionScheduler | null {\n return null;\n }\n\n createLocalStore(cfg: ComponentConfiguration): LocalStore {\n return new LocalStore(\n this.persistence,\n new IndexFreeQueryEngine(),\n cfg.initialUser\n );\n }\n\n createPersistence(cfg: ComponentConfiguration): Persistence {\n debugAssert(\n !cfg.persistenceSettings.durable,\n 'Can only start memory persistence'\n );\n return new MemoryPersistence(MemoryEagerDelegate.factory);\n }\n\n createRemoteStore(cfg: ComponentConfiguration): RemoteStore {\n return new RemoteStore(\n this.localStore,\n cfg.datastore,\n cfg.asyncQueue,\n onlineState =>\n this.syncEngine.applyOnlineStateChange(\n onlineState,\n OnlineStateSource.RemoteStore\n ),\n cfg.platform.newConnectivityMonitor()\n );\n }\n\n createSharedClientState(cfg: ComponentConfiguration): SharedClientState {\n return new MemorySharedClientState();\n }\n\n createSyncEngine(cfg: ComponentConfiguration): SyncEngine {\n return new SyncEngine(\n this.localStore,\n this.remoteStore,\n this.sharedClientState,\n cfg.initialUser,\n cfg.maxConcurrentLimboResolutions\n );\n }\n\n clearPersistence(databaseInfo: DatabaseInfo): Promise<void> {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n MEMORY_ONLY_PERSISTENCE_ERROR_MESSAGE\n );\n }\n}\n\n/**\n * Provides all components needed for Firestore with IndexedDB persistence.\n */\nexport class IndexedDbComponentProvider extends MemoryComponentProvider {\n persistence!: IndexedDbPersistence;\n\n // TODO(tree-shaking): Create an IndexedDbComponentProvider and a\n // MultiTabComponentProvider. The IndexedDbComponentProvider should depend\n // on LocalStore and SyncEngine.\n localStore!: MultiTabLocalStore;\n syncEngine!: MultiTabSyncEngine;\n\n async initialize(cfg: ComponentConfiguration): Promise<void> {\n await super.initialize(cfg);\n\n // NOTE: This will immediately call the listener, so we make sure to\n // set it after localStore / remoteStore are started.\n await this.persistence.setPrimaryStateListener(async isPrimary => {\n await (this.syncEngine as MultiTabSyncEngine).applyPrimaryState(\n isPrimary\n );\n if (this.gcScheduler) {\n if (isPrimary && !this.gcScheduler.started) {\n this.gcScheduler.start(this.localStore);\n } else if (!isPrimary) {\n this.gcScheduler.stop();\n }\n }\n });\n }\n\n createLocalStore(cfg: ComponentConfiguration): LocalStore {\n return new MultiTabLocalStore(\n this.persistence,\n new IndexFreeQueryEngine(),\n cfg.initialUser\n );\n }\n\n createSyncEngine(cfg: ComponentConfiguration): SyncEngine {\n const syncEngine = new MultiTabSyncEngine(\n this.localStore,\n this.remoteStore,\n this.sharedClientState,\n cfg.initialUser,\n cfg.maxConcurrentLimboResolutions\n );\n if (this.sharedClientState instanceof WebStorageSharedClientState) {\n this.sharedClientState.syncEngine = syncEngine;\n }\n return syncEngine;\n }\n\n createGarbageCollectionScheduler(\n cfg: ComponentConfiguration\n ): GarbageCollectionScheduler | null {\n const garbageCollector = this.persistence.referenceDelegate\n .garbageCollector;\n return new LruScheduler(garbageCollector, cfg.asyncQueue);\n }\n\n createPersistence(cfg: ComponentConfiguration): Persistence {\n debugAssert(\n cfg.persistenceSettings.durable,\n 'Can only start durable persistence'\n );\n\n const persistenceKey = IndexedDbPersistence.buildStoragePrefix(\n cfg.databaseInfo\n );\n const serializer = cfg.platform.newSerializer(cfg.databaseInfo.databaseId);\n return new IndexedDbPersistence(\n cfg.persistenceSettings.synchronizeTabs,\n persistenceKey,\n cfg.clientId,\n cfg.platform,\n LruParams.withCacheSize(cfg.persistenceSettings.cacheSizeBytes),\n cfg.asyncQueue,\n serializer,\n this.sharedClientState\n );\n }\n\n createSharedClientState(cfg: ComponentConfiguration): SharedClientState {\n if (\n cfg.persistenceSettings.durable &&\n cfg.persistenceSettings.synchronizeTabs\n ) {\n if (!WebStorageSharedClientState.isAvailable(cfg.platform)) {\n throw new FirestoreError(\n Code.UNIMPLEMENTED,\n 'IndexedDB persistence is only available on platforms that support LocalStorage.'\n );\n }\n const persistenceKey = IndexedDbPersistence.buildStoragePrefix(\n cfg.databaseInfo\n );\n return new WebStorageSharedClientState(\n cfg.asyncQueue,\n cfg.platform,\n persistenceKey,\n cfg.clientId,\n cfg.initialUser\n );\n }\n return new MemorySharedClientState();\n }\n\n clearPersistence(databaseInfo: DatabaseInfo): Promise<void> {\n const persistenceKey = IndexedDbPersistence.buildStoragePrefix(\n databaseInfo\n );\n return IndexedDbPersistence.clearPersistence(persistenceKey);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { CredentialsProvider } from '../api/credentials';\nimport { User } from '../auth/user';\nimport { LocalStore } from '../local/local_store';\nimport { GarbageCollectionScheduler, Persistence } from '../local/persistence';\nimport { Document, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { Mutation } from '../model/mutation';\nimport { Platform } from '../platform/platform';\nimport { newDatastore } from '../remote/datastore';\nimport { RemoteStore } from '../remote/remote_store';\nimport { AsyncQueue, wrapInUserErrorIfRecoverable } from '../util/async_queue';\nimport { Code, FirestoreError } from '../util/error';\nimport { logDebug } from '../util/log';\nimport { Deferred } from '../util/promise';\nimport {\n EventManager,\n ListenOptions,\n Observer,\n QueryListener\n} from './event_manager';\nimport { SyncEngine } from './sync_engine';\nimport { View } from './view';\n\nimport { SharedClientState } from '../local/shared_client_state';\nimport { AutoId } from '../util/misc';\nimport { DatabaseId, DatabaseInfo } from './database_info';\nimport { Query } from './query';\nimport { Transaction } from './transaction';\nimport { ViewSnapshot } from './view_snapshot';\nimport {\n ComponentProvider,\n MemoryComponentProvider\n} from './component_provider';\n\nconst LOG_TAG = 'FirestoreClient';\nconst MAX_CONCURRENT_LIMBO_RESOLUTIONS = 100;\n\n/** DOMException error code constants. */\nconst DOM_EXCEPTION_INVALID_STATE = 11;\nconst DOM_EXCEPTION_ABORTED = 20;\nconst DOM_EXCEPTION_QUOTA_EXCEEDED = 22;\n\nexport type PersistenceSettings =\n | {\n readonly durable: false;\n }\n | {\n readonly durable: true;\n readonly cacheSizeBytes: number;\n readonly synchronizeTabs: boolean;\n };\n\n/**\n * FirestoreClient is a top-level class that constructs and owns all of the\n * pieces of the client SDK architecture. It is responsible for creating the\n * async queue that is shared by all of the other components in the system.\n */\nexport class FirestoreClient {\n // NOTE: These should technically have '|undefined' in the types, since\n // they're initialized asynchronously rather than in the constructor, but\n // given that all work is done on the async queue and we assert that\n // initialization completes before any other work is queued, we're cheating\n // with the types rather than littering the code with '!' or unnecessary\n // undefined checks.\n private eventMgr!: EventManager;\n private persistence!: Persistence;\n private localStore!: LocalStore;\n private remoteStore!: RemoteStore;\n private syncEngine!: SyncEngine;\n private gcScheduler!: GarbageCollectionScheduler | null;\n\n // PORTING NOTE: SharedClientState is only used for multi-tab web.\n private sharedClientState!: SharedClientState;\n\n private readonly clientId = AutoId.newId();\n\n constructor(\n private platform: Platform,\n private databaseInfo: DatabaseInfo,\n private credentials: CredentialsProvider,\n /**\n * Asynchronous queue responsible for all of our internal processing. When\n * we get incoming work from the user (via public API) or the network\n * (incoming GRPC messages), we should always schedule onto this queue.\n * This ensures all of our work is properly serialized (e.g. we don't\n * start processing a new operation while the previous one is waiting for\n * an async I/O to complete).\n */\n private asyncQueue: AsyncQueue\n ) {}\n\n /**\n * Starts up the FirestoreClient, returning only whether or not enabling\n * persistence succeeded.\n *\n * The intent here is to \"do the right thing\" as far as users are concerned.\n * Namely, in cases where offline persistence is requested and possible,\n * enable it, but otherwise fall back to persistence disabled. For the most\n * part we expect this to succeed one way or the other so we don't expect our\n * users to actually wait on the firestore.enablePersistence Promise since\n * they generally won't care.\n *\n * Of course some users actually do care about whether or not persistence\n * was successfully enabled, so the Promise returned from this method\n * indicates this outcome.\n *\n * This presents a problem though: even before enablePersistence resolves or\n * rejects, users may have made calls to e.g. firestore.collection() which\n * means that the FirestoreClient in there will be available and will be\n * enqueuing actions on the async queue.\n *\n * Meanwhile any failure of an operation on the async queue causes it to\n * panic and reject any further work, on the premise that unhandled errors\n * are fatal.\n *\n * Consequently the fallback is handled internally here in start, and if the\n * fallback succeeds we signal success to the async queue even though the\n * start() itself signals failure.\n *\n * @param componentProvider Provider that returns all core components.\n * @param persistenceSettings Settings object to configure offline\n * persistence.\n * @returns A deferred result indicating the user-visible result of enabling\n * offline persistence. This method will reject this if IndexedDB fails to\n * start for any reason. If usePersistence is false this is\n * unconditionally resolved.\n */\n start(\n componentProvider: ComponentProvider,\n persistenceSettings: PersistenceSettings\n ): Promise<void> {\n this.verifyNotTerminated();\n // We defer our initialization until we get the current user from\n // setChangeListener(). We block the async queue until we got the initial\n // user and the initialization is completed. This will prevent any scheduled\n // work from happening before initialization is completed.\n //\n // If initializationDone resolved then the FirestoreClient is in a usable\n // state.\n const initializationDone = new Deferred<void>();\n\n // If usePersistence is true, certain classes of errors while starting are\n // recoverable but only by falling back to persistence disabled.\n //\n // If there's an error in the first case but not in recovery we cannot\n // reject the promise blocking the async queue because this will cause the\n // async queue to panic.\n const persistenceResult = new Deferred<void>();\n\n let initialized = false;\n this.credentials.setChangeListener(user => {\n if (!initialized) {\n initialized = true;\n\n logDebug(LOG_TAG, 'Initializing. user=', user.uid);\n\n return this.initializeComponents(\n componentProvider,\n persistenceSettings,\n user,\n persistenceResult\n ).then(initializationDone.resolve, initializationDone.reject);\n } else {\n this.asyncQueue.enqueueRetryable(() => {\n return this.handleCredentialChange(user);\n });\n }\n });\n\n // Block the async queue until initialization is done\n this.asyncQueue.enqueueAndForget(() => {\n return initializationDone.promise;\n });\n\n // Return only the result of enabling persistence. Note that this does not\n // need to await the completion of initializationDone because the result of\n // this method should not reflect any other kind of failure to start.\n return persistenceResult.promise;\n }\n\n /** Enables the network connection and requeues all pending operations. */\n enableNetwork(): Promise<void> {\n this.verifyNotTerminated();\n return this.asyncQueue.enqueue(() => {\n return this.syncEngine.enableNetwork();\n });\n }\n\n /**\n * Initializes persistent storage, attempting to use IndexedDB if\n * usePersistence is true or memory-only if false.\n *\n * If IndexedDB fails because it's already open in another tab or because the\n * platform can't possibly support our implementation then this method rejects\n * the persistenceResult and falls back on memory-only persistence.\n *\n * @param componentProvider The provider that provides all core componennts\n * for IndexedDB or memory-backed persistence\n * @param persistenceSettings Settings object to configure offline persistence\n * @param user The initial user\n * @param persistenceResult A deferred result indicating the user-visible\n * result of enabling offline persistence. This method will reject this if\n * IndexedDB fails to start for any reason. If usePersistence is false\n * this is unconditionally resolved.\n * @returns a Promise indicating whether or not initialization should\n * continue, i.e. that one of the persistence implementations actually\n * succeeded.\n */\n private async initializeComponents(\n componentProvider: ComponentProvider,\n persistenceSettings: PersistenceSettings,\n user: User,\n persistenceResult: Deferred<void>\n ): Promise<void> {\n try {\n // TODO(mrschmidt): Ideally, ComponentProvider would also initialize\n // Datastore (without duplicating the initializing logic once per\n // provider).\n\n const connection = await this.platform.loadConnection(this.databaseInfo);\n const serializer = this.platform.newSerializer(\n this.databaseInfo.databaseId\n );\n const datastore = newDatastore(connection, this.credentials, serializer);\n\n await componentProvider.initialize({\n asyncQueue: this.asyncQueue,\n databaseInfo: this.databaseInfo,\n platform: this.platform,\n datastore,\n clientId: this.clientId,\n initialUser: user,\n maxConcurrentLimboResolutions: MAX_CONCURRENT_LIMBO_RESOLUTIONS,\n persistenceSettings\n });\n\n this.persistence = componentProvider.persistence;\n this.sharedClientState = componentProvider.sharedClientState;\n this.localStore = componentProvider.localStore;\n this.remoteStore = componentProvider.remoteStore;\n this.syncEngine = componentProvider.syncEngine;\n this.gcScheduler = componentProvider.gcScheduler;\n this.eventMgr = componentProvider.eventManager;\n\n // When a user calls clearPersistence() in one client, all other clients\n // need to be terminated to allow the delete to succeed.\n this.persistence.setDatabaseDeletedListener(async () => {\n await this.terminate();\n });\n\n persistenceResult.resolve();\n } catch (error) {\n // Regardless of whether or not the retry succeeds, from an user\n // perspective, offline persistence has failed.\n persistenceResult.reject(error);\n\n // An unknown failure on the first stage shuts everything down.\n if (!this.canFallback(error)) {\n throw error;\n }\n console.warn(\n 'Error enabling offline persistence. Falling back to' +\n ' persistence disabled: ' +\n error\n );\n return this.initializeComponents(\n new MemoryComponentProvider(),\n { durable: false },\n user,\n persistenceResult\n );\n }\n }\n\n /**\n * Decides whether the provided error allows us to gracefully disable\n * persistence (as opposed to crashing the client).\n */\n private canFallback(error: FirestoreError | DOMException): boolean {\n if (error.name === 'FirebaseError') {\n return (\n error.code === Code.FAILED_PRECONDITION ||\n error.code === Code.UNIMPLEMENTED\n );\n } else if (\n typeof DOMException !== 'undefined' &&\n error instanceof DOMException\n ) {\n // There are a few known circumstances where we can open IndexedDb but\n // trying to read/write will fail (e.g. quota exceeded). For\n // well-understood cases, we attempt to detect these and then gracefully\n // fall back to memory persistence.\n // NOTE: Rather than continue to add to this list, we could decide to\n // always fall back, with the risk that we might accidentally hide errors\n // representing actual SDK bugs.\n return (\n // When the browser is out of quota we could get either quota exceeded\n // or an aborted error depending on whether the error happened during\n // schema migration.\n error.code === DOM_EXCEPTION_QUOTA_EXCEEDED ||\n error.code === DOM_EXCEPTION_ABORTED ||\n // Firefox Private Browsing mode disables IndexedDb and returns\n // INVALID_STATE for any usage.\n error.code === DOM_EXCEPTION_INVALID_STATE\n );\n }\n\n return true;\n }\n\n /**\n * Checks that the client has not been terminated. Ensures that other methods on\n * this class cannot be called after the client is terminated.\n */\n private verifyNotTerminated(): void {\n if (this.asyncQueue.isShuttingDown) {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'The client has already been terminated.'\n );\n }\n }\n\n private handleCredentialChange(user: User): Promise<void> {\n this.asyncQueue.verifyOperationInProgress();\n\n logDebug(LOG_TAG, 'Credential Changed. Current user: ' + user.uid);\n return this.syncEngine.handleCredentialChange(user);\n }\n\n /** Disables the network connection. Pending operations will not complete. */\n disableNetwork(): Promise<void> {\n this.verifyNotTerminated();\n return this.asyncQueue.enqueue(() => {\n return this.syncEngine.disableNetwork();\n });\n }\n\n terminate(): Promise<void> {\n return this.asyncQueue.enqueueAndInitiateShutdown(async () => {\n // PORTING NOTE: LocalStore does not need an explicit shutdown on web.\n if (this.gcScheduler) {\n this.gcScheduler.stop();\n }\n\n await this.remoteStore.shutdown();\n await this.sharedClientState.shutdown();\n await this.persistence.shutdown();\n\n // `removeChangeListener` must be called after shutting down the\n // RemoteStore as it will prevent the RemoteStore from retrieving\n // auth tokens.\n this.credentials.removeChangeListener();\n });\n }\n\n /**\n * Returns a Promise that resolves when all writes that were pending at the time this\n * method was called received server acknowledgement. An acknowledgement can be either acceptance\n * or rejection.\n */\n waitForPendingWrites(): Promise<void> {\n this.verifyNotTerminated();\n\n const deferred = new Deferred<void>();\n this.asyncQueue.enqueueAndForget(() => {\n return this.syncEngine.registerPendingWritesCallback(deferred);\n });\n return deferred.promise;\n }\n\n listen(\n query: Query,\n observer: Observer<ViewSnapshot>,\n options: ListenOptions\n ): QueryListener {\n this.verifyNotTerminated();\n const listener = new QueryListener(query, observer, options);\n this.asyncQueue.enqueueAndForget(() => this.eventMgr.listen(listener));\n return listener;\n }\n\n unlisten(listener: QueryListener): void {\n // Checks for termination but does not raise error, allowing unlisten after\n // termination to be a no-op.\n if (this.clientTerminated) {\n return;\n }\n this.asyncQueue.enqueueAndForget(() => {\n return this.eventMgr.unlisten(listener);\n });\n }\n\n async getDocumentFromLocalCache(\n docKey: DocumentKey\n ): Promise<Document | null> {\n this.verifyNotTerminated();\n const deferred = new Deferred<Document | null>();\n await this.asyncQueue.enqueue(async () => {\n try {\n const maybeDoc = await this.localStore.readDocument(docKey);\n if (maybeDoc instanceof Document) {\n deferred.resolve(maybeDoc);\n } else if (maybeDoc instanceof NoDocument) {\n deferred.resolve(null);\n } else {\n deferred.reject(\n new FirestoreError(\n Code.UNAVAILABLE,\n 'Failed to get document from cache. (However, this document may ' +\n \"exist on the server. Run again without setting 'source' in \" +\n 'the GetOptions to attempt to retrieve the document from the ' +\n 'server.)'\n )\n );\n }\n } catch (e) {\n const firestoreError = wrapInUserErrorIfRecoverable(\n e,\n `Failed to get document '${docKey} from cache`\n );\n deferred.reject(firestoreError);\n }\n });\n\n return deferred.promise;\n }\n\n async getDocumentsFromLocalCache(query: Query): Promise<ViewSnapshot> {\n this.verifyNotTerminated();\n const deferred = new Deferred<ViewSnapshot>();\n await this.asyncQueue.enqueue(async () => {\n try {\n const queryResult = await this.localStore.executeQuery(\n query,\n /* usePreviousResults= */ true\n );\n const view = new View(query, queryResult.remoteKeys);\n const viewDocChanges = view.computeDocChanges(queryResult.documents);\n const viewChange = view.applyChanges(\n viewDocChanges,\n /* updateLimboDocuments= */ false\n );\n deferred.resolve(viewChange.snapshot!);\n } catch (e) {\n const firestoreError = wrapInUserErrorIfRecoverable(\n e,\n `Failed to execute query '${query} against cache`\n );\n deferred.reject(firestoreError);\n }\n });\n return deferred.promise;\n }\n\n write(mutations: Mutation[]): Promise<void> {\n this.verifyNotTerminated();\n const deferred = new Deferred<void>();\n this.asyncQueue.enqueueAndForget(() =>\n this.syncEngine.write(mutations, deferred)\n );\n return deferred.promise;\n }\n\n databaseId(): DatabaseId {\n return this.databaseInfo.databaseId;\n }\n\n addSnapshotsInSyncListener(observer: Observer<void>): void {\n this.verifyNotTerminated();\n this.asyncQueue.enqueueAndForget(() => {\n this.eventMgr.addSnapshotsInSyncListener(observer);\n return Promise.resolve();\n });\n }\n\n removeSnapshotsInSyncListener(observer: Observer<void>): void {\n // Checks for shutdown but does not raise error, allowing remove after\n // shutdown to be a no-op.\n if (this.clientTerminated) {\n return;\n }\n this.asyncQueue.enqueueAndForget(() => {\n this.eventMgr.removeSnapshotsInSyncListener(observer);\n return Promise.resolve();\n });\n }\n\n get clientTerminated(): boolean {\n // Technically, the asyncQueue is still running, but only accepting operations\n // related to termination or supposed to be run after termination. It is effectively\n // terminated to the eyes of users.\n return this.asyncQueue.isShuttingDown;\n }\n\n transaction<T>(\n updateFunction: (transaction: Transaction) => Promise<T>\n ): Promise<T> {\n this.verifyNotTerminated();\n const deferred = new Deferred<T>();\n this.asyncQueue.enqueueAndForget(() => {\n this.syncEngine.runTransaction(this.asyncQueue, updateFunction, deferred);\n return Promise.resolve();\n });\n return deferred.promise;\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Observer } from '../core/event_manager';\nimport { EventHandler } from './misc';\n\n/*\n * A wrapper implementation of Observer<T> that will dispatch events\n * asynchronously. To allow immediate silencing, a mute call is added which\n * causes events scheduled to no longer be raised.\n */\nexport class AsyncObserver<T> implements Observer<T> {\n /**\n * When set to true, will not raise future events. Necessary to deal with\n * async detachment of listener.\n */\n private muted = false;\n\n constructor(private observer: Observer<T>) {}\n\n next(value: T): void {\n this.scheduleEvent(this.observer.next, value);\n }\n\n error(error: Error): void {\n this.scheduleEvent(this.observer.error, error);\n }\n\n mute(): void {\n this.muted = true;\n }\n\n private scheduleEvent<E>(eventHandler: EventHandler<E>, event: E): void {\n if (!this.muted) {\n setTimeout(() => {\n if (!this.muted) {\n eventHandler(event);\n }\n }, 0);\n }\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { JsonObject } from '../model/object_value';\n\n/**\n * Observer/Subscribe interfaces.\n */\nexport type NextFn<T> = (value: T) => void;\nexport type ErrorFn = (error: Error) => void;\nexport type CompleteFn = () => void;\n\n// Allow for any of the Observer methods to be undefined.\nexport interface PartialObserver<T> {\n next?: NextFn<T>;\n error?: ErrorFn;\n complete?: CompleteFn;\n}\n\nexport interface Unsubscribe {\n (): void;\n}\n\nexport function isPartialObserver(obj: unknown): boolean {\n return implementsAnyMethods(obj, ['next', 'error', 'complete']);\n}\n\n/**\n * Returns true if obj is an object and contains at least one of the specified\n * methods.\n */\nfunction implementsAnyMethods(obj: unknown, methods: string[]): boolean {\n if (typeof obj !== 'object' || obj === null) {\n return false;\n }\n\n const object = obj as JsonObject<unknown>;\n for (const method of methods) {\n if (method in object && typeof object[method] === 'function') {\n return true;\n }\n }\n return false;\n}\n","/**\n * @license\n * Copyright 2020 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as firestore from '@firebase/firestore-types';\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { DocumentReference, Firestore } from './database';\nimport { Blob } from './blob';\nimport { GeoPoint } from './geo_point';\nimport { Timestamp } from './timestamp';\nimport { DatabaseId } from '../core/database_info';\nimport { DocumentKey } from '../model/document_key';\nimport {\n normalizeByteString,\n normalizeNumber,\n normalizeTimestamp,\n typeOrder\n} from '../model/values';\nimport {\n getLocalWriteTime,\n getPreviousValue\n} from '../model/server_timestamps';\nimport { fail, hardAssert } from '../util/assert';\nimport { forEach } from '../util/obj';\nimport { TypeOrder } from '../model/object_value';\nimport { ResourcePath } from '../model/path';\nimport { isValidResourceName } from '../remote/serializer';\nimport { logError } from '../util/log';\n\nexport type ServerTimestampBehavior = 'estimate' | 'previous' | 'none';\n\n/**\n * Converts Firestore's internal types to the JavaScript types that we expose\n * to the user.\n */\nexport class UserDataWriter<T = firestore.DocumentData> {\n constructor(\n private readonly firestore: Firestore,\n private readonly timestampsInSnapshots: boolean,\n private readonly serverTimestampBehavior?: ServerTimestampBehavior,\n private readonly converter?: firestore.FirestoreDataConverter<T>\n ) {}\n\n convertValue(value: api.Value): unknown {\n switch (typeOrder(value)) {\n case TypeOrder.NullValue:\n return null;\n case TypeOrder.BooleanValue:\n return value.booleanValue!;\n case TypeOrder.NumberValue:\n return normalizeNumber(value.integerValue || value.doubleValue);\n case TypeOrder.TimestampValue:\n return this.convertTimestamp(value.timestampValue!);\n case TypeOrder.ServerTimestampValue:\n return this.convertServerTimestamp(value);\n case TypeOrder.StringValue:\n return value.stringValue!;\n case TypeOrder.BlobValue:\n return new Blob(normalizeByteString(value.bytesValue!));\n case TypeOrder.RefValue:\n return this.convertReference(value.referenceValue!);\n case TypeOrder.GeoPointValue:\n return this.convertGeoPoint(value.geoPointValue!);\n case TypeOrder.ArrayValue:\n return this.convertArray(value.arrayValue!);\n case TypeOrder.ObjectValue:\n return this.convertObject(value.mapValue!);\n default:\n throw fail('Invalid value type: ' + JSON.stringify(value));\n }\n }\n\n private convertObject(mapValue: api.MapValue): firestore.DocumentData {\n const result: firestore.DocumentData = {};\n forEach(mapValue.fields || {}, (key, value) => {\n result[key] = this.convertValue(value);\n });\n return result;\n }\n\n private convertGeoPoint(value: api.LatLng): GeoPoint {\n return new GeoPoint(\n normalizeNumber(value.latitude),\n normalizeNumber(value.longitude)\n );\n }\n\n private convertArray(arrayValue: api.ArrayValue): unknown[] {\n return (arrayValue.values || []).map(value => this.convertValue(value));\n }\n\n private convertServerTimestamp(value: api.Value): unknown {\n switch (this.serverTimestampBehavior) {\n case 'previous':\n const previousValue = getPreviousValue(value);\n if (previousValue == null) {\n return null;\n }\n return this.convertValue(previousValue);\n case 'estimate':\n return this.convertTimestamp(getLocalWriteTime(value));\n default:\n return null;\n }\n }\n\n private convertTimestamp(value: api.Timestamp): Timestamp | Date {\n const normalizedValue = normalizeTimestamp(value);\n const timestamp = new Timestamp(\n normalizedValue.seconds,\n normalizedValue.nanos\n );\n if (this.timestampsInSnapshots) {\n return timestamp;\n } else {\n return timestamp.toDate();\n }\n }\n\n private convertReference(name: string): DocumentReference<T> {\n const resourcePath = ResourcePath.fromString(name);\n hardAssert(\n isValidResourceName(resourcePath),\n 'ReferenceValue is not valid ' + name\n );\n const databaseId = new DatabaseId(resourcePath.get(1), resourcePath.get(3));\n const key = new DocumentKey(resourcePath.popFirst(5));\n\n if (!databaseId.isEqual(this.firestore._databaseId)) {\n // TODO(b/64130202): Somehow support foreign references.\n logError(\n `Document ${key} contains a document ` +\n `reference within a different database (` +\n `${databaseId.projectId}/${databaseId.database}) which is not ` +\n `supported. It will be treated as a reference in the current ` +\n `database (${this.firestore._databaseId.projectId}/${this.firestore._databaseId.database}) ` +\n `instead.`\n );\n }\n\n return new DocumentReference(key, this.firestore, this.converter);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport * as firestore from '@firebase/firestore-types';\n\nimport * as api from '../protos/firestore_proto_api';\n\nimport { FirebaseApp } from '@firebase/app-types';\nimport { _FirebaseApp, FirebaseService } from '@firebase/app-types/private';\nimport { DatabaseId, DatabaseInfo } from '../core/database_info';\nimport { ListenOptions } from '../core/event_manager';\nimport {\n ComponentProvider,\n MemoryComponentProvider\n} from '../core/component_provider';\nimport { FirestoreClient, PersistenceSettings } from '../core/firestore_client';\nimport {\n Bound,\n Direction,\n FieldFilter,\n Filter,\n Operator,\n OrderBy,\n Query as InternalQuery\n} from '../core/query';\nimport { Transaction as InternalTransaction } from '../core/transaction';\nimport { ChangeType, ViewSnapshot } from '../core/view_snapshot';\nimport { LruParams } from '../local/lru_garbage_collector';\nimport { Document, MaybeDocument, NoDocument } from '../model/document';\nimport { DocumentKey } from '../model/document_key';\nimport { DeleteMutation, Mutation, Precondition } from '../model/mutation';\nimport { FieldPath, ResourcePath } from '../model/path';\nimport { isServerTimestamp } from '../model/server_timestamps';\nimport { refValue } from '../model/values';\nimport { PlatformSupport } from '../platform/platform';\nimport { debugAssert, fail } from '../util/assert';\nimport { AsyncObserver } from '../util/async_observer';\nimport { AsyncQueue } from '../util/async_queue';\nimport { Code, FirestoreError } from '../util/error';\nimport {\n invalidClassError,\n validateArgType,\n validateAtLeastNumberOfArgs,\n validateBetweenNumberOfArgs,\n validateDefined,\n validateExactNumberOfArgs,\n validateNamedOptionalPropertyEquals,\n validateNamedOptionalType,\n validateNamedType,\n validateOptionalArgType,\n validateOptionalArrayElements,\n validateOptionNames,\n validatePositiveNumber,\n validateStringEnum,\n valueDescription\n} from '../util/input_validation';\nimport { getLogLevel, logError, LogLevel, setLogLevel } from '../util/log';\nimport { AutoId } from '../util/misc';\nimport { Deferred, Rejecter, Resolver } from '../util/promise';\nimport { FieldPath as ExternalFieldPath } from './field_path';\n\nimport {\n CredentialsProvider,\n CredentialsSettings,\n EmptyCredentialsProvider,\n FirebaseCredentialsProvider,\n makeCredentialsProvider\n} from './credentials';\nimport {\n CompleteFn,\n ErrorFn,\n isPartialObserver,\n NextFn,\n PartialObserver,\n Unsubscribe\n} from './observer';\nimport { fieldPathFromArgument, UserDataReader } from './user_data_reader';\nimport { UserDataWriter } from './user_data_writer';\nimport { FirebaseAuthInternalName } from '@firebase/auth-interop-types';\nimport { Provider } from '@firebase/component';\n\n// settings() defaults:\nconst DEFAULT_HOST = 'firestore.googleapis.com';\nconst DEFAULT_SSL = true;\nconst DEFAULT_TIMESTAMPS_IN_SNAPSHOTS = true;\nconst DEFAULT_FORCE_LONG_POLLING = false;\nconst DEFAULT_IGNORE_UNDEFINED_PROPERTIES = false;\n\n/**\n * Constant used to indicate the LRU garbage collection should be disabled.\n * Set this value as the `cacheSizeBytes` on the settings passed to the\n * `Firestore` instance.\n */\nexport const CACHE_SIZE_UNLIMITED = LruParams.COLLECTION_DISABLED;\n\n// enablePersistence() defaults:\nconst DEFAULT_SYNCHRONIZE_TABS = false;\n\n/** Undocumented, private additional settings not exposed in our public API. */\ninterface PrivateSettings extends firestore.Settings {\n // Can be a google-auth-library or gapi client.\n credentials?: CredentialsSettings;\n}\n\n/**\n * Options that can be provided in the Firestore constructor when not using\n * Firebase (aka standalone mode).\n */\nexport interface FirestoreDatabase {\n projectId: string;\n database?: string;\n}\n\n/**\n * A concrete type describing all the values that can be applied via a\n * user-supplied firestore.Settings object. This is a separate type so that\n * defaults can be supplied and the value can be checked for equality.\n */\nclass FirestoreSettings {\n /** The hostname to connect to. */\n readonly host: string;\n\n /** Whether to use SSL when connecting. */\n readonly ssl: boolean;\n\n readonly timestampsInSnapshots: boolean;\n\n readonly cacheSizeBytes: number;\n\n readonly forceLongPolling: boolean;\n\n readonly ignoreUndefinedProperties: boolean;\n\n // Can be a google-auth-library or gapi client.\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n credentials?: any;\n\n constructor(settings: PrivateSettings) {\n if (settings.host === undefined) {\n if (settings.ssl !== undefined) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n \"Can't provide ssl option if host option is not set\"\n );\n }\n this.host = DEFAULT_HOST;\n this.ssl = DEFAULT_SSL;\n } else {\n validateNamedType('settings', 'non-empty string', 'host', settings.host);\n this.host = settings.host;\n\n validateNamedOptionalType('settings', 'boolean', 'ssl', settings.ssl);\n this.ssl = settings.ssl ?? DEFAULT_SSL;\n }\n validateOptionNames('settings', settings, [\n 'host',\n 'ssl',\n 'credentials',\n 'timestampsInSnapshots',\n 'cacheSizeBytes',\n 'experimentalForceLongPolling',\n 'ignoreUndefinedProperties'\n ]);\n\n validateNamedOptionalType(\n 'settings',\n 'object',\n 'credentials',\n settings.credentials\n );\n this.credentials = settings.credentials;\n\n validateNamedOptionalType(\n 'settings',\n 'boolean',\n 'timestampsInSnapshots',\n settings.timestampsInSnapshots\n );\n\n validateNamedOptionalType(\n 'settings',\n 'boolean',\n 'ignoreUndefinedProperties',\n settings.ignoreUndefinedProperties\n );\n\n // Nobody should set timestampsInSnapshots anymore, but the error depends on\n // whether they set it to true or false...\n if (settings.timestampsInSnapshots === true) {\n logError(\n \"The setting 'timestampsInSnapshots: true' is no longer required \" +\n 'and should be removed.'\n );\n } else if (settings.timestampsInSnapshots === false) {\n logError(\n \"Support for 'timestampsInSnapshots: false' will be removed soon. \" +\n 'You must update your code to handle Timestamp objects.'\n );\n }\n this.timestampsInSnapshots =\n settings.timestampsInSnapshots ?? DEFAULT_TIMESTAMPS_IN_SNAPSHOTS;\n this.ignoreUndefinedProperties =\n settings.ignoreUndefinedProperties ?? DEFAULT_IGNORE_UNDEFINED_PROPERTIES;\n\n validateNamedOptionalType(\n 'settings',\n 'number',\n 'cacheSizeBytes',\n settings.cacheSizeBytes\n );\n if (settings.cacheSizeBytes === undefined) {\n this.cacheSizeBytes = LruParams.DEFAULT_CACHE_SIZE_BYTES;\n } else {\n if (\n settings.cacheSizeBytes !== CACHE_SIZE_UNLIMITED &&\n settings.cacheSizeBytes < LruParams.MINIMUM_CACHE_SIZE_BYTES\n ) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `cacheSizeBytes must be at least ${LruParams.MINIMUM_CACHE_SIZE_BYTES}`\n );\n } else {\n this.cacheSizeBytes = settings.cacheSizeBytes;\n }\n }\n\n validateNamedOptionalType(\n 'settings',\n 'boolean',\n 'experimentalForceLongPolling',\n settings.experimentalForceLongPolling\n );\n this.forceLongPolling =\n settings.experimentalForceLongPolling ?? DEFAULT_FORCE_LONG_POLLING;\n }\n\n isEqual(other: FirestoreSettings): boolean {\n return (\n this.host === other.host &&\n this.ssl === other.ssl &&\n this.timestampsInSnapshots === other.timestampsInSnapshots &&\n this.credentials === other.credentials &&\n this.cacheSizeBytes === other.cacheSizeBytes &&\n this.forceLongPolling === other.forceLongPolling &&\n this.ignoreUndefinedProperties === other.ignoreUndefinedProperties\n );\n }\n}\n\n/**\n * The root reference to the database.\n */\nexport class Firestore implements firestore.FirebaseFirestore, FirebaseService {\n // The objects that are a part of this API are exposed to third-parties as\n // compiled javascript so we want to flag our private members with a leading\n // underscore to discourage their use.\n readonly _databaseId: DatabaseId;\n private readonly _persistenceKey: string;\n private readonly _componentProvider: ComponentProvider;\n private _credentials: CredentialsProvider;\n private readonly _firebaseApp: FirebaseApp | null = null;\n private _settings: FirestoreSettings;\n\n // The firestore client instance. This will be available as soon as\n // configureClient is called, but any calls against it will block until\n // setup has completed.\n //\n // Operations on the _firestoreClient don't block on _firestoreReady. Those\n // are already set to synchronize on the async queue.\n private _firestoreClient: FirestoreClient | undefined;\n\n // Public for use in tests.\n // TODO(mikelehen): Use modularized initialization instead.\n readonly _queue = new AsyncQueue();\n\n _userDataReader: UserDataReader | undefined;\n\n // Note: We are using `MemoryComponentProvider` as a default\n // ComponentProvider to ensure backwards compatibility with the format\n // expected by the console build.\n constructor(\n databaseIdOrApp: FirestoreDatabase | FirebaseApp,\n authProvider: Provider<FirebaseAuthInternalName>,\n componentProvider: ComponentProvider = new MemoryComponentProvider()\n ) {\n if (typeof (databaseIdOrApp as FirebaseApp).options === 'object') {\n // This is very likely a Firebase app object\n // TODO(b/34177605): Can we somehow use instanceof?\n const app = databaseIdOrApp as FirebaseApp;\n this._firebaseApp = app;\n this._databaseId = Firestore.databaseIdFromApp(app);\n this._persistenceKey = app.name;\n this._credentials = new FirebaseCredentialsProvider(authProvider);\n } else {\n const external = databaseIdOrApp as FirestoreDatabase;\n if (!external.projectId) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Must provide projectId'\n );\n }\n\n this._databaseId = new DatabaseId(external.projectId, external.database);\n // Use a default persistenceKey that lines up with FirebaseApp.\n this._persistenceKey = '[DEFAULT]';\n this._credentials = new EmptyCredentialsProvider();\n }\n\n this._componentProvider = componentProvider;\n this._settings = new FirestoreSettings({});\n }\n\n get _dataReader(): UserDataReader {\n debugAssert(\n !!this._firestoreClient,\n 'Cannot obtain UserDataReader before instance is intitialized'\n );\n if (!this._userDataReader) {\n // Lazy initialize UserDataReader once the settings are frozen\n this._userDataReader = new UserDataReader(\n this._databaseId,\n this._settings.ignoreUndefinedProperties\n );\n }\n return this._userDataReader;\n }\n\n settings(settingsLiteral: firestore.Settings): void {\n validateExactNumberOfArgs('Firestore.settings', arguments, 1);\n validateArgType('Firestore.settings', 'object', 1, settingsLiteral);\n\n const newSettings = new FirestoreSettings(settingsLiteral);\n if (this._firestoreClient && !this._settings.isEqual(newSettings)) {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'Firestore has already been started and its settings can no longer ' +\n 'be changed. You can only call settings() before calling any other ' +\n 'methods on a Firestore object.'\n );\n }\n\n this._settings = newSettings;\n if (newSettings.credentials !== undefined) {\n this._credentials = makeCredentialsProvider(newSettings.credentials);\n }\n }\n\n enableNetwork(): Promise<void> {\n this.ensureClientConfigured();\n return this._firestoreClient!.enableNetwork();\n }\n\n disableNetwork(): Promise<void> {\n this.ensureClientConfigured();\n return this._firestoreClient!.disableNetwork();\n }\n\n enablePersistence(settings?: firestore.PersistenceSettings): Promise<void> {\n if (this._firestoreClient) {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'Firestore has already been started and persistence can no longer ' +\n 'be enabled. You can only call enablePersistence() before calling ' +\n 'any other methods on a Firestore object.'\n );\n }\n\n let synchronizeTabs = false;\n\n if (settings) {\n if (settings.experimentalTabSynchronization !== undefined) {\n logError(\n \"The 'experimentalTabSynchronization' setting will be removed. Use 'synchronizeTabs' instead.\"\n );\n }\n synchronizeTabs =\n settings.synchronizeTabs ??\n settings.experimentalTabSynchronization ??\n DEFAULT_SYNCHRONIZE_TABS;\n }\n\n return this.configureClient(this._componentProvider, {\n durable: true,\n cacheSizeBytes: this._settings.cacheSizeBytes,\n synchronizeTabs\n });\n }\n\n async clearPersistence(): Promise<void> {\n if (\n this._firestoreClient !== undefined &&\n !this._firestoreClient.clientTerminated\n ) {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'Persistence cannot be cleared after this Firestore instance is initialized.'\n );\n }\n\n const deferred = new Deferred<void>();\n this._queue.enqueueAndForgetEvenAfterShutdown(async () => {\n try {\n const databaseInfo = this.makeDatabaseInfo();\n await this._componentProvider.clearPersistence(databaseInfo);\n deferred.resolve();\n } catch (e) {\n deferred.reject(e);\n }\n });\n return deferred.promise;\n }\n\n terminate(): Promise<void> {\n (this.app as _FirebaseApp)._removeServiceInstance('firestore');\n return this.INTERNAL.delete();\n }\n\n get _isTerminated(): boolean {\n this.ensureClientConfigured();\n return this._firestoreClient!.clientTerminated;\n }\n\n waitForPendingWrites(): Promise<void> {\n this.ensureClientConfigured();\n return this._firestoreClient!.waitForPendingWrites();\n }\n\n onSnapshotsInSync(observer: PartialObserver<void>): Unsubscribe;\n onSnapshotsInSync(onSync: () => void): Unsubscribe;\n onSnapshotsInSync(arg: unknown): Unsubscribe {\n this.ensureClientConfigured();\n\n if (isPartialObserver(arg)) {\n return this.onSnapshotsInSyncInternal(arg as PartialObserver<void>);\n } else {\n validateArgType('Firestore.onSnapshotsInSync', 'function', 1, arg);\n const observer: PartialObserver<void> = {\n next: arg as () => void\n };\n return this.onSnapshotsInSyncInternal(observer);\n }\n }\n\n private onSnapshotsInSyncInternal(\n observer: PartialObserver<void>\n ): Unsubscribe {\n const errHandler = (err: Error): void => {\n throw fail('Uncaught Error in onSnapshotsInSync');\n };\n const asyncObserver = new AsyncObserver<void>({\n next: () => {\n if (observer.next) {\n observer.next();\n }\n },\n error: errHandler\n });\n this._firestoreClient!.addSnapshotsInSyncListener(asyncObserver);\n return () => {\n asyncObserver.mute();\n this._firestoreClient!.removeSnapshotsInSyncListener(asyncObserver);\n };\n }\n\n ensureClientConfigured(): FirestoreClient {\n if (!this._firestoreClient) {\n // Kick off starting the client but don't actually wait for it.\n // eslint-disable-next-line @typescript-eslint/no-floating-promises\n this.configureClient(new MemoryComponentProvider(), {\n durable: false\n });\n }\n return this._firestoreClient as FirestoreClient;\n }\n\n private makeDatabaseInfo(): DatabaseInfo {\n return new DatabaseInfo(\n this._databaseId,\n this._persistenceKey,\n this._settings.host,\n this._settings.ssl,\n this._settings.forceLongPolling\n );\n }\n\n private configureClient(\n componentProvider: ComponentProvider,\n persistenceSettings: PersistenceSettings\n ): Promise<void> {\n debugAssert(!!this._settings.host, 'FirestoreSettings.host is not set');\n\n debugAssert(\n !this._firestoreClient,\n 'configureClient() called multiple times'\n );\n\n const databaseInfo = this.makeDatabaseInfo();\n\n this._firestoreClient = new FirestoreClient(\n PlatformSupport.getPlatform(),\n databaseInfo,\n this._credentials,\n this._queue\n );\n\n return this._firestoreClient.start(componentProvider, persistenceSettings);\n }\n\n private static databaseIdFromApp(app: FirebaseApp): DatabaseId {\n if (!contains(app.options, 'projectId')) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n '\"projectId\" not provided in firebase.initializeApp.'\n );\n }\n\n const projectId = app.options.projectId;\n if (!projectId || typeof projectId !== 'string') {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'projectId must be a string in FirebaseApp.options'\n );\n }\n return new DatabaseId(projectId);\n }\n\n get app(): FirebaseApp {\n if (!this._firebaseApp) {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n \"Firestore was not initialized using the Firebase SDK. 'app' is \" +\n 'not available'\n );\n }\n return this._firebaseApp;\n }\n\n INTERNAL = {\n delete: async (): Promise<void> => {\n // The client must be initalized to ensure that all subsequent API usage\n // throws an exception.\n this.ensureClientConfigured();\n await this._firestoreClient!.terminate();\n }\n };\n\n collection(pathString: string): firestore.CollectionReference {\n validateExactNumberOfArgs('Firestore.collection', arguments, 1);\n validateArgType('Firestore.collection', 'non-empty string', 1, pathString);\n this.ensureClientConfigured();\n return new CollectionReference(ResourcePath.fromString(pathString), this);\n }\n\n doc(pathString: string): firestore.DocumentReference {\n validateExactNumberOfArgs('Firestore.doc', arguments, 1);\n validateArgType('Firestore.doc', 'non-empty string', 1, pathString);\n this.ensureClientConfigured();\n return DocumentReference.forPath(ResourcePath.fromString(pathString), this);\n }\n\n collectionGroup(collectionId: string): firestore.Query {\n validateExactNumberOfArgs('Firestore.collectionGroup', arguments, 1);\n validateArgType(\n 'Firestore.collectionGroup',\n 'non-empty string',\n 1,\n collectionId\n );\n if (collectionId.indexOf('/') >= 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid collection ID '${collectionId}' passed to function ` +\n `Firestore.collectionGroup(). Collection IDs must not contain '/'.`\n );\n }\n this.ensureClientConfigured();\n return new Query(\n new InternalQuery(ResourcePath.EMPTY_PATH, collectionId),\n this\n );\n }\n\n runTransaction<T>(\n updateFunction: (transaction: firestore.Transaction) => Promise<T>\n ): Promise<T> {\n validateExactNumberOfArgs('Firestore.runTransaction', arguments, 1);\n validateArgType('Firestore.runTransaction', 'function', 1, updateFunction);\n return this.ensureClientConfigured().transaction(\n (transaction: InternalTransaction) => {\n return updateFunction(new Transaction(this, transaction));\n }\n );\n }\n\n batch(): firestore.WriteBatch {\n this.ensureClientConfigured();\n\n return new WriteBatch(this);\n }\n\n static get logLevel(): firestore.LogLevel {\n switch (getLogLevel()) {\n case LogLevel.DEBUG:\n return 'debug';\n case LogLevel.SILENT:\n return 'silent';\n default:\n // The default log level is error\n return 'error';\n }\n }\n\n static setLogLevel(level: firestore.LogLevel): void {\n validateExactNumberOfArgs('Firestore.setLogLevel', arguments, 1);\n validateArgType('Firestore.setLogLevel', 'non-empty string', 1, level);\n switch (level) {\n case 'debug':\n setLogLevel(LogLevel.DEBUG);\n break;\n case 'error':\n setLogLevel(LogLevel.ERROR);\n break;\n case 'silent':\n setLogLevel(LogLevel.SILENT);\n break;\n default:\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid log level: ' + level\n );\n }\n }\n\n // Note: this is not a property because the minifier can't work correctly with\n // the way TypeScript compiler outputs properties.\n _areTimestampsInSnapshotsEnabled(): boolean {\n return this._settings.timestampsInSnapshots;\n }\n}\n\n/**\n * A reference to a transaction.\n */\nexport class Transaction implements firestore.Transaction {\n constructor(\n private _firestore: Firestore,\n private _transaction: InternalTransaction\n ) {}\n\n get<T>(\n documentRef: firestore.DocumentReference<T>\n ): Promise<firestore.DocumentSnapshot<T>> {\n validateExactNumberOfArgs('Transaction.get', arguments, 1);\n const ref = validateReference(\n 'Transaction.get',\n documentRef,\n this._firestore\n );\n return this._transaction\n .lookup([ref._key])\n .then((docs: MaybeDocument[]) => {\n if (!docs || docs.length !== 1) {\n return fail('Mismatch in docs returned from document lookup.');\n }\n const doc = docs[0];\n if (doc instanceof NoDocument) {\n return new DocumentSnapshot<T>(\n this._firestore,\n ref._key,\n null,\n /* fromCache= */ false,\n /* hasPendingWrites= */ false,\n ref._converter\n );\n } else if (doc instanceof Document) {\n return new DocumentSnapshot<T>(\n this._firestore,\n ref._key,\n doc,\n /* fromCache= */ false,\n /* hasPendingWrites= */ false,\n ref._converter\n );\n } else {\n throw fail(\n `BatchGetDocumentsRequest returned unexpected document type: ${doc.constructor.name}`\n );\n }\n });\n }\n\n set<T>(\n documentRef: firestore.DocumentReference<T>,\n value: T,\n options?: firestore.SetOptions\n ): Transaction {\n validateBetweenNumberOfArgs('Transaction.set', arguments, 2, 3);\n const ref = validateReference(\n 'Transaction.set',\n documentRef,\n this._firestore\n );\n options = validateSetOptions('Transaction.set', options);\n const [convertedValue, functionName] = applyFirestoreDataConverter(\n ref._converter,\n value,\n 'Transaction.set'\n );\n const parsed =\n options.merge || options.mergeFields\n ? this._firestore._dataReader.parseMergeData(\n functionName,\n convertedValue,\n options.mergeFields\n )\n : this._firestore._dataReader.parseSetData(\n functionName,\n convertedValue\n );\n this._transaction.set(ref._key, parsed);\n return this;\n }\n\n update(\n documentRef: firestore.DocumentReference<unknown>,\n value: firestore.UpdateData\n ): Transaction;\n update(\n documentRef: firestore.DocumentReference<unknown>,\n field: string | ExternalFieldPath,\n value: unknown,\n ...moreFieldsAndValues: unknown[]\n ): Transaction;\n update(\n documentRef: firestore.DocumentReference<unknown>,\n fieldOrUpdateData: string | ExternalFieldPath | firestore.UpdateData,\n value?: unknown,\n ...moreFieldsAndValues: unknown[]\n ): Transaction {\n let ref;\n let parsed;\n\n if (\n typeof fieldOrUpdateData === 'string' ||\n fieldOrUpdateData instanceof ExternalFieldPath\n ) {\n validateAtLeastNumberOfArgs('Transaction.update', arguments, 3);\n ref = validateReference(\n 'Transaction.update',\n documentRef,\n this._firestore\n );\n parsed = this._firestore._dataReader.parseUpdateVarargs(\n 'Transaction.update',\n fieldOrUpdateData,\n value,\n moreFieldsAndValues\n );\n } else {\n validateExactNumberOfArgs('Transaction.update', arguments, 2);\n ref = validateReference(\n 'Transaction.update',\n documentRef,\n this._firestore\n );\n parsed = this._firestore._dataReader.parseUpdateData(\n 'Transaction.update',\n fieldOrUpdateData\n );\n }\n\n this._transaction.update(ref._key, parsed);\n return this;\n }\n\n delete(documentRef: firestore.DocumentReference<unknown>): Transaction {\n validateExactNumberOfArgs('Transaction.delete', arguments, 1);\n const ref = validateReference(\n 'Transaction.delete',\n documentRef,\n this._firestore\n );\n this._transaction.delete(ref._key);\n return this;\n }\n}\n\nexport class WriteBatch implements firestore.WriteBatch {\n private _mutations = [] as Mutation[];\n private _committed = false;\n\n constructor(private _firestore: Firestore) {}\n\n set<T>(\n documentRef: firestore.DocumentReference<T>,\n value: T,\n options?: firestore.SetOptions\n ): WriteBatch {\n validateBetweenNumberOfArgs('WriteBatch.set', arguments, 2, 3);\n this.verifyNotCommitted();\n const ref = validateReference(\n 'WriteBatch.set',\n documentRef,\n this._firestore\n );\n options = validateSetOptions('WriteBatch.set', options);\n const [convertedValue, functionName] = applyFirestoreDataConverter(\n ref._converter,\n value,\n 'WriteBatch.set'\n );\n const parsed =\n options.merge || options.mergeFields\n ? this._firestore._dataReader.parseMergeData(\n functionName,\n convertedValue,\n options.mergeFields\n )\n : this._firestore._dataReader.parseSetData(\n functionName,\n convertedValue\n );\n this._mutations = this._mutations.concat(\n parsed.toMutations(ref._key, Precondition.none())\n );\n return this;\n }\n\n update(\n documentRef: firestore.DocumentReference<unknown>,\n value: firestore.UpdateData\n ): WriteBatch;\n update(\n documentRef: firestore.DocumentReference<unknown>,\n field: string | ExternalFieldPath,\n value: unknown,\n ...moreFieldsAndValues: unknown[]\n ): WriteBatch;\n update(\n documentRef: firestore.DocumentReference<unknown>,\n fieldOrUpdateData: string | ExternalFieldPath | firestore.UpdateData,\n value?: unknown,\n ...moreFieldsAndValues: unknown[]\n ): WriteBatch {\n this.verifyNotCommitted();\n\n let ref;\n let parsed;\n\n if (\n typeof fieldOrUpdateData === 'string' ||\n fieldOrUpdateData instanceof ExternalFieldPath\n ) {\n validateAtLeastNumberOfArgs('WriteBatch.update', arguments, 3);\n ref = validateReference(\n 'WriteBatch.update',\n documentRef,\n this._firestore\n );\n parsed = this._firestore._dataReader.parseUpdateVarargs(\n 'WriteBatch.update',\n fieldOrUpdateData,\n value,\n moreFieldsAndValues\n );\n } else {\n validateExactNumberOfArgs('WriteBatch.update', arguments, 2);\n ref = validateReference(\n 'WriteBatch.update',\n documentRef,\n this._firestore\n );\n parsed = this._firestore._dataReader.parseUpdateData(\n 'WriteBatch.update',\n fieldOrUpdateData\n );\n }\n\n this._mutations = this._mutations.concat(\n parsed.toMutations(ref._key, Precondition.exists(true))\n );\n return this;\n }\n\n delete(documentRef: firestore.DocumentReference<unknown>): WriteBatch {\n validateExactNumberOfArgs('WriteBatch.delete', arguments, 1);\n this.verifyNotCommitted();\n const ref = validateReference(\n 'WriteBatch.delete',\n documentRef,\n this._firestore\n );\n this._mutations = this._mutations.concat(\n new DeleteMutation(ref._key, Precondition.none())\n );\n return this;\n }\n\n commit(): Promise<void> {\n this.verifyNotCommitted();\n this._committed = true;\n if (this._mutations.length > 0) {\n return this._firestore.ensureClientConfigured().write(this._mutations);\n }\n\n return Promise.resolve();\n }\n\n private verifyNotCommitted(): void {\n if (this._committed) {\n throw new FirestoreError(\n Code.FAILED_PRECONDITION,\n 'A write batch can no longer be used after commit() ' +\n 'has been called.'\n );\n }\n }\n}\n\n/**\n * A reference to a particular document in a collection in the database.\n */\nexport class DocumentReference<T = firestore.DocumentData>\n implements firestore.DocumentReference<T> {\n private _firestoreClient: FirestoreClient;\n\n constructor(\n public _key: DocumentKey,\n readonly firestore: Firestore,\n readonly _converter?: firestore.FirestoreDataConverter<T>\n ) {\n this._firestoreClient = this.firestore.ensureClientConfigured();\n }\n\n static forPath<U>(\n path: ResourcePath,\n firestore: Firestore,\n converter?: firestore.FirestoreDataConverter<U>\n ): DocumentReference<U> {\n if (path.length % 2 !== 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid document reference. Document ' +\n 'references must have an even number of segments, but ' +\n `${path.canonicalString()} has ${path.length}`\n );\n }\n return new DocumentReference(new DocumentKey(path), firestore, converter);\n }\n\n get id(): string {\n return this._key.path.lastSegment();\n }\n\n get parent(): firestore.CollectionReference<T> {\n return new CollectionReference(\n this._key.path.popLast(),\n this.firestore,\n this._converter\n );\n }\n\n get path(): string {\n return this._key.path.canonicalString();\n }\n\n collection(\n pathString: string\n ): firestore.CollectionReference<firestore.DocumentData> {\n validateExactNumberOfArgs('DocumentReference.collection', arguments, 1);\n validateArgType(\n 'DocumentReference.collection',\n 'non-empty string',\n 1,\n pathString\n );\n if (!pathString) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Must provide a non-empty collection name to collection()'\n );\n }\n const path = ResourcePath.fromString(pathString);\n return new CollectionReference(this._key.path.child(path), this.firestore);\n }\n\n isEqual(other: firestore.DocumentReference<T>): boolean {\n if (!(other instanceof DocumentReference)) {\n throw invalidClassError('isEqual', 'DocumentReference', 1, other);\n }\n return (\n this.firestore === other.firestore &&\n this._key.isEqual(other._key) &&\n this._converter === other._converter\n );\n }\n\n set(\n value: firestore.DocumentData,\n options?: firestore.SetOptions\n ): Promise<void>;\n set(value: T, options?: firestore.SetOptions): Promise<void> {\n validateBetweenNumberOfArgs('DocumentReference.set', arguments, 1, 2);\n options = validateSetOptions('DocumentReference.set', options);\n const [convertedValue, functionName] = applyFirestoreDataConverter(\n this._converter,\n value,\n 'DocumentReference.set'\n );\n const parsed =\n options.merge || options.mergeFields\n ? this.firestore._dataReader.parseMergeData(\n functionName,\n convertedValue,\n options.mergeFields\n )\n : this.firestore._dataReader.parseSetData(functionName, convertedValue);\n return this._firestoreClient.write(\n parsed.toMutations(this._key, Precondition.none())\n );\n }\n\n update(value: firestore.UpdateData): Promise<void>;\n update(\n field: string | ExternalFieldPath,\n value: unknown,\n ...moreFieldsAndValues: unknown[]\n ): Promise<void>;\n update(\n fieldOrUpdateData: string | ExternalFieldPath | firestore.UpdateData,\n value?: unknown,\n ...moreFieldsAndValues: unknown[]\n ): Promise<void> {\n let parsed;\n\n if (\n typeof fieldOrUpdateData === 'string' ||\n fieldOrUpdateData instanceof ExternalFieldPath\n ) {\n validateAtLeastNumberOfArgs('DocumentReference.update', arguments, 2);\n parsed = this.firestore._dataReader.parseUpdateVarargs(\n 'DocumentReference.update',\n fieldOrUpdateData,\n value,\n moreFieldsAndValues\n );\n } else {\n validateExactNumberOfArgs('DocumentReference.update', arguments, 1);\n parsed = this.firestore._dataReader.parseUpdateData(\n 'DocumentReference.update',\n fieldOrUpdateData\n );\n }\n\n return this._firestoreClient.write(\n parsed.toMutations(this._key, Precondition.exists(true))\n );\n }\n\n delete(): Promise<void> {\n validateExactNumberOfArgs('DocumentReference.delete', arguments, 0);\n return this._firestoreClient.write([\n new DeleteMutation(this._key, Precondition.none())\n ]);\n }\n\n onSnapshot(\n observer: PartialObserver<firestore.DocumentSnapshot<T>>\n ): Unsubscribe;\n onSnapshot(\n options: firestore.SnapshotListenOptions,\n observer: PartialObserver<firestore.DocumentSnapshot<T>>\n ): Unsubscribe;\n onSnapshot(\n onNext: NextFn<firestore.DocumentSnapshot<T>>,\n onError?: ErrorFn,\n onCompletion?: CompleteFn\n ): Unsubscribe;\n onSnapshot(\n options: firestore.SnapshotListenOptions,\n onNext: NextFn<firestore.DocumentSnapshot<T>>,\n onError?: ErrorFn,\n onCompletion?: CompleteFn\n ): Unsubscribe;\n\n onSnapshot(...args: unknown[]): Unsubscribe {\n validateBetweenNumberOfArgs(\n 'DocumentReference.onSnapshot',\n arguments,\n 1,\n 4\n );\n let options: firestore.SnapshotListenOptions = {\n includeMetadataChanges: false\n };\n let observer: PartialObserver<firestore.DocumentSnapshot<T>>;\n let currArg = 0;\n if (\n typeof args[currArg] === 'object' &&\n !isPartialObserver(args[currArg])\n ) {\n options = args[currArg] as firestore.SnapshotListenOptions;\n validateOptionNames('DocumentReference.onSnapshot', options, [\n 'includeMetadataChanges'\n ]);\n validateNamedOptionalType(\n 'DocumentReference.onSnapshot',\n 'boolean',\n 'includeMetadataChanges',\n options.includeMetadataChanges\n );\n currArg++;\n }\n\n const internalOptions = {\n includeMetadataChanges: options.includeMetadataChanges\n };\n\n if (isPartialObserver(args[currArg])) {\n observer = args[currArg] as PartialObserver<\n firestore.DocumentSnapshot<T>\n >;\n } else {\n validateArgType(\n 'DocumentReference.onSnapshot',\n 'function',\n currArg,\n args[currArg]\n );\n validateOptionalArgType(\n 'DocumentReference.onSnapshot',\n 'function',\n currArg + 1,\n args[currArg + 1]\n );\n validateOptionalArgType(\n 'DocumentReference.onSnapshot',\n 'function',\n currArg + 2,\n args[currArg + 2]\n );\n observer = {\n next: args[currArg] as NextFn<firestore.DocumentSnapshot<T>>,\n error: args[currArg + 1] as ErrorFn,\n complete: args[currArg + 2] as CompleteFn\n };\n }\n return this.onSnapshotInternal(internalOptions, observer);\n }\n\n private onSnapshotInternal(\n options: ListenOptions,\n observer: PartialObserver<firestore.DocumentSnapshot<T>>\n ): Unsubscribe {\n let errHandler = (err: Error): void => {\n console.error('Uncaught Error in onSnapshot:', err);\n };\n if (observer.error) {\n errHandler = observer.error.bind(observer);\n }\n\n const asyncObserver = new AsyncObserver<ViewSnapshot>({\n next: snapshot => {\n if (observer.next) {\n debugAssert(\n snapshot.docs.size <= 1,\n 'Too many documents returned on a document query'\n );\n const doc = snapshot.docs.get(this._key);\n\n observer.next(\n new DocumentSnapshot(\n this.firestore,\n this._key,\n doc,\n snapshot.fromCache,\n snapshot.hasPendingWrites,\n this._converter\n )\n );\n }\n },\n error: errHandler\n });\n const internalListener = this._firestoreClient.listen(\n InternalQuery.atPath(this._key.path),\n asyncObserver,\n options\n );\n\n return () => {\n asyncObserver.mute();\n this._firestoreClient.unlisten(internalListener);\n };\n }\n\n get(options?: firestore.GetOptions): Promise<firestore.DocumentSnapshot<T>> {\n validateBetweenNumberOfArgs('DocumentReference.get', arguments, 0, 1);\n validateGetOptions('DocumentReference.get', options);\n return new Promise(\n (resolve: Resolver<firestore.DocumentSnapshot<T>>, reject: Rejecter) => {\n if (options && options.source === 'cache') {\n this.firestore\n .ensureClientConfigured()\n .getDocumentFromLocalCache(this._key)\n .then(doc => {\n resolve(\n new DocumentSnapshot(\n this.firestore,\n this._key,\n doc,\n /*fromCache=*/ true,\n doc instanceof Document ? doc.hasLocalMutations : false,\n this._converter\n )\n );\n }, reject);\n } else {\n this.getViaSnapshotListener(resolve, reject, options);\n }\n }\n );\n }\n\n private getViaSnapshotListener(\n resolve: Resolver<firestore.DocumentSnapshot<T>>,\n reject: Rejecter,\n options?: firestore.GetOptions\n ): void {\n const unlisten = this.onSnapshotInternal(\n {\n includeMetadataChanges: true,\n waitForSyncWhenOnline: true\n },\n {\n next: (snap: firestore.DocumentSnapshot<T>) => {\n // Remove query first before passing event to user to avoid\n // user actions affecting the now stale query.\n unlisten();\n\n if (!snap.exists && snap.metadata.fromCache) {\n // TODO(dimond): If we're online and the document doesn't\n // exist then we resolve with a doc.exists set to false. If\n // we're offline however, we reject the Promise in this\n // case. Two options: 1) Cache the negative response from\n // the server so we can deliver that even when you're\n // offline 2) Actually reject the Promise in the online case\n // if the document doesn't exist.\n reject(\n new FirestoreError(\n Code.UNAVAILABLE,\n 'Failed to get document because the client is ' + 'offline.'\n )\n );\n } else if (\n snap.exists &&\n snap.metadata.fromCache &&\n options &&\n options.source === 'server'\n ) {\n reject(\n new FirestoreError(\n Code.UNAVAILABLE,\n 'Failed to get document from server. (However, this ' +\n 'document does exist in the local cache. Run again ' +\n 'without setting source to \"server\" to ' +\n 'retrieve the cached document.)'\n )\n );\n } else {\n resolve(snap);\n }\n },\n error: reject\n }\n );\n }\n\n withConverter<U>(\n converter: firestore.FirestoreDataConverter<U>\n ): firestore.DocumentReference<U> {\n return new DocumentReference<U>(this._key, this.firestore, converter);\n }\n}\n\nclass SnapshotMetadata implements firestore.SnapshotMetadata {\n constructor(\n readonly hasPendingWrites: boolean,\n readonly fromCache: boolean\n ) {}\n\n isEqual(other: firestore.SnapshotMetadata): boolean {\n return (\n this.hasPendingWrites === other.hasPendingWrites &&\n this.fromCache === other.fromCache\n );\n }\n}\n\n/**\n * Options interface that can be provided to configure the deserialization of\n * DocumentSnapshots.\n */\nexport interface SnapshotOptions extends firestore.SnapshotOptions {}\n\nexport class DocumentSnapshot<T = firestore.DocumentData>\n implements firestore.DocumentSnapshot<T> {\n constructor(\n private _firestore: Firestore,\n private _key: DocumentKey,\n public _document: Document | null,\n private _fromCache: boolean,\n private _hasPendingWrites: boolean,\n private readonly _converter?: firestore.FirestoreDataConverter<T>\n ) {}\n\n data(options?: firestore.SnapshotOptions): T | undefined {\n validateBetweenNumberOfArgs('DocumentSnapshot.data', arguments, 0, 1);\n options = validateSnapshotOptions('DocumentSnapshot.data', options);\n if (!this._document) {\n return undefined;\n } else {\n // We only want to use the converter and create a new DocumentSnapshot\n // if a converter has been provided.\n if (this._converter) {\n const snapshot = new QueryDocumentSnapshot(\n this._firestore,\n this._key,\n this._document,\n this._fromCache,\n this._hasPendingWrites\n );\n return this._converter.fromFirestore(snapshot, options);\n } else {\n const userDataWriter = new UserDataWriter(\n this._firestore,\n this._firestore._areTimestampsInSnapshotsEnabled(),\n options.serverTimestamps,\n /* converter= */ undefined\n );\n return userDataWriter.convertValue(this._document.toProto()) as T;\n }\n }\n }\n\n get(\n fieldPath: string | ExternalFieldPath,\n options?: firestore.SnapshotOptions\n ): unknown {\n validateBetweenNumberOfArgs('DocumentSnapshot.get', arguments, 1, 2);\n options = validateSnapshotOptions('DocumentSnapshot.get', options);\n if (this._document) {\n const value = this._document\n .data()\n .field(fieldPathFromArgument('DocumentSnapshot.get', fieldPath));\n if (value !== null) {\n const userDataWriter = new UserDataWriter(\n this._firestore,\n this._firestore._areTimestampsInSnapshotsEnabled(),\n options.serverTimestamps,\n this._converter\n );\n return userDataWriter.convertValue(value);\n }\n }\n return undefined;\n }\n\n get id(): string {\n return this._key.path.lastSegment();\n }\n\n get ref(): firestore.DocumentReference<T> {\n return new DocumentReference<T>(\n this._key,\n this._firestore,\n this._converter\n );\n }\n\n get exists(): boolean {\n return this._document !== null;\n }\n\n get metadata(): firestore.SnapshotMetadata {\n return new SnapshotMetadata(this._hasPendingWrites, this._fromCache);\n }\n\n isEqual(other: firestore.DocumentSnapshot<T>): boolean {\n if (!(other instanceof DocumentSnapshot)) {\n throw invalidClassError('isEqual', 'DocumentSnapshot', 1, other);\n }\n return (\n this._firestore === other._firestore &&\n this._fromCache === other._fromCache &&\n this._key.isEqual(other._key) &&\n (this._document === null\n ? other._document === null\n : this._document.isEqual(other._document)) &&\n this._converter === other._converter\n );\n }\n}\n\nexport class QueryDocumentSnapshot<T = firestore.DocumentData>\n extends DocumentSnapshot<T>\n implements firestore.QueryDocumentSnapshot<T> {\n data(options?: SnapshotOptions): T {\n const data = super.data(options);\n debugAssert(\n data !== undefined,\n 'Document in a QueryDocumentSnapshot should exist'\n );\n return data;\n }\n}\n\nexport class Query<T = firestore.DocumentData> implements firestore.Query<T> {\n constructor(\n public _query: InternalQuery,\n readonly firestore: Firestore,\n protected readonly _converter?: firestore.FirestoreDataConverter<T>\n ) {}\n\n where(\n field: string | ExternalFieldPath,\n opStr: firestore.WhereFilterOp,\n value: unknown\n ): firestore.Query<T> {\n validateExactNumberOfArgs('Query.where', arguments, 3);\n validateDefined('Query.where', 3, value);\n\n // Enumerated from the WhereFilterOp type in index.d.ts.\n const whereFilterOpEnums = [\n Operator.LESS_THAN,\n Operator.LESS_THAN_OR_EQUAL,\n Operator.EQUAL,\n Operator.GREATER_THAN_OR_EQUAL,\n Operator.GREATER_THAN,\n Operator.ARRAY_CONTAINS,\n Operator.IN,\n Operator.ARRAY_CONTAINS_ANY\n ];\n const op = validateStringEnum('Query.where', whereFilterOpEnums, 2, opStr);\n\n let fieldValue: api.Value;\n const fieldPath = fieldPathFromArgument('Query.where', field);\n if (fieldPath.isKeyField()) {\n if (\n op === Operator.ARRAY_CONTAINS ||\n op === Operator.ARRAY_CONTAINS_ANY\n ) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid Query. You can't perform '${op}' ` +\n 'queries on FieldPath.documentId().'\n );\n } else if (op === Operator.IN) {\n this.validateDisjunctiveFilterElements(value, op);\n const referenceList: api.Value[] = [];\n for (const arrayValue of value as api.Value[]) {\n referenceList.push(this.parseDocumentIdValue(arrayValue));\n }\n fieldValue = { arrayValue: { values: referenceList } };\n } else {\n fieldValue = this.parseDocumentIdValue(value);\n }\n } else {\n if (op === Operator.IN || op === Operator.ARRAY_CONTAINS_ANY) {\n this.validateDisjunctiveFilterElements(value, op);\n }\n fieldValue = this.firestore._dataReader.parseQueryValue(\n 'Query.where',\n value,\n // We only allow nested arrays for IN queries.\n /** allowArrays = */ op === Operator.IN\n );\n }\n const filter = FieldFilter.create(fieldPath, op, fieldValue);\n this.validateNewFilter(filter);\n return new Query(\n this._query.addFilter(filter),\n this.firestore,\n this._converter\n );\n }\n\n orderBy(\n field: string | ExternalFieldPath,\n directionStr?: firestore.OrderByDirection\n ): firestore.Query<T> {\n validateBetweenNumberOfArgs('Query.orderBy', arguments, 1, 2);\n validateOptionalArgType(\n 'Query.orderBy',\n 'non-empty string',\n 2,\n directionStr\n );\n let direction: Direction;\n if (directionStr === undefined || directionStr === 'asc') {\n direction = Direction.ASCENDING;\n } else if (directionStr === 'desc') {\n direction = Direction.DESCENDING;\n } else {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Function Query.orderBy() has unknown direction '${directionStr}', ` +\n `expected 'asc' or 'desc'.`\n );\n }\n if (this._query.startAt !== null) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. You must not call Query.startAt() or ' +\n 'Query.startAfter() before calling Query.orderBy().'\n );\n }\n if (this._query.endAt !== null) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. You must not call Query.endAt() or ' +\n 'Query.endBefore() before calling Query.orderBy().'\n );\n }\n const fieldPath = fieldPathFromArgument('Query.orderBy', field);\n const orderBy = new OrderBy(fieldPath, direction);\n this.validateNewOrderBy(orderBy);\n return new Query(\n this._query.addOrderBy(orderBy),\n this.firestore,\n this._converter\n );\n }\n\n limit(n: number): firestore.Query<T> {\n validateExactNumberOfArgs('Query.limit', arguments, 1);\n validateArgType('Query.limit', 'number', 1, n);\n validatePositiveNumber('Query.limit', 1, n);\n return new Query(\n this._query.withLimitToFirst(n),\n this.firestore,\n this._converter\n );\n }\n\n limitToLast(n: number): firestore.Query<T> {\n validateExactNumberOfArgs('Query.limitToLast', arguments, 1);\n validateArgType('Query.limitToLast', 'number', 1, n);\n validatePositiveNumber('Query.limitToLast', 1, n);\n return new Query(\n this._query.withLimitToLast(n),\n this.firestore,\n this._converter\n );\n }\n\n startAt(\n docOrField: unknown | firestore.DocumentSnapshot<unknown>,\n ...fields: unknown[]\n ): firestore.Query<T> {\n validateAtLeastNumberOfArgs('Query.startAt', arguments, 1);\n const bound = this.boundFromDocOrFields(\n 'Query.startAt',\n docOrField,\n fields,\n /*before=*/ true\n );\n return new Query(\n this._query.withStartAt(bound),\n this.firestore,\n this._converter\n );\n }\n\n startAfter(\n docOrField: unknown | firestore.DocumentSnapshot<unknown>,\n ...fields: unknown[]\n ): firestore.Query<T> {\n validateAtLeastNumberOfArgs('Query.startAfter', arguments, 1);\n const bound = this.boundFromDocOrFields(\n 'Query.startAfter',\n docOrField,\n fields,\n /*before=*/ false\n );\n return new Query(\n this._query.withStartAt(bound),\n this.firestore,\n this._converter\n );\n }\n\n endBefore(\n docOrField: unknown | firestore.DocumentSnapshot<unknown>,\n ...fields: unknown[]\n ): firestore.Query<T> {\n validateAtLeastNumberOfArgs('Query.endBefore', arguments, 1);\n const bound = this.boundFromDocOrFields(\n 'Query.endBefore',\n docOrField,\n fields,\n /*before=*/ true\n );\n return new Query(\n this._query.withEndAt(bound),\n this.firestore,\n this._converter\n );\n }\n\n endAt(\n docOrField: unknown | firestore.DocumentSnapshot<unknown>,\n ...fields: unknown[]\n ): firestore.Query<T> {\n validateAtLeastNumberOfArgs('Query.endAt', arguments, 1);\n const bound = this.boundFromDocOrFields(\n 'Query.endAt',\n docOrField,\n fields,\n /*before=*/ false\n );\n return new Query(\n this._query.withEndAt(bound),\n this.firestore,\n this._converter\n );\n }\n\n isEqual(other: firestore.Query<T>): boolean {\n if (!(other instanceof Query)) {\n throw invalidClassError('isEqual', 'Query', 1, other);\n }\n return (\n this.firestore === other.firestore && this._query.isEqual(other._query)\n );\n }\n\n withConverter<U>(\n converter: firestore.FirestoreDataConverter<U>\n ): firestore.Query<U> {\n return new Query<U>(this._query, this.firestore, converter);\n }\n\n /** Helper function to create a bound from a document or fields */\n private boundFromDocOrFields(\n methodName: string,\n docOrField: unknown | firestore.DocumentSnapshot<T>,\n fields: unknown[],\n before: boolean\n ): Bound {\n validateDefined(methodName, 1, docOrField);\n if (docOrField instanceof DocumentSnapshot) {\n if (fields.length > 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Too many arguments provided to ${methodName}().`\n );\n }\n const snap = docOrField;\n if (!snap.exists) {\n throw new FirestoreError(\n Code.NOT_FOUND,\n `Can't use a DocumentSnapshot that doesn't exist for ` +\n `${methodName}().`\n );\n }\n return this.boundFromDocument(snap._document!, before);\n } else {\n const allFields = [docOrField].concat(fields);\n return this.boundFromFields(methodName, allFields, before);\n }\n }\n\n /**\n * Create a Bound from a query and a document.\n *\n * Note that the Bound will always include the key of the document\n * and so only the provided document will compare equal to the returned\n * position.\n *\n * Will throw if the document does not contain all fields of the order by\n * of the query or if any of the fields in the order by are an uncommitted\n * server timestamp.\n */\n private boundFromDocument(doc: Document, before: boolean): Bound {\n const components: api.Value[] = [];\n\n // Because people expect to continue/end a query at the exact document\n // provided, we need to use the implicit sort order rather than the explicit\n // sort order, because it's guaranteed to contain the document key. That way\n // the position becomes unambiguous and the query continues/ends exactly at\n // the provided document. Without the key (by using the explicit sort\n // orders), multiple documents could match the position, yielding duplicate\n // results.\n for (const orderBy of this._query.orderBy) {\n if (orderBy.field.isKeyField()) {\n components.push(refValue(this.firestore._databaseId, doc.key));\n } else {\n const value = doc.field(orderBy.field);\n if (isServerTimestamp(value)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. You are trying to start or end a query using a ' +\n 'document for which the field \"' +\n orderBy.field +\n '\" is an uncommitted server timestamp. (Since the value of ' +\n 'this field is unknown, you cannot start/end a query with it.)'\n );\n } else if (value !== null) {\n components.push(value);\n } else {\n const field = orderBy.field.canonicalString();\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. You are trying to start or end a query using a ` +\n `document for which the field '${field}' (used as the ` +\n `orderBy) does not exist.`\n );\n }\n }\n }\n return new Bound(components, before);\n }\n\n /**\n * Converts a list of field values to a Bound for the given query.\n */\n private boundFromFields(\n methodName: string,\n values: unknown[],\n before: boolean\n ): Bound {\n // Use explicit order by's because it has to match the query the user made\n const orderBy = this._query.explicitOrderBy;\n if (values.length > orderBy.length) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Too many arguments provided to ${methodName}(). ` +\n `The number of arguments must be less than or equal to the ` +\n `number of Query.orderBy() clauses`\n );\n }\n\n const components: api.Value[] = [];\n for (let i = 0; i < values.length; i++) {\n const rawValue = values[i];\n const orderByComponent = orderBy[i];\n if (orderByComponent.field.isKeyField()) {\n if (typeof rawValue !== 'string') {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. Expected a string for document ID in ` +\n `${methodName}(), but got a ${typeof rawValue}`\n );\n }\n if (\n !this._query.isCollectionGroupQuery() &&\n rawValue.indexOf('/') !== -1\n ) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. When querying a collection and ordering by FieldPath.documentId(), ` +\n `the value passed to ${methodName}() must be a plain document ID, but ` +\n `'${rawValue}' contains a slash.`\n );\n }\n const path = this._query.path.child(ResourcePath.fromString(rawValue));\n if (!DocumentKey.isDocumentKey(path)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. When querying a collection group and ordering by ` +\n `FieldPath.documentId(), the value passed to ${methodName}() must result in a ` +\n `valid document path, but '${path}' is not because it contains an odd number ` +\n `of segments.`\n );\n }\n const key = new DocumentKey(path);\n components.push(refValue(this.firestore._databaseId, key));\n } else {\n const wrapped = this.firestore._dataReader.parseQueryValue(\n methodName,\n rawValue\n );\n components.push(wrapped);\n }\n }\n\n return new Bound(components, before);\n }\n\n onSnapshot(\n observer: PartialObserver<firestore.QuerySnapshot<T>>\n ): Unsubscribe;\n onSnapshot(\n options: firestore.SnapshotListenOptions,\n observer: PartialObserver<firestore.QuerySnapshot<T>>\n ): Unsubscribe;\n onSnapshot(\n onNext: NextFn<firestore.QuerySnapshot<T>>,\n onError?: ErrorFn,\n onCompletion?: CompleteFn\n ): Unsubscribe;\n onSnapshot(\n options: firestore.SnapshotListenOptions,\n onNext: NextFn<firestore.QuerySnapshot<T>>,\n onError?: ErrorFn,\n onCompletion?: CompleteFn\n ): Unsubscribe;\n\n onSnapshot(...args: unknown[]): Unsubscribe {\n validateBetweenNumberOfArgs('Query.onSnapshot', arguments, 1, 4);\n let options: firestore.SnapshotListenOptions = {};\n let observer: PartialObserver<firestore.QuerySnapshot<T>>;\n let currArg = 0;\n if (\n typeof args[currArg] === 'object' &&\n !isPartialObserver(args[currArg])\n ) {\n options = args[currArg] as firestore.SnapshotListenOptions;\n validateOptionNames('Query.onSnapshot', options, [\n 'includeMetadataChanges'\n ]);\n validateNamedOptionalType(\n 'Query.onSnapshot',\n 'boolean',\n 'includeMetadataChanges',\n options.includeMetadataChanges\n );\n currArg++;\n }\n\n if (isPartialObserver(args[currArg])) {\n observer = args[currArg] as PartialObserver<firestore.QuerySnapshot<T>>;\n } else {\n validateArgType('Query.onSnapshot', 'function', currArg, args[currArg]);\n validateOptionalArgType(\n 'Query.onSnapshot',\n 'function',\n currArg + 1,\n args[currArg + 1]\n );\n validateOptionalArgType(\n 'Query.onSnapshot',\n 'function',\n currArg + 2,\n args[currArg + 2]\n );\n observer = {\n next: args[currArg] as NextFn<firestore.QuerySnapshot<T>>,\n error: args[currArg + 1] as ErrorFn,\n complete: args[currArg + 2] as CompleteFn\n };\n }\n this.validateHasExplicitOrderByForLimitToLast(this._query);\n return this.onSnapshotInternal(options, observer);\n }\n\n private onSnapshotInternal(\n options: ListenOptions,\n observer: PartialObserver<firestore.QuerySnapshot<T>>\n ): Unsubscribe {\n let errHandler = (err: Error): void => {\n console.error('Uncaught Error in onSnapshot:', err);\n };\n if (observer.error) {\n errHandler = observer.error.bind(observer);\n }\n\n const asyncObserver = new AsyncObserver<ViewSnapshot>({\n next: (result: ViewSnapshot): void => {\n if (observer.next) {\n observer.next(\n new QuerySnapshot(\n this.firestore,\n this._query,\n result,\n this._converter\n )\n );\n }\n },\n error: errHandler\n });\n\n const firestoreClient = this.firestore.ensureClientConfigured();\n const internalListener = firestoreClient.listen(\n this._query,\n asyncObserver,\n options\n );\n return (): void => {\n asyncObserver.mute();\n firestoreClient.unlisten(internalListener);\n };\n }\n\n private validateHasExplicitOrderByForLimitToLast(query: InternalQuery): void {\n if (query.hasLimitToLast() && query.explicitOrderBy.length === 0) {\n throw new FirestoreError(\n Code.UNIMPLEMENTED,\n 'limitToLast() queries require specifying at least one orderBy() clause'\n );\n }\n }\n\n get(options?: firestore.GetOptions): Promise<firestore.QuerySnapshot<T>> {\n validateBetweenNumberOfArgs('Query.get', arguments, 0, 1);\n validateGetOptions('Query.get', options);\n this.validateHasExplicitOrderByForLimitToLast(this._query);\n return new Promise(\n (resolve: Resolver<firestore.QuerySnapshot<T>>, reject: Rejecter) => {\n if (options && options.source === 'cache') {\n this.firestore\n .ensureClientConfigured()\n .getDocumentsFromLocalCache(this._query)\n .then((viewSnap: ViewSnapshot) => {\n resolve(\n new QuerySnapshot(\n this.firestore,\n this._query,\n viewSnap,\n this._converter\n )\n );\n }, reject);\n } else {\n this.getViaSnapshotListener(resolve, reject, options);\n }\n }\n );\n }\n\n private getViaSnapshotListener(\n resolve: Resolver<firestore.QuerySnapshot<T>>,\n reject: Rejecter,\n options?: firestore.GetOptions\n ): void {\n const unlisten = this.onSnapshotInternal(\n {\n includeMetadataChanges: true,\n waitForSyncWhenOnline: true\n },\n {\n next: (result: firestore.QuerySnapshot<T>) => {\n // Remove query first before passing event to user to avoid\n // user actions affecting the now stale query.\n unlisten();\n\n if (\n result.metadata.fromCache &&\n options &&\n options.source === 'server'\n ) {\n reject(\n new FirestoreError(\n Code.UNAVAILABLE,\n 'Failed to get documents from server. (However, these ' +\n 'documents may exist in the local cache. Run again ' +\n 'without setting source to \"server\" to ' +\n 'retrieve the cached documents.)'\n )\n );\n } else {\n resolve(result);\n }\n },\n error: reject\n }\n );\n }\n\n /**\n * Parses the given documentIdValue into a ReferenceValue, throwing\n * appropriate errors if the value is anything other than a DocumentReference\n * or String, or if the string is malformed.\n */\n private parseDocumentIdValue(documentIdValue: unknown): api.Value {\n if (typeof documentIdValue === 'string') {\n if (documentIdValue === '') {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. When querying with FieldPath.documentId(), you ' +\n 'must provide a valid document ID, but it was an empty string.'\n );\n }\n if (\n !this._query.isCollectionGroupQuery() &&\n documentIdValue.indexOf('/') !== -1\n ) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. When querying a collection by ` +\n `FieldPath.documentId(), you must provide a plain document ID, but ` +\n `'${documentIdValue}' contains a '/' character.`\n );\n }\n const path = this._query.path.child(\n ResourcePath.fromString(documentIdValue)\n );\n if (!DocumentKey.isDocumentKey(path)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. When querying a collection group by ` +\n `FieldPath.documentId(), the value provided must result in a valid document path, ` +\n `but '${path}' is not because it has an odd number of segments (${path.length}).`\n );\n }\n return refValue(this.firestore._databaseId, new DocumentKey(path));\n } else if (documentIdValue instanceof DocumentReference) {\n const ref = documentIdValue as DocumentReference<T>;\n return refValue(this.firestore._databaseId, ref._key);\n } else {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. When querying with FieldPath.documentId(), you must provide a valid ` +\n `string or a DocumentReference, but it was: ` +\n `${valueDescription(documentIdValue)}.`\n );\n }\n }\n\n /**\n * Validates that the value passed into a disjunctrive filter satisfies all\n * array requirements.\n */\n private validateDisjunctiveFilterElements(\n value: unknown,\n operator: Operator\n ): void {\n if (!Array.isArray(value) || value.length === 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid Query. A non-empty array is required for ' +\n `'${operator.toString()}' filters.`\n );\n }\n if (value.length > 10) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid Query. '${operator.toString()}' filters support a ` +\n 'maximum of 10 elements in the value array.'\n );\n }\n if (value.indexOf(null) >= 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid Query. '${operator.toString()}' filters cannot contain 'null' ` +\n 'in the value array.'\n );\n }\n if (value.filter(element => Number.isNaN(element)).length > 0) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid Query. '${operator.toString()}' filters cannot contain 'NaN' ` +\n 'in the value array.'\n );\n }\n }\n\n private validateNewFilter(filter: Filter): void {\n if (filter instanceof FieldFilter) {\n const arrayOps = [Operator.ARRAY_CONTAINS, Operator.ARRAY_CONTAINS_ANY];\n const disjunctiveOps = [Operator.IN, Operator.ARRAY_CONTAINS_ANY];\n const isArrayOp = arrayOps.indexOf(filter.op) >= 0;\n const isDisjunctiveOp = disjunctiveOps.indexOf(filter.op) >= 0;\n\n if (filter.isInequality()) {\n const existingField = this._query.getInequalityFilterField();\n if (existingField !== null && !existingField.isEqual(filter.field)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. All where filters with an inequality' +\n ' (<, <=, >, or >=) must be on the same field. But you have' +\n ` inequality filters on '${existingField.toString()}'` +\n ` and '${filter.field.toString()}'`\n );\n }\n\n const firstOrderByField = this._query.getFirstOrderByField();\n if (firstOrderByField !== null) {\n this.validateOrderByAndInequalityMatch(\n filter.field,\n firstOrderByField\n );\n }\n } else if (isDisjunctiveOp || isArrayOp) {\n // You can have at most 1 disjunctive filter and 1 array filter. Check if\n // the new filter conflicts with an existing one.\n let conflictingOp: Operator | null = null;\n if (isDisjunctiveOp) {\n conflictingOp = this._query.findFilterOperator(disjunctiveOps);\n }\n if (conflictingOp === null && isArrayOp) {\n conflictingOp = this._query.findFilterOperator(arrayOps);\n }\n if (conflictingOp != null) {\n // We special case when it's a duplicate op to give a slightly clearer error message.\n if (conflictingOp === filter.op) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid query. You cannot use more than one ' +\n `'${filter.op.toString()}' filter.`\n );\n } else {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. You cannot use '${filter.op.toString()}' filters ` +\n `with '${conflictingOp.toString()}' filters.`\n );\n }\n }\n }\n }\n }\n\n private validateNewOrderBy(orderBy: OrderBy): void {\n if (this._query.getFirstOrderByField() === null) {\n // This is the first order by. It must match any inequality.\n const inequalityField = this._query.getInequalityFilterField();\n if (inequalityField !== null) {\n this.validateOrderByAndInequalityMatch(inequalityField, orderBy.field);\n }\n }\n }\n\n private validateOrderByAndInequalityMatch(\n inequality: FieldPath,\n orderBy: FieldPath\n ): void {\n if (!orderBy.isEqual(inequality)) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid query. You have a where filter with an inequality ` +\n `(<, <=, >, or >=) on field '${inequality.toString()}' ` +\n `and so you must also use '${inequality.toString()}' ` +\n `as your first Query.orderBy(), but your first Query.orderBy() ` +\n `is on field '${orderBy.toString()}' instead.`\n );\n }\n }\n}\n\nexport class QuerySnapshot<T = firestore.DocumentData>\n implements firestore.QuerySnapshot<T> {\n private _cachedChanges: Array<firestore.DocumentChange<T>> | null = null;\n private _cachedChangesIncludeMetadataChanges: boolean | null = null;\n\n readonly metadata: firestore.SnapshotMetadata;\n\n constructor(\n private readonly _firestore: Firestore,\n private readonly _originalQuery: InternalQuery,\n private readonly _snapshot: ViewSnapshot,\n private readonly _converter?: firestore.FirestoreDataConverter<T>\n ) {\n this.metadata = new SnapshotMetadata(\n _snapshot.hasPendingWrites,\n _snapshot.fromCache\n );\n }\n\n get docs(): Array<firestore.QueryDocumentSnapshot<T>> {\n const result: Array<firestore.QueryDocumentSnapshot<T>> = [];\n this.forEach(doc => result.push(doc));\n return result;\n }\n\n get empty(): boolean {\n return this._snapshot.docs.isEmpty();\n }\n\n get size(): number {\n return this._snapshot.docs.size;\n }\n\n forEach(\n callback: (result: firestore.QueryDocumentSnapshot<T>) => void,\n thisArg?: unknown\n ): void {\n validateBetweenNumberOfArgs('QuerySnapshot.forEach', arguments, 1, 2);\n validateArgType('QuerySnapshot.forEach', 'function', 1, callback);\n this._snapshot.docs.forEach(doc => {\n callback.call(thisArg, this.convertToDocumentImpl(doc));\n });\n }\n\n get query(): firestore.Query<T> {\n return new Query(this._originalQuery, this._firestore, this._converter);\n }\n\n docChanges(\n options?: firestore.SnapshotListenOptions\n ): Array<firestore.DocumentChange<T>> {\n if (options) {\n validateOptionNames('QuerySnapshot.docChanges', options, [\n 'includeMetadataChanges'\n ]);\n validateNamedOptionalType(\n 'QuerySnapshot.docChanges',\n 'boolean',\n 'includeMetadataChanges',\n options.includeMetadataChanges\n );\n }\n\n const includeMetadataChanges = !!(\n options && options.includeMetadataChanges\n );\n\n if (includeMetadataChanges && this._snapshot.excludesMetadataChanges) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'To include metadata changes with your document changes, you must ' +\n 'also pass { includeMetadataChanges:true } to onSnapshot().'\n );\n }\n\n if (\n !this._cachedChanges ||\n this._cachedChangesIncludeMetadataChanges !== includeMetadataChanges\n ) {\n this._cachedChanges = changesFromSnapshot<T>(\n this._firestore,\n includeMetadataChanges,\n this._snapshot,\n this._converter\n );\n this._cachedChangesIncludeMetadataChanges = includeMetadataChanges;\n }\n\n return this._cachedChanges;\n }\n\n /** Check the equality. The call can be very expensive. */\n isEqual(other: firestore.QuerySnapshot<T>): boolean {\n if (!(other instanceof QuerySnapshot)) {\n throw invalidClassError('isEqual', 'QuerySnapshot', 1, other);\n }\n\n return (\n this._firestore === other._firestore &&\n this._originalQuery.isEqual(other._originalQuery) &&\n this._snapshot.isEqual(other._snapshot) &&\n this._converter === other._converter\n );\n }\n\n private convertToDocumentImpl(doc: Document): QueryDocumentSnapshot<T> {\n return new QueryDocumentSnapshot(\n this._firestore,\n doc.key,\n doc,\n this.metadata.fromCache,\n this._snapshot.mutatedKeys.has(doc.key),\n this._converter\n );\n }\n}\n\nexport class CollectionReference<T = firestore.DocumentData> extends Query<T>\n implements firestore.CollectionReference<T> {\n constructor(\n readonly _path: ResourcePath,\n firestore: Firestore,\n _converter?: firestore.FirestoreDataConverter<T>\n ) {\n super(InternalQuery.atPath(_path), firestore, _converter);\n if (_path.length % 2 !== 1) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Invalid collection reference. Collection ' +\n 'references must have an odd number of segments, but ' +\n `${_path.canonicalString()} has ${_path.length}`\n );\n }\n }\n\n get id(): string {\n return this._query.path.lastSegment();\n }\n\n get parent(): firestore.DocumentReference<firestore.DocumentData> | null {\n const parentPath = this._query.path.popLast();\n if (parentPath.isEmpty()) {\n return null;\n } else {\n return new DocumentReference<firestore.DocumentData>(\n new DocumentKey(parentPath),\n this.firestore\n );\n }\n }\n\n get path(): string {\n return this._query.path.canonicalString();\n }\n\n doc(pathString?: string): firestore.DocumentReference<T> {\n validateBetweenNumberOfArgs('CollectionReference.doc', arguments, 0, 1);\n // We allow omission of 'pathString' but explicitly prohibit passing in both\n // 'undefined' and 'null'.\n if (arguments.length === 0) {\n pathString = AutoId.newId();\n }\n validateArgType(\n 'CollectionReference.doc',\n 'non-empty string',\n 1,\n pathString\n );\n if (pathString === '') {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Document path must be a non-empty string'\n );\n }\n const path = ResourcePath.fromString(pathString!);\n return DocumentReference.forPath<T>(\n this._query.path.child(path),\n this.firestore,\n this._converter\n );\n }\n\n add(value: T): Promise<firestore.DocumentReference<T>> {\n validateExactNumberOfArgs('CollectionReference.add', arguments, 1);\n const convertedValue = this._converter\n ? this._converter.toFirestore(value)\n : value;\n validateArgType('CollectionReference.add', 'object', 1, convertedValue);\n const docRef = this.doc();\n return docRef.set(value).then(() => docRef);\n }\n\n withConverter<U>(\n converter: firestore.FirestoreDataConverter<U>\n ): firestore.CollectionReference<U> {\n return new CollectionReference<U>(this._path, this.firestore, converter);\n }\n}\n\nfunction validateSetOptions(\n methodName: string,\n options: firestore.SetOptions | undefined\n): firestore.SetOptions {\n if (options === undefined) {\n return {\n merge: false\n };\n }\n\n validateOptionNames(methodName, options, ['merge', 'mergeFields']);\n validateNamedOptionalType(methodName, 'boolean', 'merge', options.merge);\n validateOptionalArrayElements(\n methodName,\n 'mergeFields',\n 'a string or a FieldPath',\n options.mergeFields,\n element =>\n typeof element === 'string' || element instanceof ExternalFieldPath\n );\n\n if (options.mergeFields !== undefined && options.merge !== undefined) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n `Invalid options passed to function ${methodName}(): You cannot specify both \"merge\" ` +\n `and \"mergeFields\".`\n );\n }\n\n return options;\n}\n\nfunction validateSnapshotOptions(\n methodName: string,\n options: firestore.SnapshotOptions | undefined\n): firestore.SnapshotOptions {\n if (options === undefined) {\n return {};\n }\n\n validateOptionNames(methodName, options, ['serverTimestamps']);\n validateNamedOptionalPropertyEquals(\n methodName,\n 'options',\n 'serverTimestamps',\n options.serverTimestamps,\n ['estimate', 'previous', 'none']\n );\n return options;\n}\n\nfunction validateGetOptions(\n methodName: string,\n options: firestore.GetOptions | undefined\n): void {\n validateOptionalArgType(methodName, 'object', 1, options);\n if (options) {\n validateOptionNames(methodName, options, ['source']);\n validateNamedOptionalPropertyEquals(\n methodName,\n 'options',\n 'source',\n options.source,\n ['default', 'server', 'cache']\n );\n }\n}\n\nfunction validateReference<T>(\n methodName: string,\n documentRef: firestore.DocumentReference<T>,\n firestore: Firestore\n): DocumentReference<T> {\n if (!(documentRef instanceof DocumentReference)) {\n throw invalidClassError(methodName, 'DocumentReference', 1, documentRef);\n } else if (documentRef.firestore !== firestore) {\n throw new FirestoreError(\n Code.INVALID_ARGUMENT,\n 'Provided document reference is from a different Firestore instance.'\n );\n } else {\n return documentRef;\n }\n}\n\n/**\n * Calculates the array of firestore.DocumentChange's for a given ViewSnapshot.\n *\n * Exported for testing.\n */\nexport function changesFromSnapshot<T>(\n firestore: Firestore,\n includeMetadataChanges: boolean,\n snapshot: ViewSnapshot,\n converter?: firestore.FirestoreDataConverter<T>\n): Array<firestore.DocumentChange<T>> {\n if (snapshot.oldDocs.isEmpty()) {\n // Special case the first snapshot because index calculation is easy and\n // fast\n let lastDoc: Document;\n let index = 0;\n return snapshot.docChanges.map(change => {\n const doc = new QueryDocumentSnapshot<T>(\n firestore,\n change.doc.key,\n change.doc,\n snapshot.fromCache,\n snapshot.mutatedKeys.has(change.doc.key),\n converter\n );\n debugAssert(\n change.type === ChangeType.Added,\n 'Invalid event type for first snapshot'\n );\n debugAssert(\n !lastDoc || snapshot.query.docComparator(lastDoc, change.doc) < 0,\n 'Got added events in wrong order'\n );\n lastDoc = change.doc;\n return {\n type: 'added' as firestore.DocumentChangeType,\n doc,\n oldIndex: -1,\n newIndex: index++\n };\n });\n } else {\n // A DocumentSet that is updated incrementally as changes are applied to use\n // to lookup the index of a document.\n let indexTracker = snapshot.oldDocs;\n return snapshot.docChanges\n .filter(\n change => includeMetadataChanges || change.type !== ChangeType.Metadata\n )\n .map(change => {\n const doc = new QueryDocumentSnapshot<T>(\n firestore,\n change.doc.key,\n change.doc,\n snapshot.fromCache,\n snapshot.mutatedKeys.has(change.doc.key),\n converter\n );\n let oldIndex = -1;\n let newIndex = -1;\n if (change.type !== ChangeType.Added) {\n oldIndex = indexTracker.indexOf(change.doc.key);\n debugAssert(oldIndex >= 0, 'Index for document not found');\n indexTracker = indexTracker.delete(change.doc.key);\n }\n if (change.type !== ChangeType.Removed) {\n indexTracker = indexTracker.add(change.doc);\n newIndex = indexTracker.indexOf(change.doc.key);\n }\n return { type: resultChangeType(change.type), doc, oldIndex, newIndex };\n });\n }\n}\n\nfunction resultChangeType(type: ChangeType): firestore.DocumentChangeType {\n switch (type) {\n case ChangeType.Added:\n return 'added';\n case ChangeType.Modified:\n case ChangeType.Metadata:\n return 'modified';\n case ChangeType.Removed:\n return 'removed';\n default:\n return fail('Unknown change type: ' + type);\n }\n}\n\n/**\n * Converts custom model object of type T into DocumentData by applying the\n * converter if it exists.\n *\n * This function is used when converting user objects to DocumentData\n * because we want to provide the user with a more specific error message if\n * their set() or fails due to invalid data originating from a toFirestore()\n * call.\n */\nfunction applyFirestoreDataConverter<T>(\n converter: firestore.FirestoreDataConverter<T> | undefined,\n value: T,\n functionName: string\n): [firestore.DocumentData, string] {\n let convertedValue;\n if (converter) {\n convertedValue = converter.toFirestore(value);\n functionName = 'toFirestore() in ' + functionName;\n } else {\n convertedValue = value as firestore.DocumentData;\n }\n return [convertedValue, functionName];\n}\n\nfunction contains(obj: object, key: string): obj is { key: unknown } {\n return Object.prototype.hasOwnProperty.call(obj, key);\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { Code, FirestoreError } from './error';\n\n/**\n * Helper function to prevent instantiation through the constructor.\n *\n * This method creates a new constructor that throws when it's invoked.\n * The prototype of that constructor is then set to the prototype of the hidden\n * \"class\" to expose all the prototype methods and allow for instanceof\n * checks.\n *\n * To also make all the static methods available, all properties of the\n * original constructor are copied to the new constructor.\n */\nexport function makeConstructorPrivate<T extends Function>(\n cls: T,\n optionalMessage?: string\n): T {\n function PublicConstructor(): never {\n let error = 'This constructor is private.';\n if (optionalMessage) {\n error += ' ';\n error += optionalMessage;\n }\n throw new FirestoreError(Code.INVALID_ARGUMENT, error);\n }\n\n // Make sure instanceof checks work and all methods are exposed on the public\n // constructor\n PublicConstructor.prototype = cls.prototype;\n\n // Copy any static methods/members\n Object.assign(PublicConstructor, cls);\n\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n return PublicConstructor as any;\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { FirebaseApp, FirebaseNamespace } from '@firebase/app-types';\nimport { FirebaseAuthInternalName } from '@firebase/auth-interop-types';\nimport { _FirebaseNamespace } from '@firebase/app-types/private';\nimport { Component, ComponentType, Provider } from '@firebase/component';\nimport {\n CACHE_SIZE_UNLIMITED,\n Firestore,\n DocumentReference,\n DocumentSnapshot,\n QueryDocumentSnapshot,\n Query,\n QuerySnapshot,\n CollectionReference,\n Transaction,\n WriteBatch\n} from '../api/database';\nimport { Blob } from '../api/blob';\nimport { FieldPath } from '../api/field_path';\nimport { GeoPoint } from '../api/geo_point';\nimport { Timestamp } from '../api/timestamp';\nimport { makeConstructorPrivate } from '../util/api';\nimport { FieldValue } from '../api/field_value';\n\n// Public instance that disallows construction at runtime. Note that this still\n// allows instanceof checks.\nexport const PublicFirestore = makeConstructorPrivate(\n Firestore,\n 'Use firebase.firestore() instead.'\n);\nexport const PublicTransaction = makeConstructorPrivate(\n Transaction,\n 'Use firebase.firestore().runTransaction() instead.'\n);\nexport const PublicWriteBatch = makeConstructorPrivate(\n WriteBatch,\n 'Use firebase.firestore().batch() instead.'\n);\nexport const PublicDocumentReference = makeConstructorPrivate(\n DocumentReference,\n 'Use firebase.firestore().doc() instead.'\n);\nexport const PublicDocumentSnapshot = makeConstructorPrivate(DocumentSnapshot);\nexport const PublicQueryDocumentSnapshot = makeConstructorPrivate(\n QueryDocumentSnapshot\n);\nexport const PublicQuery = makeConstructorPrivate(Query);\nexport const PublicQuerySnapshot = makeConstructorPrivate(QuerySnapshot);\nexport const PublicCollectionReference = makeConstructorPrivate(\n CollectionReference,\n 'Use firebase.firestore().collection() instead.'\n);\nexport const PublicFieldValue = makeConstructorPrivate(\n FieldValue,\n 'Use FieldValue.<field>() instead.'\n);\nexport const PublicBlob = makeConstructorPrivate(\n Blob,\n 'Use Blob.fromUint8Array() or Blob.fromBase64String() instead.'\n);\n\nconst firestoreNamespace = {\n Firestore: PublicFirestore,\n GeoPoint,\n Timestamp,\n Blob: PublicBlob,\n Transaction: PublicTransaction,\n WriteBatch: PublicWriteBatch,\n DocumentReference: PublicDocumentReference,\n DocumentSnapshot: PublicDocumentSnapshot,\n Query: PublicQuery,\n QueryDocumentSnapshot: PublicQueryDocumentSnapshot,\n QuerySnapshot: PublicQuerySnapshot,\n CollectionReference: PublicCollectionReference,\n FieldPath,\n FieldValue: PublicFieldValue,\n setLogLevel: Firestore.setLogLevel,\n CACHE_SIZE_UNLIMITED\n};\n\n/**\n * Configures Firestore as part of the Firebase SDK by calling registerService.\n *\n * @param firebase The FirebaseNamespace to register Firestore with\n * @param firestoreFactory A factory function that returns a new Firestore\n * instance.\n */\nexport function configureForFirebase(\n firebase: FirebaseNamespace,\n firestoreFactory: (\n app: FirebaseApp,\n auth: Provider<FirebaseAuthInternalName>\n ) => Firestore\n): void {\n (firebase as _FirebaseNamespace).INTERNAL.registerComponent(\n new Component(\n 'firestore',\n container => {\n const app = container.getProvider('app').getImmediate()!;\n return firestoreFactory(app, container.getProvider('auth-internal'));\n },\n ComponentType.PUBLIC\n ).setServiceProps({ ...firestoreNamespace })\n );\n}\n","/**\n * @license\n * Copyright 2019 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { ConnectivityMonitor, NetworkStatus } from './connectivity_monitor';\n\nexport class NoopConnectivityMonitor implements ConnectivityMonitor {\n addCallback(callback: (status: NetworkStatus) => void): void {\n // No-op.\n }\n\n shutdown(): void {\n // No-op.\n }\n}\n","/**\n * @license\n * Copyright 2019 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { logDebug } from '../util/log';\nimport {\n ConnectivityMonitor,\n ConnectivityMonitorCallback,\n NetworkStatus\n} from './../remote/connectivity_monitor';\n\n// References to `window` are guarded by BrowserConnectivityMonitor.isAvailable()\n/* eslint-disable no-restricted-globals */\n\nconst LOG_TAG = 'ConnectivityMonitor';\n\n/**\n * Browser implementation of ConnectivityMonitor.\n */\nexport class BrowserConnectivityMonitor implements ConnectivityMonitor {\n private readonly networkAvailableListener = (): void =>\n this.onNetworkAvailable();\n private readonly networkUnavailableListener = (): void =>\n this.onNetworkUnavailable();\n private callbacks: ConnectivityMonitorCallback[] = [];\n\n constructor() {\n this.configureNetworkMonitoring();\n }\n\n addCallback(callback: (status: NetworkStatus) => void): void {\n this.callbacks.push(callback);\n }\n\n shutdown(): void {\n window.removeEventListener('online', this.networkAvailableListener);\n window.removeEventListener('offline', this.networkUnavailableListener);\n }\n\n private configureNetworkMonitoring(): void {\n window.addEventListener('online', this.networkAvailableListener);\n window.addEventListener('offline', this.networkUnavailableListener);\n }\n\n private onNetworkAvailable(): void {\n logDebug(LOG_TAG, 'Network connectivity changed: AVAILABLE');\n for (const callback of this.callbacks) {\n callback(NetworkStatus.AVAILABLE);\n }\n }\n\n private onNetworkUnavailable(): void {\n logDebug(LOG_TAG, 'Network connectivity changed: UNAVAILABLE');\n for (const callback of this.callbacks) {\n callback(NetworkStatus.UNAVAILABLE);\n }\n }\n\n // TODO(chenbrian): Consider passing in window either into this component or\n // here for testing via FakeWindow.\n /** Checks that all used attributes of window are available. */\n static isAvailable(): boolean {\n return (\n typeof window !== 'undefined' &&\n window.addEventListener !== undefined &&\n window.removeEventListener !== undefined\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { debugAssert } from '../util/assert';\nimport { FirestoreError } from '../util/error';\n\nimport { Stream } from './connection';\n\n/**\n * Provides a simple helper class that implements the Stream interface to\n * bridge to other implementations that are streams but do not implement the\n * interface. The stream callbacks are invoked with the callOn... methods.\n */\nexport class StreamBridge<I, O> implements Stream<I, O> {\n private wrappedOnOpen: (() => void) | undefined;\n private wrappedOnClose: ((err?: FirestoreError) => void) | undefined;\n private wrappedOnMessage: ((msg: O) => void) | undefined;\n\n private sendFn: (msg: I) => void;\n private closeFn: () => void;\n\n constructor(args: { sendFn: (msg: I) => void; closeFn: () => void }) {\n this.sendFn = args.sendFn;\n this.closeFn = args.closeFn;\n }\n\n onOpen(callback: () => void): void {\n debugAssert(!this.wrappedOnOpen, 'Called onOpen on stream twice!');\n this.wrappedOnOpen = callback;\n }\n\n onClose(callback: (err?: FirestoreError) => void): void {\n debugAssert(!this.wrappedOnClose, 'Called onClose on stream twice!');\n this.wrappedOnClose = callback;\n }\n\n onMessage(callback: (msg: O) => void): void {\n debugAssert(!this.wrappedOnMessage, 'Called onMessage on stream twice!');\n this.wrappedOnMessage = callback;\n }\n\n close(): void {\n this.closeFn();\n }\n\n send(msg: I): void {\n this.sendFn(msg);\n }\n\n callOnOpen(): void {\n debugAssert(\n this.wrappedOnOpen !== undefined,\n 'Cannot call onOpen because no callback was set'\n );\n this.wrappedOnOpen();\n }\n\n callOnClose(err?: FirestoreError): void {\n debugAssert(\n this.wrappedOnClose !== undefined,\n 'Cannot call onClose because no callback was set'\n );\n this.wrappedOnClose(err);\n }\n\n callOnMessage(msg: O): void {\n debugAssert(\n this.wrappedOnMessage !== undefined,\n 'Cannot call onMessage because no callback was set'\n );\n this.wrappedOnMessage(msg);\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport {\n createWebChannelTransport,\n ErrorCode,\n EventType,\n WebChannel,\n WebChannelError,\n WebChannelOptions,\n XhrIo\n} from '@firebase/webchannel-wrapper';\n\nimport {\n isBrowserExtension,\n isElectron,\n isIE,\n isMobileCordova,\n isReactNative,\n isUWP\n} from '@firebase/util';\n\nimport { Token } from '../api/credentials';\nimport { DatabaseId, DatabaseInfo } from '../core/database_info';\nimport { SDK_VERSION } from '../core/version';\nimport { Connection, Stream } from '../remote/connection';\nimport {\n mapCodeFromRpcStatus,\n mapCodeFromHttpResponseErrorStatus\n} from '../remote/rpc_error';\nimport { StreamBridge } from '../remote/stream_bridge';\nimport { debugAssert, fail, hardAssert } from '../util/assert';\nimport { Code, FirestoreError } from '../util/error';\nimport { logDebug } from '../util/log';\nimport { Indexable } from '../util/misc';\nimport { Rejecter, Resolver } from '../util/promise';\nimport { StringMap } from '../util/types';\n\nconst LOG_TAG = 'Connection';\n\nconst RPC_STREAM_SERVICE = 'google.firestore.v1.Firestore';\nconst RPC_URL_VERSION = 'v1';\n\n/**\n * Maps RPC names to the corresponding REST endpoint name.\n * Uses Object Literal notation to avoid renaming.\n */\nconst RPC_NAME_REST_MAPPING: { [key: string]: string } = {};\nRPC_NAME_REST_MAPPING['BatchGetDocuments'] = 'batchGet';\nRPC_NAME_REST_MAPPING['Commit'] = 'commit';\n\n// TODO(b/38203344): The SDK_VERSION is set independently from Firebase because\n// we are doing out-of-band releases. Once we release as part of Firebase, we\n// should use the Firebase version instead.\nconst X_GOOG_API_CLIENT_VALUE = 'gl-js/ fire/' + SDK_VERSION;\n\nconst XHR_TIMEOUT_SECS = 15;\n\nexport class WebChannelConnection implements Connection {\n private readonly databaseId: DatabaseId;\n private readonly baseUrl: string;\n private readonly forceLongPolling: boolean;\n\n constructor(info: DatabaseInfo) {\n this.databaseId = info.databaseId;\n const proto = info.ssl ? 'https' : 'http';\n this.baseUrl = proto + '://' + info.host;\n this.forceLongPolling = info.forceLongPolling;\n }\n\n /**\n * Modifies the headers for a request, adding any authorization token if\n * present and any additional headers for the request.\n */\n private modifyHeadersForRequest(\n headers: StringMap,\n token: Token | null\n ): void {\n if (token) {\n for (const header in token.authHeaders) {\n if (token.authHeaders.hasOwnProperty(header)) {\n headers[header] = token.authHeaders[header];\n }\n }\n }\n headers['X-Goog-Api-Client'] = X_GOOG_API_CLIENT_VALUE;\n }\n\n invokeRPC<Req, Resp>(\n rpcName: string,\n request: Req,\n token: Token | null\n ): Promise<Resp> {\n const url = this.makeUrl(rpcName);\n\n return new Promise((resolve: Resolver<Resp>, reject: Rejecter) => {\n const xhr = new XhrIo();\n xhr.listenOnce(EventType.COMPLETE, () => {\n try {\n switch (xhr.getLastErrorCode()) {\n case ErrorCode.NO_ERROR:\n const json = xhr.getResponseJson() as Resp;\n logDebug(LOG_TAG, 'XHR received:', JSON.stringify(json));\n resolve(json);\n break;\n case ErrorCode.TIMEOUT:\n logDebug(LOG_TAG, 'RPC \"' + rpcName + '\" timed out');\n reject(\n new FirestoreError(Code.DEADLINE_EXCEEDED, 'Request time out')\n );\n break;\n case ErrorCode.HTTP_ERROR:\n const status = xhr.getStatus();\n logDebug(\n LOG_TAG,\n 'RPC \"' + rpcName + '\" failed with status:',\n status,\n 'response text:',\n xhr.getResponseText()\n );\n if (status > 0) {\n const responseError = (xhr.getResponseJson() as WebChannelError)\n .error;\n if (\n !!responseError &&\n !!responseError.status &&\n !!responseError.message\n ) {\n const firestoreErrorCode = mapCodeFromHttpResponseErrorStatus(\n responseError.status\n );\n reject(\n new FirestoreError(\n firestoreErrorCode,\n responseError.message\n )\n );\n } else {\n reject(\n new FirestoreError(\n Code.UNKNOWN,\n 'Server responded with status ' + xhr.getStatus()\n )\n );\n }\n } else {\n // If we received an HTTP_ERROR but there's no status code,\n // it's most probably a connection issue\n logDebug(LOG_TAG, 'RPC \"' + rpcName + '\" failed');\n reject(\n new FirestoreError(Code.UNAVAILABLE, 'Connection failed.')\n );\n }\n break;\n default:\n fail(\n 'RPC \"' +\n rpcName +\n '\" failed with unanticipated ' +\n 'webchannel error ' +\n xhr.getLastErrorCode() +\n ': ' +\n xhr.getLastError() +\n ', giving up.'\n );\n }\n } finally {\n logDebug(LOG_TAG, 'RPC \"' + rpcName + '\" completed.');\n }\n });\n\n // The database field is already encoded in URL. Specifying it again in\n // the body is not necessary in production, and will cause duplicate field\n // errors in the Firestore Emulator. Let's remove it.\n const jsonObj = ({ ...request } as unknown) as Indexable;\n delete jsonObj.database;\n\n const requestString = JSON.stringify(jsonObj);\n logDebug(LOG_TAG, 'XHR sending: ', url + ' ' + requestString);\n // Content-Type: text/plain will avoid preflight requests which might\n // mess with CORS and redirects by proxies. If we add custom headers\n // we will need to change this code to potentially use the\n // $httpOverwrite parameter supported by ESF to avoid\n // triggering preflight requests.\n const headers: StringMap = { 'Content-Type': 'text/plain' };\n\n this.modifyHeadersForRequest(headers, token);\n\n xhr.send(url, 'POST', requestString, headers, XHR_TIMEOUT_SECS);\n });\n }\n\n invokeStreamingRPC<Req, Resp>(\n rpcName: string,\n request: Req,\n token: Token | null\n ): Promise<Resp[]> {\n // The REST API automatically aggregates all of the streamed results, so we\n // can just use the normal invoke() method.\n return this.invokeRPC<Req, Resp[]>(rpcName, request, token);\n }\n\n openStream<Req, Resp>(\n rpcName: string,\n token: Token | null\n ): Stream<Req, Resp> {\n const urlParts = [\n this.baseUrl,\n '/',\n RPC_STREAM_SERVICE,\n '/',\n rpcName,\n '/channel'\n ];\n const webchannelTransport = createWebChannelTransport();\n const request: WebChannelOptions = {\n // Required for backend stickiness, routing behavior is based on this\n // parameter.\n httpSessionIdParam: 'gsessionid',\n initMessageHeaders: {},\n messageUrlParams: {\n // This param is used to improve routing and project isolation by the\n // backend and must be included in every request.\n database: `projects/${this.databaseId.projectId}/databases/${this.databaseId.database}`\n },\n sendRawJson: true,\n supportsCrossDomainXhr: true,\n internalChannelParams: {\n // Override the default timeout (randomized between 10-20 seconds) since\n // a large write batch on a slow internet connection may take a long\n // time to send to the backend. Rather than have WebChannel impose a\n // tight timeout which could lead to infinite timeouts and retries, we\n // set it very large (5-10 minutes) and rely on the browser's builtin\n // timeouts to kick in if the request isn't working.\n forwardChannelRequestTimeoutMs: 10 * 60 * 1000\n },\n forceLongPolling: this.forceLongPolling\n };\n\n this.modifyHeadersForRequest(request.initMessageHeaders!, token);\n\n // Sending the custom headers we just added to request.initMessageHeaders\n // (Authorization, etc.) will trigger the browser to make a CORS preflight\n // request because the XHR will no longer meet the criteria for a \"simple\"\n // CORS request:\n // https://developer.mozilla.org/en-US/docs/Web/HTTP/CORS#Simple_requests\n //\n // Therefore to avoid the CORS preflight request (an extra network\n // roundtrip), we use the httpHeadersOverwriteParam option to specify that\n // the headers should instead be encoded into a special \"$httpHeaders\" query\n // parameter, which is recognized by the webchannel backend. This is\n // formally defined here:\n // https://github.com/google/closure-library/blob/b0e1815b13fb92a46d7c9b3c30de5d6a396a3245/closure/goog/net/rpc/httpcors.js#L32\n //\n // TODO(b/145624756): There is a backend bug where $httpHeaders isn't respected if the request\n // doesn't have an Origin header. So we have to exclude a few browser environments that are\n // known to (sometimes) not include an Origin. See\n // https://github.com/firebase/firebase-js-sdk/issues/1491.\n if (\n !isMobileCordova() &&\n !isReactNative() &&\n !isElectron() &&\n !isIE() &&\n !isUWP() &&\n !isBrowserExtension()\n ) {\n request.httpHeadersOverwriteParam = '$httpHeaders';\n }\n\n const url = urlParts.join('');\n logDebug(LOG_TAG, 'Creating WebChannel: ' + url + ' ' + request);\n const channel = webchannelTransport.createWebChannel(url, request);\n\n // WebChannel supports sending the first message with the handshake - saving\n // a network round trip. However, it will have to call send in the same\n // JS event loop as open. In order to enforce this, we delay actually\n // opening the WebChannel until send is called. Whether we have called\n // open is tracked with this variable.\n let opened = false;\n\n // A flag to determine whether the stream was closed (by us or through an\n // error/close event) to avoid delivering multiple close events or sending\n // on a closed stream\n let closed = false;\n\n const streamBridge = new StreamBridge<Req, Resp>({\n sendFn: (msg: Req) => {\n if (!closed) {\n if (!opened) {\n logDebug(LOG_TAG, 'Opening WebChannel transport.');\n channel.open();\n opened = true;\n }\n logDebug(LOG_TAG, 'WebChannel sending:', msg);\n channel.send(msg);\n } else {\n logDebug(LOG_TAG, 'Not sending because WebChannel is closed:', msg);\n }\n },\n closeFn: () => channel.close()\n });\n\n // Closure events are guarded and exceptions are swallowed, so catch any\n // exception and rethrow using a setTimeout so they become visible again.\n // Note that eventually this function could go away if we are confident\n // enough the code is exception free.\n const unguardedEventListen = <T>(\n type: string,\n fn: (param?: T) => void\n ): void => {\n // TODO(dimond): closure typing seems broken because WebChannel does\n // not implement goog.events.Listenable\n channel.listen(type, (param: unknown) => {\n try {\n fn(param as T);\n } catch (e) {\n setTimeout(() => {\n throw e;\n }, 0);\n }\n });\n };\n\n unguardedEventListen(WebChannel.EventType.OPEN, () => {\n if (!closed) {\n logDebug(LOG_TAG, 'WebChannel transport opened.');\n }\n });\n\n unguardedEventListen(WebChannel.EventType.CLOSE, () => {\n if (!closed) {\n closed = true;\n logDebug(LOG_TAG, 'WebChannel transport closed');\n streamBridge.callOnClose();\n }\n });\n\n unguardedEventListen<Error>(WebChannel.EventType.ERROR, err => {\n if (!closed) {\n closed = true;\n logDebug(LOG_TAG, 'WebChannel transport errored:', err);\n streamBridge.callOnClose(\n new FirestoreError(\n Code.UNAVAILABLE,\n 'The operation could not be completed'\n )\n );\n }\n });\n\n // WebChannel delivers message events as array. If batching is not enabled\n // (it's off by default) each message will be delivered alone, resulting in\n // a single element array.\n interface WebChannelResponse {\n data: Resp[];\n }\n\n unguardedEventListen<WebChannelResponse>(\n WebChannel.EventType.MESSAGE,\n msg => {\n if (!closed) {\n const msgData = msg!.data[0];\n hardAssert(!!msgData, 'Got a webchannel message without data.');\n // TODO(b/35143891): There is a bug in One Platform that caused errors\n // (and only errors) to be wrapped in an extra array. To be forward\n // compatible with the bug we need to check either condition. The latter\n // can be removed once the fix has been rolled out.\n // Use any because msgData.error is not typed.\n const msgDataOrError: WebChannelError | object = msgData;\n const error =\n msgDataOrError.error ||\n (msgDataOrError as WebChannelError[])[0]?.error;\n if (error) {\n logDebug(LOG_TAG, 'WebChannel received error:', error);\n // error.status will be a string like 'OK' or 'NOT_FOUND'.\n const status: string = error.status;\n let code = mapCodeFromRpcStatus(status);\n let message = error.message;\n if (code === undefined) {\n code = Code.INTERNAL;\n message =\n 'Unknown error status: ' +\n status +\n ' with message ' +\n error.message;\n }\n // Mark closed so no further events are propagated\n closed = true;\n streamBridge.callOnClose(new FirestoreError(code, message));\n channel.close();\n } else {\n logDebug(LOG_TAG, 'WebChannel received:', msgData);\n streamBridge.callOnMessage(msgData);\n }\n }\n }\n );\n\n setTimeout(() => {\n // Technically we could/should wait for the WebChannel opened event,\n // but because we want to send the first message with the WebChannel\n // handshake we pretend the channel opened here (asynchronously), and\n // then delay the actual open until the first message is sent.\n streamBridge.callOnOpen();\n }, 0);\n return streamBridge;\n }\n\n // visible for testing\n makeUrl(rpcName: string): string {\n const urlRpcName = RPC_NAME_REST_MAPPING[rpcName];\n debugAssert(\n urlRpcName !== undefined,\n 'Unknown REST mapping for: ' + rpcName\n );\n return (\n this.baseUrl +\n '/' +\n RPC_URL_VERSION +\n '/projects/' +\n this.databaseId.projectId +\n '/databases/' +\n this.databaseId.database +\n '/documents:' +\n urlRpcName\n );\n }\n}\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { PlatformSupport } from '../platform/platform';\nimport { BrowserPlatform } from './browser_platform';\n\n/**\n * This code needs to run before Firestore is used. This can be achieved in\n * several ways:\n * 1) Through the JSCompiler compiling this code and then (automatically)\n * executing it before exporting the Firestore symbols.\n * 2) Through importing this module first in a Firestore main module\n */\nPlatformSupport.setPlatform(new BrowserPlatform());\n","/**\n * @license\n * Copyright 2017 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport { DatabaseId, DatabaseInfo } from '../core/database_info';\nimport { Platform } from '../platform/platform';\nimport { Connection } from '../remote/connection';\nimport { JsonProtoSerializer } from '../remote/serializer';\nimport { ConnectivityMonitor } from './../remote/connectivity_monitor';\n\nimport { NoopConnectivityMonitor } from '../remote/connectivity_monitor_noop';\nimport { BrowserConnectivityMonitor } from './browser_connectivity_monitor';\nimport { WebChannelConnection } from './webchannel_connection';\nimport { debugAssert } from '../util/assert';\n\n// Implements the Platform API for browsers and some browser-like environments\n// (including ReactNative).\nexport class BrowserPlatform implements Platform {\n readonly base64Available: boolean;\n\n constructor() {\n this.base64Available = typeof atob !== 'undefined';\n }\n\n get document(): Document | null {\n // `document` is not always available, e.g. in ReactNative and WebWorkers.\n // eslint-disable-next-line no-restricted-globals\n return typeof document !== 'undefined' ? document : null;\n }\n\n get window(): Window | null {\n // `window` is not always available, e.g. in ReactNative and WebWorkers.\n // eslint-disable-next-line no-restricted-globals\n return typeof window !== 'undefined' ? window : null;\n }\n\n loadConnection(databaseInfo: DatabaseInfo): Promise<Connection> {\n return Promise.resolve(new WebChannelConnection(databaseInfo));\n }\n\n newConnectivityMonitor(): ConnectivityMonitor {\n if (BrowserConnectivityMonitor.isAvailable()) {\n return new BrowserConnectivityMonitor();\n } else {\n return new NoopConnectivityMonitor();\n }\n }\n\n newSerializer(databaseId: DatabaseId): JsonProtoSerializer {\n return new JsonProtoSerializer(databaseId, { useProto3Json: true });\n }\n\n formatJSON(value: unknown): string {\n return JSON.stringify(value);\n }\n\n atob(encoded: string): string {\n return atob(encoded);\n }\n\n btoa(raw: string): string {\n return btoa(raw);\n }\n\n randomBytes(nBytes: number): Uint8Array {\n debugAssert(nBytes >= 0, `Expecting non-negative nBytes, got: ${nBytes}`);\n\n // Polyfills for IE and WebWorker by using `self` and `msCrypto` when `crypto` is not available.\n const crypto =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n typeof self !== 'undefined' && (self.crypto || (self as any)['msCrypto']);\n const bytes = new Uint8Array(nBytes);\n if (crypto) {\n crypto.getRandomValues(bytes);\n } else {\n // Falls back to Math.random\n for (let i = 0; i < nBytes; i++) {\n bytes[i] = Math.floor(Math.random() * 256);\n }\n }\n return bytes;\n }\n}\n","/**\n * @license\n * Copyright 2020 Google LLC\n *\n * Licensed under the Apache License, Version 2.0 (the \"License\");\n * you may not use this file except in compliance with the License.\n * You may obtain a copy of the License at\n *\n * http://www.apache.org/licenses/LICENSE-2.0\n *\n * Unless required by applicable law or agreed to in writing, software\n * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\n\nimport firebase from '@firebase/app';\nimport { FirebaseNamespace } from '@firebase/app-types';\n\nimport { Firestore } from './src/api/database';\nimport { MemoryComponentProvider } from './src/core/component_provider';\nimport { configureForFirebase } from './src/platform/config';\n\nimport './register-module';\nimport './src/platform_browser/browser_init';\n\nimport { name, version } from './package.json';\n\n/**\n * Registers the memory-only Firestore build with the components framework.\n */\nexport function registerFirestore(instance: FirebaseNamespace): void {\n configureForFirebase(\n instance,\n (app, auth) => new Firestore(app, auth, new MemoryComponentProvider())\n );\n instance.registerVersion(name, version);\n}\n\nregisterFirestore(firebase);\n"],"names":["SDK_VERSION","firebase","User","[object Object]","uid","this","__PRIVATE_isAuthenticated","__PRIVATE_otherUser","Code","OK","CANCELLED","UNKNOWN","INVALID_ARGUMENT","DEADLINE_EXCEEDED","NOT_FOUND","ALREADY_EXISTS","PERMISSION_DENIED","UNAUTHENTICATED","RESOURCE_EXHAUSTED","FAILED_PRECONDITION","ABORTED","OUT_OF_RANGE","UNIMPLEMENTED","INTERNAL","UNAVAILABLE","DATA_LOSS","FirestoreError","Error","code","message","super","toString","name","__PRIVATE_OAuthToken","value","user","__PRIVATE_authHeaders","__PRIVATE_EmptyCredentialsProvider","Promise","resolve","__PRIVATE_changeListener","__PRIVATE_FirebaseCredentialsProvider","__PRIVATE_authProvider","__PRIVATE_tokenListener","__PRIVATE_tokenCounter","currentUser","__PRIVATE_getUser","__PRIVATE_receivedInitialUser","auth","getImmediate","optional","addAuthTokenListener","get","then","__PRIVATE_initialTokenCounter","forceRefresh","getToken","__PRIVATE_tokenData","__PRIVATE_hardAssert","accessToken","removeAuthTokenListener","__PRIVATE_currentUid","getUid","__PRIVATE_FirstPartyToken","__PRIVATE_gapi","__PRIVATE_sessionIndex","__PRIVATE_FIRST_PARTY","o","headers","X-Goog-AuthUser","__PRIVATE_authHeader","__PRIVATE_getAuthHeaderValueForFirstParty","__PRIVATE_FirstPartyCredentialsProvider","Timestamp","seconds","nanoseconds","fromMillis","Date","now","date","getTime","milliseconds","Math","floor","toMillis","other","__PRIVATE_primitiveComparator","__PRIVATE_adjustedSeconds","String","padStart","__PRIVATE_SnapshotVersion","timestamp","__PRIVATE__compareTo","isEqual","__PRIVATE_BasePath","segments","offset","length","undefined","fail","__PRIVATE_len","__PRIVATE_comparator","__PRIVATE_nameOrPath","slice","limit","forEach","__PRIVATE_segment","push","__PRIVATE_construct","size","index","__PRIVATE_i","__PRIVATE_potentialChild","fn","end","p1","p2","min","left","right","ResourcePath","__PRIVATE_toArray","join","__PRIVATE_canonicalString","path","indexOf","split","filter","__PRIVATE_identifierRegExp","FieldPath","test","map","str","replace","__PRIVATE_isValidIdentifier","__PRIVATE_current","__PRIVATE_addCurrentSegment","__PRIVATE_inBackticks","c","next","__PRIVATE_DocumentKey","__PRIVATE_fromString","__PRIVATE_popFirst","collectionId","k1","k2","__PRIVATE_objectSize","obj","count","key","Object","prototype","hasOwnProperty","call","__PRIVATE_isEmpty","__PRIVATE_ByteString","__PRIVATE_binaryString","base64","__PRIVATE_PlatformSupport","__PRIVATE_getPlatform","atob","array","fromCharCode","__PRIVATE_binaryStringFromUint8Array","btoa","buffer","Uint8Array","charCodeAt","__PRIVATE_uint8ArrayFromBinaryString","__PRIVATE_isNullOrUndefined","__PRIVATE_isNegativeZero","__PRIVATE_isServerTimestamp","mapValue","fields","stringValue","__PRIVATE_getLocalWriteTime","__PRIVATE_localWriteTime","__PRIVATE_normalizeTimestamp","nanos","__PRIVATE_ISO_TIMESTAMP_REG_EXP","RegExp","__PRIVATE_typeOrder","__PRIVATE_valueEquals","__PRIVATE_leftType","booleanValue","timestampValue","__PRIVATE_leftTimestamp","__PRIVATE_rightTimestamp","__PRIVATE_timestampEquals","__PRIVATE_normalizeByteString","__PRIVATE_blobEquals","referenceValue","__PRIVATE_normalizeNumber","geoPointValue","latitude","longitude","__PRIVATE_geoPointEquals","integerValue","__PRIVATE_n1","__PRIVATE_n2","isNaN","__PRIVATE_numberEquals","__PRIVATE_arrayEquals","arrayValue","values","__PRIVATE_leftMap","__PRIVATE_rightMap","__PRIVATE_objectEquals","__PRIVATE_arrayValueContains","__PRIVATE_haystack","__PRIVATE_needle","find","v","__PRIVATE_valueCompare","__PRIVATE_rightType","__PRIVATE_leftNumber","doubleValue","__PRIVATE_rightNumber","__PRIVATE_compareNumbers","__PRIVATE_compareTimestamps","__PRIVATE_leftBytes","__PRIVATE_rightBytes","__PRIVATE_compareTo","__PRIVATE_compareBlobs","__PRIVATE_leftPath","__PRIVATE_rightPath","__PRIVATE_leftSegments","__PRIVATE_rightSegments","__PRIVATE_comparison","__PRIVATE_compareReferences","__PRIVATE_compareGeoPoints","__PRIVATE_leftArray","__PRIVATE_rightArray","compare","__PRIVATE_compareArrays","__PRIVATE_leftKeys","keys","__PRIVATE_rightKeys","sort","__PRIVATE_keyCompare","__PRIVATE_compareMaps","canonicalId","__PRIVATE_canonifyValue","__PRIVATE_normalizedTimestamp","__PRIVATE_canonifyTimestamp","toBase64","__PRIVATE_fromName","__PRIVATE_geoPoint","result","first","__PRIVATE_canonifyArray","__PRIVATE_sortedKeys","__PRIVATE_canonifyMap","__PRIVATE_fraction","exec","__PRIVATE_nanoStr","substr","Number","__PRIVATE_parsedDate","blob","fromBase64String","fromUint8Array","__PRIVATE_refValue","__PRIVATE_databaseId","projectId","database","isInteger","isArray","__PRIVATE_isNullValue","__PRIVATE_isNanValue","__PRIVATE_isMapValue","__PRIVATE_ServerTimestampTransform","previousValue","__type__","__local_write_time__","serverTimestamp","__PRIVATE_transformResult","__PRIVATE_ArrayUnionTransformOperation","elements","apply","__PRIVATE_coercedFieldValuesArray","__PRIVATE_toUnion","some","element","__PRIVATE_ArrayRemoveTransformOperation","__PRIVATE_toRemove","__PRIVATE_NumericIncrementTransformOperation","serializer","__PRIVATE_operand","__PRIVATE_baseValue","__PRIVATE_computeBaseValue","__PRIVATE_sum","asNumber","__PRIVATE_toInteger","__PRIVATE_toDouble","__PRIVATE_isDouble","__PRIVATE_FieldMask","fieldPath","__PRIVATE_fieldMaskPath","__PRIVATE_isPrefixOf","__PRIVATE_l","r","FieldTransform","field","transform","__PRIVATE_MutationResult","version","transformResults","Precondition","updateTime","exists","wt","__PRIVATE_maybeDoc","Document","__PRIVATE_Mutation","__PRIVATE_SetMutation","__PRIVATE_precondition","__PRIVATE_mutationResult","__PRIVATE_verifyKeyMatches","hasCommittedMutations","__PRIVATE_baseDoc","__PRIVATE_isValidFor","__PRIVATE_getPostMutationVersion","At","__PRIVATE_PatchMutation","data","__PRIVATE_fieldMask","__PRIVATE_UnknownDocument","__PRIVATE_newData","__PRIVATE_patchDocument","__PRIVATE_ObjectValue","empty","__PRIVATE_patchObject","__PRIVATE_builder","__PRIVATE_ObjectValueBuilder","newValue","set","delete","__PRIVATE_build","__PRIVATE_TransformMutation","fieldTransforms","doc","__PRIVATE_requireDocument","__PRIVATE_serverTransformResults","__PRIVATE_transformObject","__PRIVATE_localTransformResults","__PRIVATE_baseObject","__PRIVATE_fieldTransform","__PRIVATE_existingValue","__PRIVATE_coercedValue","__PRIVATE_applyToRemoteDocument","__PRIVATE_applyToLocalView","__PRIVATE_DeleteMutation","__PRIVATE_NoDocument","__PRIVATE_VerifyMutation","proto","__PRIVATE_lastSegment","Map","__PRIVATE_setOverlay","__PRIVATE_currentLevel","__PRIVATE_overlayMap","__PRIVATE_currentSegment","currentValue","entries","__PRIVATE_mergedResult","__PRIVATE_applyOverlay","__PRIVATE_EMPTY_PATH","__PRIVATE_currentPath","__PRIVATE_currentOverlays","__PRIVATE_modified","__PRIVATE_resultAtPath","__PRIVATE_pathSegment","__PRIVATE_nested","child","__PRIVATE_extractFieldMask","__PRIVATE_nestedFields","__PRIVATE_nestedPath","__PRIVATE_MaybeDocument","__PRIVATE_objectValue","options","__PRIVATE_hasLocalMutations","hasPendingWrites","Target","collectionGroup","orderBy","filters","startAt","endAt","__PRIVATE_memoizedCanonicalId","f","__PRIVATE_isDocumentKey","Query","__PRIVATE_explicitOrderBy","__PRIVATE_limitType","__PRIVATE_assertValidBound","__PRIVATE_memoizedOrderBy","__PRIVATE_inequalityField","__PRIVATE_getInequalityFilterField","__PRIVATE_firstOrderByField","__PRIVATE_getFirstOrderByField","__PRIVATE_isKeyField","__PRIVATE_OrderBy","__PRIVATE_keyField","__PRIVATE_foundKeyOrdering","__PRIVATE_lastDirection","dir","__PRIVATE_newFilters","concat","__PRIVATE_newOrderBy","bound","__PRIVATE_toTarget","__PRIVATE_d1","__PRIVATE_d2","__PRIVATE_comparedOnKeyField","__PRIVATE_comp","__PRIVATE_matchesPathAndCollectionGroup","__PRIVATE_matchesOrderBy","__PRIVATE_matchesFilters","__PRIVATE_matchesBounds","FieldFilter","__PRIVATE_isInequality","__PRIVATE_operators","op","__PRIVATE_isDocumentQuery","__PRIVATE_memoizedTarget","__PRIVATE_orderBys","__PRIVATE_Bound","position","before","__PRIVATE_docPath","__PRIVATE_hasCollectionId","__PRIVATE_isImmediateParentOf","matches","__PRIVATE_sortsBeforeDocument","__PRIVATE_KeyFieldInFilter","__PRIVATE_KeyFieldFilter","__PRIVATE_ArrayContainsFilter","__PRIVATE_InFilter","__PRIVATE_ArrayContainsAnyFilter","__PRIVATE_matchesComparison","val","p","__PRIVATE_orderByComponent","component","__PRIVATE_isKeyOrderBy","v1","v2","__PRIVATE_compareDocumentsByField","__PRIVATE_TargetData","target","targetId","__PRIVATE_purpose","sequenceNumber","__PRIVATE_snapshotVersion","lastLimboFreeSnapshotVersion","resumeToken","__PRIVATE_EMPTY_BYTE_STRING","ExistenceFilter","__PRIVATE_RpcCode","__PRIVATE_isPermanentError","__PRIVATE_mapCodeFromRpcCode","__PRIVATE_logError","RpcCode","__PRIVATE_SortedMap","root","__PRIVATE_LLRBNode","EMPTY","__PRIVATE_insert","__PRIVATE_copy","__PRIVATE_BLACK","remove","node","cmp","__PRIVATE_prunedNodes","__PRIVATE_minKey","__PRIVATE_maxKey","action","__PRIVATE_inorderTraversal","k","__PRIVATE_descriptions","__PRIVATE_reverseTraversal","__PRIVATE_SortedMapIterator","__PRIVATE_startKey","__PRIVATE_isReverse","__PRIVATE_nodeStack","pop","color","RED","n","__PRIVATE_fixUp","__PRIVATE_isRed","__PRIVATE_moveRedLeft","__PRIVATE_removeMin","__PRIVATE_smallest","__PRIVATE_rotateRight","__PRIVATE_moveRedRight","__PRIVATE_rotateLeft","__PRIVATE_colorFlip","__PRIVATE_nl","__PRIVATE_nr","__PRIVATE_blackDepth","__PRIVATE_check","pow","__PRIVATE_SortedSet","__PRIVATE_elem","cb","range","__PRIVATE_iter","__PRIVATE_getIteratorFrom","__PRIVATE_hasNext","__PRIVATE_getNext","start","__PRIVATE_getIterator","__PRIVATE_SortedSetIterator","has","add","__PRIVATE_thisIt","__PRIVATE_otherIt","__PRIVATE_thisElem","__PRIVATE_otherElem","__PRIVATE_res","__PRIVATE_EMPTY_MAYBE_DOCUMENT_MAP","__PRIVATE_maybeDocumentMap","__PRIVATE_nullableMaybeDocumentMap","__PRIVATE_EMPTY_DOCUMENT_MAP","__PRIVATE_documentMap","__PRIVATE_EMPTY_DOCUMENT_VERSION_MAP","__PRIVATE_documentVersionMap","__PRIVATE_EMPTY_DOCUMENT_KEY_SET","__PRIVATE_documentKeySet","__PRIVATE_EMPTY_TARGET_ID_SET","__PRIVATE_targetIdSet","__PRIVATE_DocumentSet","__PRIVATE_keyedMap","__PRIVATE_sortedSet","__PRIVATE_oldSet","__PRIVATE_thisDoc","__PRIVATE_otherDoc","__PRIVATE_docStrings","__PRIVATE_newSet","__PRIVATE_DocumentChangeSet","__PRIVATE_change","__PRIVATE_oldChange","__PRIVATE_changeMap","type","__PRIVATE_changes","__PRIVATE_ViewSnapshot","query","docs","__PRIVATE_oldDocs","docChanges","__PRIVATE_mutatedKeys","fromCache","__PRIVATE_syncStateChanged","__PRIVATE_excludesMetadataChanges","documents","__PRIVATE_emptySet","__PRIVATE_otherChanges","__PRIVATE_RemoteEvent","__PRIVATE_targetChanges","__PRIVATE_targetMismatches","__PRIVATE_documentUpdates","__PRIVATE_resolvedLimboDocuments","TargetChange","__PRIVATE_createSynthesizedTargetChangeForCurrentChange","__PRIVATE_addedDocuments","__PRIVATE_modifiedDocuments","__PRIVATE_removedDocuments","__PRIVATE_DocumentWatchChange","__PRIVATE_updatedTargetIds","removedTargetIds","__PRIVATE_newDoc","__PRIVATE_ExistenceFilterChange","__PRIVATE_existenceFilter","__PRIVATE_WatchTargetChange","state","targetIds","cause","__PRIVATE_TargetState","__PRIVATE_snapshotChangesMap","ds","__PRIVATE__current","__PRIVATE__resumeToken","gs","__PRIVATE_pendingResponses","bs","__PRIVATE__hasPendingChanges","__PRIVATE_approximateByteSize","__PRIVATE_documentChanges","__PRIVATE_changeType","__PRIVATE_WatchChangeAggregator","__PRIVATE_metadataProvider","__PRIVATE_documentTargetMap","__PRIVATE_docChange","__PRIVATE_addDocumentToTarget","__PRIVATE_removeDocumentFromTarget","targetChange","__PRIVATE_forEachTarget","__PRIVATE_targetState","__PRIVATE_ensureTargetState","__PRIVATE_isActiveTarget","__PRIVATE_updateResumeToken","__PRIVATE_recordTargetResponse","__PRIVATE_isPending","__PRIVATE_clearPendingChanges","removeTarget","__PRIVATE_markCurrent","__PRIVATE_resetTarget","__PRIVATE_targetStates","__PRIVATE__","__PRIVATE_watchChange","__PRIVATE_expectedCount","__PRIVATE_targetData","__PRIVATE_targetDataForActiveTarget","__PRIVATE_getCurrentDocumentCountForTarget","__PRIVATE_pendingTargetResets","__PRIVATE_pendingDocumentUpdates","__PRIVATE_targetContainsDocument","__PRIVATE_hasPendingChanges","__PRIVATE_toTargetChange","__PRIVATE_pendingDocumentTargetMapping","__PRIVATE_targets","__PRIVATE_isOnlyLimboTarget","__PRIVATE_forEachWhile","__PRIVATE_remoteEvent","document","__PRIVATE_addDocumentChange","__PRIVATE_ensureDocumentTargetMapping","__PRIVATE_updatedDocument","__PRIVATE_removeDocumentChange","__PRIVATE_getRemoteKeysForTarget","__PRIVATE_recordPendingTargetRequest","__PRIVATE_targetMapping","__PRIVATE_targetActive","__PRIVATE_logDebug","__PRIVATE_getTargetDataForTarget","__PRIVATE_DIRECTIONS","__PRIVATE_dirs","asc","desc","__PRIVATE_OPERATORS","__PRIVATE_ops","<","<=",">",">=","==","array-contains","in","array-contains-any","__PRIVATE_JsonProtoSerializer","status","__PRIVATE_useProto3Json","Infinity","MAX_SAFE_INTEGER","MIN_SAFE_INTEGER","isSafeInteger","toISOString","bytes","toUint8Array","__PRIVATE_toTimestamp","__PRIVATE_fromTimestamp","__PRIVATE_fullyQualifiedPrefixPath","__PRIVATE_resource","__PRIVATE_isValidResourceName","__PRIVATE_toResourceName","__PRIVATE_fromResourceName","__PRIVATE_extractLocalPathFromResourceName","__PRIVATE_resourceName","mi","__PRIVATE_toName","__PRIVATE_toProto","fromVersion","found","missing","readTime","__PRIVATE_fromFound","__PRIVATE_fromMissing","__PRIVATE_fromWatchTargetChangeState","targetChangeType","__PRIVATE_fromBytes","__PRIVATE_causeProto","__PRIVATE_fromRpcStatus","documentChange","__PRIVATE_entityChange","documentDelete","__PRIVATE_docDelete","documentRemove","__PRIVATE_docRemove","__PRIVATE_mutation","update","__PRIVATE_toMutationDocument","updateMask","__PRIVATE_toDocumentMask","__PRIVATE_toFieldTransform","verify","__PRIVATE_isNone","currentDocument","__PRIVATE_toPrecondition","__PRIVATE_fromPrecondition","__PRIVATE_none","__PRIVATE_fromDocumentMask","__PRIVATE_fromFieldTransform","toVersion","commitTime","__PRIVATE_protos","__PRIVATE_fromWriteResult","setToServerValue","appendMissingElements","removeAllFromArray","increment","instance","__PRIVATE_fromServerFormat","__PRIVATE_toQueryPath","__PRIVATE_documentsTarget","__PRIVATE_atPath","__PRIVATE_fromQueryPath","structuredQuery","parent","from","allDescendants","__PRIVATE_popLast","where","__PRIVATE_toFilter","__PRIVATE_toOrder","__PRIVATE_toInt32Proto","__PRIVATE_toCursor","__PRIVATE_fromCount","__PRIVATE_filterBy","__PRIVATE_fromFilter","__PRIVATE_fromOrder","__PRIVATE_fromInt32Proto","__PRIVATE_fromCursor","__PRIVATE_toLabel","goog-listen-tags","__PRIVATE_toDocumentsTarget","__PRIVATE_toQueryTarget","__PRIVATE_toBytes","__PRIVATE_toUnaryOrFieldFilter","compositeFilter","unaryFilter","__PRIVATE_fromUnaryFilter","fieldFilter","__PRIVATE_fromFieldFilter","reduce","__PRIVATE_accum","order","__PRIVATE_toPropertyOrder","__PRIVATE_fromPropertyOrder","cursor","__PRIVATE_fieldReference","__PRIVATE_toFieldPathReference","direction","__PRIVATE_toDirection","__PRIVATE_fromFieldPathReference","__PRIVATE_fromDirection","create","__PRIVATE_fromOperatorName","__PRIVATE_toOperatorName","__PRIVATE_nanField","NaN","__PRIVATE_nullField","nullValue","__PRIVATE_canonicalFields","fieldPaths","__PRIVATE_paths","platform","__PRIVATE_logClient","Logger","__PRIVATE_getLogLevel","logLevel","setLogLevel","__PRIVATE_newLevel","msg","LogLevel","DEBUG","args","__PRIVATE_argToString","debug","ERROR","error","__PRIVATE_formatJSON","e","__PRIVATE_failure","assertion","__PRIVATE_debugCast","constructor","__PRIVATE_AutoId","__PRIVATE_chars","__PRIVATE_maxMultiple","__PRIVATE_autoId","__PRIVATE_randomBytes","charAt","every","__PRIVATE_DatabaseInfo","persistenceKey","host","ssl","forceLongPolling","__PRIVATE_DatabaseId","_n","__PRIVATE_ObjectMap","__PRIVATE_mapKeyFn","id","__PRIVATE_inner","__PRIVATE_otherKey","splice","__PRIVATE_MutationBatch","batchId","baseMutations","mutations","__PRIVATE_docKey","__PRIVATE_batchResult","__PRIVATE_mutationResults","__PRIVATE_maybeDocs","__PRIVATE_mutatedDocuments","__PRIVATE_m","__PRIVATE_mutatedDocument","__PRIVATE_MutationBatchResult","batch","__PRIVATE_commitVersion","streamToken","__PRIVATE_docVersions","results","__PRIVATE_versionMap","PersistencePromise","callback","__PRIVATE_isDone","__PRIVATE_nextCallback","__PRIVATE_catchCallback","__PRIVATE_nextFn","__PRIVATE_catchFn","__PRIVATE_callbackAttached","__PRIVATE_wrapFailure","__PRIVATE_wrapSuccess","reject","__PRIVATE_wrapUserFunction","all","__PRIVATE_resolvedCount","done","__PRIVATE_err","__PRIVATE_predicates","predicate","__PRIVATE_isTrue","collection","__PRIVATE_promises","s","__PRIVATE_waitFor","__PRIVATE_LocalDocumentsView","__PRIVATE_remoteDocumentCache","__PRIVATE_mutationQueue","__PRIVATE_indexManager","transaction","__PRIVATE_getAllMutationBatchesAffectingDocumentKey","__PRIVATE_batches","__PRIVATE_getDocumentInternal","__PRIVATE_inBatches","__PRIVATE_getEntry","__PRIVATE_localView","getEntries","__PRIVATE_getLocalViewOfDocuments","__PRIVATE_baseDocs","__PRIVATE_getAllMutationBatchesAffectingDocumentKeys","__PRIVATE_applyLocalMutationsToDocuments","__PRIVATE_sinceReadTime","__PRIVATE_getDocumentsMatchingDocumentQuery","__PRIVATE_isCollectionGroupQuery","__PRIVATE_getDocumentsMatchingCollectionGroupQuery","__PRIVATE_getDocumentsMatchingCollectionQuery","__PRIVATE_getDocument","__PRIVATE_getCollectionParents","__PRIVATE_parents","__PRIVATE_collectionQuery","__PRIVATE_asCollectionQueryAtPath","__PRIVATE_mutationBatches","__PRIVATE_getDocumentsMatchingQuery","__PRIVATE_queryResults","__PRIVATE_getAllMutationBatchesAffectingQuery","__PRIVATE_matchingMutationBatches","__PRIVATE_addMissingBaseDocuments","__PRIVATE_mergedDocuments","__PRIVATE_mutatedDoc","__PRIVATE_existingDocuments","__PRIVATE_missingBaseDocEntriesForPatching","__PRIVATE_missingBaseDocs","__PRIVATE_LocalViewChanges","__PRIVATE_addedKeys","__PRIVATE_removedKeys","__PRIVATE_viewSnapshot","__PRIVATE_ListenSequence","__PRIVATE_sequenceNumberSyncer","__PRIVATE_sequenceNumberHandler","__PRIVATE_setPreviousValue","__PRIVATE_writeNewSequenceNumber","__PRIVATE_writeSequenceNumber","__PRIVATE_externalPreviousValue","max","__PRIVATE_nextValue","__PRIVATE_Deferred","promise","__PRIVATE_ExponentialBackoff","__PRIVATE_queue","__PRIVATE_timerId","__PRIVATE_initialDelayMs","__PRIVATE_backoffFactor","__PRIVATE_maxDelayMs","reset","__PRIVATE_currentBaseMs","cancel","__PRIVATE_desiredDelayWithJitterMs","__PRIVATE_jitterDelayMs","__PRIVATE_delaySoFarMs","__PRIVATE_lastAttemptTime","__PRIVATE_remainingDelayMs","__PRIVATE_timerPromise","__PRIVATE_enqueueAfterDelay","__PRIVATE_skipDelay","random","__PRIVATE_MemoryIndexManager","__PRIVATE_MemoryCollectionParentIndex","collectionPath","__PRIVATE_collectionParentIndex","parentPath","__PRIVATE_existingParents","__PRIVATE_added","__PRIVATE_TargetIdGenerator","__PRIVATE_lastId","__PRIVATE_isIndexedDbTransactionError","__PRIVATE_DelayedOperation","__PRIVATE_asyncQueue","__PRIVATE_targetTimeMs","__PRIVATE_removalCallback","__PRIVATE_deferred","bind","catch","__PRIVATE_delayMs","__PRIVATE_targetTime","__PRIVATE_delayedOp","__PRIVATE_timerHandle","setTimeout","__PRIVATE_handleDelayElapsed","reason","clearTimeout","__PRIVATE_enqueueAndForget","__PRIVATE_AsyncQueue","__PRIVATE_backoff","__PRIVATE_skipBackoff","window","addEventListener","__PRIVATE_visibilityHandler","Mr","__PRIVATE__isShuttingDown","enqueue","__PRIVATE_verifyNotFailed","__PRIVATE_enqueueInternal","removeEventListener","__PRIVATE_enqueueEvenAfterShutdown","__PRIVATE_retryableTail","__PRIVATE_retryingOp","async","__PRIVATE_backoffAndRun","__PRIVATE_newTail","__PRIVATE_tail","__PRIVATE_operationInProgress","stack","__PRIVATE_timerIdsToSkip","__PRIVATE_createAndSchedule","__PRIVATE_removedOp","__PRIVATE_removeDelayedOperation","__PRIVATE_delayedOperations","__PRIVATE_currentTail","__PRIVATE_lastTimerId","__PRIVATE_drain","a","b","__PRIVATE_wrapInUserErrorIfRecoverable","__PRIVATE_LruParams","__PRIVATE_cacheSizeCollectionThreshold","__PRIVATE_percentileToCollect","__PRIVATE_maximumSequenceNumbersToCollect","__PRIVATE_cacheSize","__PRIVATE_DEFAULT_COLLECTION_PERCENTILE","__PRIVATE_DEFAULT_MAX_SEQUENCE_NUMBERS_TO_COLLECT","__PRIVATE_DEFAULT_CACHE_SIZE_BYTES","__PRIVATE_COLLECTION_DISABLED","__PRIVATE_LocalStore","persistence","__PRIVATE_queryEngine","__PRIVATE_initialUser","t","__PRIVATE_getMutationQueue","__PRIVATE_remoteDocuments","__PRIVATE_getRemoteDocumentCache","__PRIVATE_targetCache","__PRIVATE_getTargetCache","__PRIVATE_localDocuments","__PRIVATE_getIndexManager","__PRIVATE_setLocalDocumentsView","__PRIVATE_newMutationQueue","__PRIVATE_newLocalDocuments","runTransaction","txn","__PRIVATE_oldBatches","__PRIVATE_getAllMutationBatches","__PRIVATE_promisedOldBatches","__PRIVATE_newBatches","__PRIVATE_removedBatchIds","__PRIVATE_addedBatchIds","__PRIVATE_changedKeys","__PRIVATE_getDocuments","__PRIVATE_affectedDocuments","Vh","ph","yh","__PRIVATE_existingDocs","__PRIVATE_extractBaseValue","__PRIVATE_addMutationBatch","__PRIVATE_applyToLocalDocumentSet","vh","__PRIVATE_affected","__PRIVATE_documentBuffer","__PRIVATE_newChangeBuffer","Dh","__PRIVATE_acknowledgeBatch","__PRIVATE_applyWriteToRemoteDocuments","__PRIVATE_performConsistencyCheck","__PRIVATE_affectedKeys","__PRIVATE_lookupMutationBatch","__PRIVATE_removeMutationBatch","__PRIVATE_getHighestUnacknowledgedBatchId","__PRIVATE_getLastStreamToken","__PRIVATE_setLastStreamToken","__PRIVATE_getLastRemoteSnapshotVersion","__PRIVATE_remoteVersion","__PRIVATE_newTargetDataByTargetMap","__PRIVATE_targetDataByTarget","__PRIVATE_oldTargetData","__PRIVATE_removeMatchingKeys","__PRIVATE_addMatchingKeys","__PRIVATE_newTargetData","__PRIVATE_withResumeToken","__PRIVATE_withSequenceNumber","__PRIVATE_currentSequenceNumber","__PRIVATE_shouldPersistTargetData","__PRIVATE_updateTargetData","__PRIVATE_changedDocs","__PRIVATE_updatedKeys","__PRIVATE_existingDoc","__PRIVATE_removeEntry","__PRIVATE_addEntry","__PRIVATE_referenceDelegate","__PRIVATE_updateLimboDocument","__PRIVATE_updateRemoteVersion","lastRemoteSnapshotVersion","__PRIVATE_setTargetsMetadata","__PRIVATE_toMicroseconds","__PRIVATE_RESUME_TOKEN_MAX_AGE_MICROS","__PRIVATE_viewChanges","__PRIVATE_viewChange","__PRIVATE_addReference","__PRIVATE_removeReference","__PRIVATE_updatedTargetData","__PRIVATE_withLastLimboFreeSnapshotVersion","__PRIVATE_afterBatchId","__PRIVATE_getNextMutationBatchAfterBatchId","__PRIVATE_getTargetData","__PRIVATE_cached","__PRIVATE_allocateTargetId","__PRIVATE_addTargetData","__PRIVATE_targetIdByTarget","__PRIVATE_keepPersistedTargetData","mode","__PRIVATE_usePreviousResults","__PRIVATE_remoteKeys","__PRIVATE_getMatchingKeysForTargetId","_o","__PRIVATE_docKeys","__PRIVATE_promiseChain","__PRIVATE_remoteDoc","__PRIVATE_ackVersion","__PRIVATE_garbageCollector","__PRIVATE_collect","__PRIVATE_ignoreIfPrimaryLeaseLoss","__PRIVATE_ReferenceSet","__PRIVATE_DocReference","__PRIVATE_compareByKey","__PRIVATE_compareByTargetId","__PRIVATE_refsByKey","ref","__PRIVATE_refsByTarget","__PRIVATE_removeRef","__PRIVATE_emptyKey","__PRIVATE_startRef","__PRIVATE_endRef","__PRIVATE_forEachInRange","__PRIVATE_firstRef","__PRIVATE_firstAfterOrEqual","__PRIVATE_targetOrBatchId","__PRIVATE_validateNoArgs","functionName","__PRIVATE_formatPlural","__PRIVATE_validateExactNumberOfArgs","__PRIVATE_numberOfArgs","__PRIVATE_validateAtLeastNumberOfArgs","__PRIVATE_minNumberOfArgs","__PRIVATE_validateBetweenNumberOfArgs","__PRIVATE_maxNumberOfArgs","__PRIVATE_validateArgType","__PRIVATE_argument","__PRIVATE_validateType","__PRIVATE_ordinal","__PRIVATE_validateOptionalArgType","__PRIVATE_validateNamedType","__PRIVATE_optionName","__PRIVATE_validateNamedOptionalType","__PRIVATE_validateOptionalArrayElements","__PRIVATE_typeDescription","__PRIVATE_validator","Array","__PRIVATE_valueDescription","__PRIVATE_validateArrayElements","__PRIVATE_validateNamedOptionalPropertyEquals","__PRIVATE_inputName","input","__PRIVATE_expected","__PRIVATE_expectedDescription","__PRIVATE_actualDescription","__PRIVATE_validateNamedPropertyEquals","valid","__PRIVATE_isPlainObject","description","getPrototypeOf","substring","JSON","stringify","__PRIVATE_customObjectName","__PRIVATE_tryGetCustomObjectType","__PRIVATE_validateDefined","__PRIVATE_validateOptionNames","__PRIVATE_optionNames","__PRIVATE_invalidClassError","__PRIVATE_validatePositiveNumber","num","__PRIVATE_assertUint8ArrayAvailable","__PRIVATE_assertBase64Available","__PRIVATE_base64Available","Blob","__PRIVATE_byteString","__PRIVATE__byteString","arguments","fieldNames","__PRIVATE_minNumberOfElements","__PRIVATE_validateNamedArrayAtLeastNumberOfElements","__PRIVATE__internalPath","__PRIVATE_InternalFieldPath","__PRIVATE__DOCUMENT_ID","__PRIVATE_RESERVED","__PRIVATE_FieldValueImpl","__PRIVATE__methodName","__PRIVATE_DeleteFieldValueImpl","context","__PRIVATE_dataSource","__PRIVATE_createError","__PRIVATE_ServerTimestampFieldValueImpl","__PRIVATE_ArrayUnionFieldValueImpl","__PRIVATE__elements","__PRIVATE_parseContext","__PRIVATE_ParseContext","No","methodName","Lo","ignoreUndefinedProperties","__PRIVATE_parsedElements","__PRIVATE_parseData","arrayUnion","__PRIVATE_ArrayRemoveFieldValueImpl","__PRIVATE_NumericIncrementFieldValueImpl","__PRIVATE__operand","__PRIVATE_numericIncrement","GeoPoint","isFinite","__PRIVATE__lat","__PRIVATE__long","__PRIVATE_RESERVED_FIELD_REGEX","__PRIVATE_ParsedSetData","__PRIVATE_ParsedUpdateData","__PRIVATE_isWrite","settings","__PRIVATE_validatePath","configuration","__PRIVATE_childPath","__PRIVATE_contextWith","__PRIVATE_validatePathSegment","__PRIVATE_fieldDescription","__PRIVATE_UserDataReader","__PRIVATE_newSerializer","__PRIVATE_createContext","__PRIVATE_validatePlainObject","__PRIVATE_updateData","__PRIVATE_parseObject","__PRIVATE_validatedFieldPaths","__PRIVATE_stringOrFieldPath","__PRIVATE_ExternalFieldPath","__PRIVATE_fieldPathFromDotSeparatedString","contains","__PRIVATE_fieldMaskContains","__PRIVATE_covers","__PRIVATE_fieldMaskPaths","__PRIVATE_childContext","__PRIVATE_childContextForFieldPath","__PRIVATE_parsedValue","mask","moreFieldsAndValues","__PRIVATE_fieldPathFromArgument","__PRIVATE_allowArrays","__PRIVATE_looksLikeJsonObject","__PRIVATE_parseSentinelFieldValue","__PRIVATE_arrayElement","__PRIVATE_entryIndex","__PRIVATE_entry","__PRIVATE_parsedEntry","__PRIVATE_childContextForArray","__PRIVATE_parseArray","__PRIVATE_toNumber","fromDate","bytesValue","DocumentReference","__PRIVATE_thisDb","__PRIVATE_otherDb","firestore","__PRIVATE__databaseId","__PRIVATE__key","__PRIVATE_parseScalarValue","__PRIVATE_childContextForField","search","__PRIVATE_fromDotSeparatedString","__PRIVATE_PersistentStream","__PRIVATE_connectionTimerId","__PRIVATE_idleTimerId","__PRIVATE_connection","__PRIVATE_credentialsProvider","listener","__PRIVATE_performBackoff","__PRIVATE_isStarted","close","__PRIVATE_isOpen","__PRIVATE_idleTimer","__PRIVATE_handleIdleCloseTimer","__PRIVATE_cancelIdleCheck","stream","send","__PRIVATE_finalState","__PRIVATE_closeCount","__PRIVATE_resetToMax","__PRIVATE_invalidateToken","__PRIVATE_tearDown","__PRIVATE_onClose","__PRIVATE_dispatchIfNotClosed","__PRIVATE_getCloseGuardedDispatcher","token","__PRIVATE_startStream","__PRIVATE_rpcError","__PRIVATE_handleStreamClose","__PRIVATE_startRpc","__PRIVATE_onOpen","onMessage","__PRIVATE_startCloseCount","__PRIVATE_PersistentListenStream","credentials","__PRIVATE_openStream","__PRIVATE_watchChangeProto","__PRIVATE_fromWatchChange","snapshot","__PRIVATE_versionFromListenResponse","__PRIVATE_onWatchChange","request","__PRIVATE_encodedDatabaseId","addTarget","labels","__PRIVATE_toListenRequestLabels","__PRIVATE_sendRequest","__PRIVATE_PersistentWriteStream","ba","__PRIVATE_handshakeComplete_","__PRIVATE_writeMutations","__PRIVATE_responseProto","lastStreamToken","__PRIVATE_fromWriteResults","writeResults","__PRIVATE_onMutationResult","__PRIVATE_onHandshakeComplete","writes","__PRIVATE_toMutation","__PRIVATE_DatastoreImpl","__PRIVATE_rpcName","__PRIVATE_invokeRPC","__PRIVATE_invokeStreamingRPC","Transaction","__PRIVATE_datastore","Set","__PRIVATE_ensureCommitNotCalled","__PRIVATE_datastoreImpl","__PRIVATE_params","response","__PRIVATE_fromMaybeDocument","__PRIVATE_invokeBatchGetDocumentsRpc","__PRIVATE_recordVersion","write","__PRIVATE_toMutations","__PRIVATE_writtenDocs","__PRIVATE_preconditionForUpdate","__PRIVATE_lastWriteError","__PRIVATE_unwritten","__PRIVATE_readVersions","__PRIVATE__version","__PRIVATE_invokeCommitRpc","__PRIVATE_committed","__PRIVATE_docVersion","__PRIVATE_existingVersion","__PRIVATE_OnlineStateTracker","__PRIVATE_onlineStateHandler","__PRIVATE_watchStreamFailures","__PRIVATE_setAndBroadcast","__PRIVATE_onlineStateTimer","__PRIVATE_logClientOfflineWarningIfNecessary","__PRIVATE_clearOnlineStateTimer","__PRIVATE_newState","__PRIVATE_shouldWarnClientIsOffline","details","__PRIVATE_RemoteStore","__PRIVATE_localStore","__PRIVATE_connectivityMonitor","__PRIVATE_addCallback","__PRIVATE_canUseNetwork","__PRIVATE_restartNetwork","__PRIVATE_onlineStateTracker","__PRIVATE_watchStream","__PRIVATE_newPersistentWatchStream","Aa","__PRIVATE_onWatchStreamOpen","Ta","__PRIVATE_onWatchStreamClose","Va","__PRIVATE_onWatchStreamChange","__PRIVATE_writeStream","__PRIVATE_newPersistentWriteStream","__PRIVATE_onWriteStreamOpen","__PRIVATE_onWriteStreamClose","Ca","__PRIVATE_onWriteHandshakeComplete","Sa","enableNetwork","networkEnabled","__PRIVATE_enableNetworkInternal","__PRIVATE_shouldStartWatchStream","__PRIVATE_startWatchStream","__PRIVATE_fillWritePipeline","__PRIVATE_disableNetworkInternal","stop","__PRIVATE_writePipeline","__PRIVATE_cleanUpWatchStreamState","__PRIVATE_shutdown","__PRIVATE_listenTargets","__PRIVATE_sendWatchRequest","__PRIVATE_sendUnwatchRequest","__PRIVATE_markIdle","__PRIVATE_syncEngine","__PRIVATE_watchChangeAggregator","__PRIVATE_watch","__PRIVATE_unwatch","__PRIVATE_handleWatchStreamStart","__PRIVATE_indexedDbFailed","isPrimary","__PRIVATE_handleWatchStreamFailure","__PRIVATE_handleTargetError","__PRIVATE_disableNetworkUntilRecovery","__PRIVATE_handleDocumentChange","__PRIVATE_handleExistenceFilter","__PRIVATE_handleTargetChange","__PRIVATE_raiseWatchSnapshot","__PRIVATE_enqueueRetryable","__PRIVATE_createRemoteEvent","__PRIVATE_requestTargetData","__PRIVATE_applyRemoteEvent","__PRIVATE_rejectListen","__PRIVATE_canAddToWritePipeline","__PRIVATE_lastBatchIdRetrieved","__PRIVATE_nextMutationBatch","__PRIVATE_addToWritePipeline","__PRIVATE_shouldStartWriteStream","__PRIVATE_startWriteStream","__PRIVATE_handshakeComplete","__PRIVATE_writeHandshake","shift","__PRIVATE_success","__PRIVATE_applySuccessfulWrite","__PRIVATE_handleWriteError","__PRIVATE_handleHandshakeError","__PRIVATE_inhibitBackoff","__PRIVATE_rejectFailedWrite","__PRIVATE_LocalClientState","activeTargetIds","updateTimeMs","__PRIVATE_MemorySharedClientState","__PRIVATE_localState","__PRIVATE_addQueryTarget","__PRIVATE_queryState","__PRIVATE_removeQueryTarget","onlineState","__PRIVATE_AddedLimboDocument","__PRIVATE_RemovedLimboDocument","__PRIVATE_View","__PRIVATE__syncedDocuments","__PRIVATE_documentSet","__PRIVATE_docComparator","uc","__PRIVATE_previousChanges","__PRIVATE_changeSet","__PRIVATE_oldDocumentSet","__PRIVATE_newMutatedKeys","__PRIVATE_newDocumentSet","__PRIVATE_needsRefill","__PRIVATE_lastDocInLimit","__PRIVATE_hasLimitToFirst","last","__PRIVATE_firstDocInLimit","__PRIVATE_hasLimitToLast","__PRIVATE_newMaybeDoc","__PRIVATE_oldDoc","__PRIVATE_oldDocHadPendingMutations","__PRIVATE_newDocHasPendingMutations","__PRIVATE_changeApplied","track","__PRIVATE_shouldWaitForSyncedDocument","ac","lc","fc","ns","__PRIVATE_updateLimboDocuments","__PRIVATE_getChanges","__PRIVATE_c1","__PRIVATE_c2","__PRIVATE_compareChangeType","__PRIVATE_applyTargetChange","__PRIVATE_limboChanges","__PRIVATE_newSyncState","__PRIVATE_limboDocuments","__PRIVATE_syncState","Ec","__PRIVATE_applyChanges","__PRIVATE_oldLimboDocuments","__PRIVATE_shouldBeInLimbo","__PRIVATE_queryResult","__PRIVATE_computeDocChanges","__PRIVATE_fromInitialDocuments","__PRIVATE_TransactionRunner","__PRIVATE_remoteStore","updateFunction","__PRIVATE_runWithBackOff","__PRIVATE_createTransaction","__PRIVATE_userPromise","__PRIVATE_tryRunUpdateFunction","commit","__PRIVATE_commitError","__PRIVATE_handleTransactionError","__PRIVATE_userPromiseError","__PRIVATE_retries","__PRIVATE_isRetryableTransactionError","__PRIVATE_QueryView","view","__PRIVATE_LimboResolution","__PRIVATE_SyncEngine","__PRIVATE_sharedClientState","__PRIVATE_maxConcurrentLimboResolutions","q","__PRIVATE_forSyncEngine","Uc","__PRIVATE_syncEngineListener","__PRIVATE_assertSubscribed","__PRIVATE_queryView","__PRIVATE_queryViewsByQuery","__PRIVATE_addLocalQueryTarget","__PRIVATE_computeInitialSnapshot","__PRIVATE_allocateTarget","__PRIVATE_initializeViewAndComputeSnapshot","__PRIVATE_isPrimaryClient","listen","__PRIVATE_executeQuery","__PRIVATE_viewDocChanges","__PRIVATE_synthesizedTargetChange","__PRIVATE_updateTrackedLimbos","__PRIVATE_queriesByTarget","__PRIVATE_queries","__PRIVATE_removeLocalQueryTarget","__PRIVATE_isActiveQueryTarget","__PRIVATE_releaseTarget","__PRIVATE_clearQueryState","__PRIVATE_unlisten","__PRIVATE_removeAndCleanupTarget","__PRIVATE_userCallback","__PRIVATE_localWrite","__PRIVATE_addPendingMutation","__PRIVATE_addMutationCallback","__PRIVATE_emitNewSnapsAndNotifyLocalStore","__PRIVATE_run","__PRIVATE_limboResolution","__PRIVATE_activeLimboResolutionsByTarget","__PRIVATE_receivedDocument","source","__PRIVATE_newViewSnapshots","__PRIVATE_applyOnlineStateChange","__PRIVATE_onOnlineStateChange","__PRIVATE_updateQueryState","__PRIVATE_limboKey","event","__PRIVATE_activeLimboTargetsByKey","__PRIVATE_pumpEnqueuedLimboResolutions","__PRIVATE_mutationBatchResult","__PRIVATE_processUserCallback","__PRIVATE_triggerPendingWritesCallbacks","__PRIVATE_updateMutationState","__PRIVATE_rejectBatch","__PRIVATE_highestBatchId","__PRIVATE_callbacks","__PRIVATE_pendingWritesCallbacks","__PRIVATE_firestoreError","__PRIVATE_errorMessage","clear","__PRIVATE_newCallbacks","__PRIVATE_mutationUserCallbacks","__PRIVATE_toKey","__PRIVATE_onWatchError","__PRIVATE_limboDocumentRefs","__PRIVATE_removeReferencesForId","__PRIVATE_containsKey","__PRIVATE_removeLimboTarget","__PRIVATE_limboTargetId","__PRIVATE_limboChange","__PRIVATE_trackLimboChange","__PRIVATE_enqueuedLimboResolutions","__PRIVATE_limboTargetIdGenerator","__PRIVATE_INVALID","__PRIVATE_newSnaps","__PRIVATE_docChangesInAllViews","__PRIVATE_queriesProcessed","__PRIVATE_fromSnapshot","__PRIVATE_notifyLocalViewChanges","__PRIVATE_fnName","__PRIVATE_handleUserChange","__PRIVATE_rejectOutstandingPendingWritesCallbacks","__PRIVATE_handleCredentialChange","disableNetwork","__PRIVATE_keySet","__PRIVATE_unionWith","__PRIVATE_syncedDocuments","__PRIVATE_QueryListenersInfo","__PRIVATE_EventManager","subscribe","__PRIVATE_firstListen","__PRIVATE_queryInfo","__PRIVATE_viewSnap","onError","__PRIVATE_listeners","__PRIVATE_onViewSnapshot","__PRIVATE_raiseSnapshotsInSyncEvent","__PRIVATE_lastListen","__PRIVATE_viewSnaps","__PRIVATE_raisedEvent","observer","__PRIVATE_snapshotsInSyncListeners","__PRIVATE_QueryListener","__PRIVATE_queryObserver","__PRIVATE_snap","includeMetadataChanges","__PRIVATE_raisedInitialEvent","__PRIVATE_shouldRaiseEvent","__PRIVATE_shouldRaiseInitialEvent","__PRIVATE_raiseInitialEvent","__PRIVATE_maybeOnline","__PRIVATE_waitForSyncWhenOnline","__PRIVATE_hasPendingWritesChanged","__PRIVATE_IndexFreeQueryEngine","__PRIVATE_localDocumentsView","__PRIVATE_matchesAllDocuments","__PRIVATE_executeFullCollectionScan","__PRIVATE_previousResults","__PRIVATE_applyQuery","__PRIVATE_updatedResults","__PRIVATE_sortedPreviousResults","__PRIVATE_limboFreeSnapshotVersion","__PRIVATE_docAtLimitEdge","__PRIVATE_MemoryMutationQueue","__PRIVATE_batchIndex","__PRIVATE_indexOfExistingBatchId","__PRIVATE_nextBatchId","__PRIVATE_batchesByDocumentKey","__PRIVATE_addToCollectionParentIndex","__PRIVATE_findMutationBatch","__PRIVATE_rawIndex","__PRIVATE_indexOfBatchId","__PRIVATE_documentKey","POSITIVE_INFINITY","__PRIVATE_documentKeys","__PRIVATE_uniqueBatchIDs","__PRIVATE_findMutationBatches","prefix","__PRIVATE_immediateChildrenPathLength","__PRIVATE_startPath","__PRIVATE_rowKeyPath","__PRIVATE_batchIDs","__PRIVATE_references","__PRIVATE_markPotentiallyOrphaned","__PRIVATE_MemoryRemoteDocumentCache","__PRIVATE_sizer","__PRIVATE_previousSize","__PRIVATE_currentSize","$l","__PRIVATE_maybeDocument","iterator","__PRIVATE_RemoteDocumentChangeBuffer","__PRIVATE__readTime","__PRIVATE_assertNotApplied","__PRIVATE_bufferedEntry","__PRIVATE_getFromCache","__PRIVATE_getAllFromCache","__PRIVATE_changesApplied","__PRIVATE_documentCache","__PRIVATE_MemoryTargetCache","__PRIVATE_forTargetCache","__PRIVATE_highestSequenceNumber","highestTargetId","__PRIVATE_targetIdGenerator","highestListenSequenceNumber","__PRIVATE_saveTargetData","targetCount","upperBound","__PRIVATE_removals","__PRIVATE_removeMatchingKeysForTargetId","__PRIVATE_addReferences","__PRIVATE_removeReferences","__PRIVATE_matchingKeys","__PRIVATE_referencesForId","__PRIVATE_MemoryPersistence","__PRIVATE_referenceDelegateFactory","__PRIVATE__started","__PRIVATE_documentSize","n_","__PRIVATE_mutationQueues","__PRIVATE_transactionOperation","__PRIVATE_MemoryTransaction","__PRIVATE_listenSequence","__PRIVATE_onTransactionStarted","__PRIVATE_onTransactionCommitted","__PRIVATE_toPromise","__PRIVATE_raiseOnCommittedEvent","__PRIVATE_or","__PRIVATE_onCommittedListeners","__PRIVATE_MemoryEagerDelegate","w_","__PRIVATE__orphanedDocuments","__PRIVATE_localViewReferences","__PRIVATE_orphanedDocuments","cache","__PRIVATE_removeTargetData","__PRIVATE_changeBuffer","__PRIVATE_isReferenced","__PRIVATE_mutationQueuesContainKey","__PRIVATE_MemoryComponentProvider","__PRIVATE_cfg","__PRIVATE_createSharedClientState","__PRIVATE_createPersistence","__PRIVATE_gcScheduler","__PRIVATE_createGarbageCollectionScheduler","__PRIVATE_createLocalStore","__PRIVATE_createRemoteStore","__PRIVATE_createSyncEngine","__PRIVATE_eventManager","__PRIVATE_createEventManager","__PRIVATE_applyPrimaryState","__PRIVATE_factory","__PRIVATE_newConnectivityMonitor","__PRIVATE_databaseInfo","__PRIVATE_FirestoreClient","__PRIVATE_newId","__PRIVATE_componentProvider","__PRIVATE_persistenceSettings","__PRIVATE_verifyNotTerminated","__PRIVATE_initializationDone","__PRIVATE_persistenceResult","__PRIVATE_initialized","__PRIVATE_setChangeListener","__PRIVATE_initializeComponents","__PRIVATE_loadConnection","__PRIVATE_newDatastore","initialize","Pr","v_","ka","clientId","g_","Dc","F_","__PRIVATE_eventMgr","__PRIVATE_setDatabaseDeletedListener","terminate","__PRIVATE_canFallback","console","warn","L_","DOMException","__PRIVATE_isShuttingDown","__PRIVATE_verifyOperationInProgress","__PRIVATE_enqueueAndInitiateShutdown","__PRIVATE_removeChangeListener","__PRIVATE_registerPendingWritesCallback","__PRIVATE_clientTerminated","__PRIVATE_readDocument","__PRIVATE_addSnapshotsInSyncListener","__PRIVATE_removeSnapshotsInSyncListener","O_","__PRIVATE_AsyncObserver","__PRIVATE_scheduleEvent","muted","eventHandler","__PRIVATE_isPartialObserver","__PRIVATE_methods","object","method","__PRIVATE_implementsAnyMethods","__PRIVATE_UserDataWriter","timestampsInSnapshots","__PRIVATE_serverTimestampBehavior","converter","__PRIVATE_convertTimestamp","__PRIVATE_convertServerTimestamp","__PRIVATE_convertReference","__PRIVATE_convertGeoPoint","__PRIVATE_convertArray","__PRIVATE_convertObject","__PRIVATE_convertValue","__PRIVATE_getPreviousValue","__PRIVATE_normalizedValue","toDate","__PRIVATE_resourcePath","CACHE_SIZE_UNLIMITED","__PRIVATE_FirestoreSettings","cacheSizeBytes","__PRIVATE_MINIMUM_CACHE_SIZE_BYTES","experimentalForceLongPolling","Firestore","__PRIVATE_databaseIdOrApp","__PRIVATE_ensureClientConfigured","__PRIVATE__firestoreClient","app","__PRIVATE__firebaseApp","__PRIVATE_databaseIdFromApp","__PRIVATE__persistenceKey","__PRIVATE__credentials","external","__PRIVATE__componentProvider","__PRIVATE__settings","rf","__PRIVATE__userDataReader","__PRIVATE_settingsLiteral","__PRIVATE_newSettings","__PRIVATE_client","__PRIVATE_makeCredentialsProvider","synchronizeTabs","experimentalTabSynchronization","__PRIVATE_configureClient","__PRIVATE__queue","__PRIVATE_enqueueAndForgetEvenAfterShutdown","__PRIVATE_makeDatabaseInfo","clearPersistence","_removeServiceInstance","lf","waitForPendingWrites","arg","__PRIVATE_onSnapshotsInSyncInternal","__PRIVATE_asyncObserver","__PRIVATE_mute","__PRIVATE_pathString","CollectionReference","__PRIVATE_forPath","__PRIVATE_InternalQuery","WriteBatch","SILENT","level","__PRIVATE__firestore","__PRIVATE__transaction","documentRef","__PRIVATE_validateReference","__PRIVATE_lookup","DocumentSnapshot","__PRIVATE__converter","__PRIVATE_validateSetOptions","__PRIVATE_convertedValue","__PRIVATE_applyFirestoreDataConverter","__PRIVATE_parsed","merge","mergeFields","__PRIVATE__dataReader","__PRIVATE_parseMergeData","__PRIVATE_parseSetData","__PRIVATE_fieldOrUpdateData","__PRIVATE_parseUpdateVarargs","__PRIVATE_parseUpdateData","__PRIVATE_verifyNotCommitted","__PRIVATE__mutations","__PRIVATE__committed","__PRIVATE_currArg","__PRIVATE_internalOptions","complete","__PRIVATE_onSnapshotInternal","__PRIVATE_errHandler","__PRIVATE_internalListener","__PRIVATE_validateGetOptions","__PRIVATE_getDocumentFromLocalCache","__PRIVATE_getViaSnapshotListener","Rl","metadata","SnapshotMetadata","__PRIVATE__document","__PRIVATE__fromCache","__PRIVATE__hasPendingWrites","__PRIVATE_validateSnapshotOptions","QueryDocumentSnapshot","fromFirestore","__PRIVATE__areTimestampsInSnapshotsEnabled","serverTimestamps","__PRIVATE__query","opStr","__PRIVATE_enums","__PRIVATE_validateStringEnum","__PRIVATE_fieldValue","__PRIVATE_validateDisjunctiveFilterElements","__PRIVATE_referenceList","__PRIVATE_parseDocumentIdValue","__PRIVATE_parseQueryValue","__PRIVATE_validateNewFilter","__PRIVATE_addFilter","directionStr","__PRIVATE_validateNewOrderBy","__PRIVATE_addOrderBy","__PRIVATE_withLimitToFirst","__PRIVATE_withLimitToLast","__PRIVATE_docOrField","__PRIVATE_boundFromDocOrFields","__PRIVATE_withStartAt","__PRIVATE_withEndAt","__PRIVATE_boundFromDocument","__PRIVATE_allFields","__PRIVATE_boundFromFields","components","__PRIVATE_rawValue","__PRIVATE_wrapped","__PRIVATE_validateHasExplicitOrderByForLimitToLast","QuerySnapshot","__PRIVATE_firestoreClient","__PRIVATE_getDocumentsFromLocalCache","__PRIVATE_documentIdValue","operator","__PRIVATE_arrayOps","__PRIVATE_disjunctiveOps","__PRIVATE_isArrayOp","__PRIVATE_isDisjunctiveOp","__PRIVATE_existingField","__PRIVATE_validateOrderByAndInequalityMatch","__PRIVATE_conflictingOp","__PRIVATE_findFilterOperator","__PRIVATE_inequality","__PRIVATE__originalQuery","__PRIVATE__snapshot","thisArg","__PRIVATE_convertToDocumentImpl","__PRIVATE__cachedChanges","__PRIVATE__cachedChangesIncludeMetadataChanges","__PRIVATE_lastDoc","oldIndex","newIndex","__PRIVATE_indexTracker","__PRIVATE_resultChangeType","__PRIVATE_changesFromSnapshot","__PRIVATE__path","toFirestore","__PRIVATE_docRef","__PRIVATE_makeConstructorPrivate","__PRIVATE_cls","__PRIVATE_optionalMessage","__PRIVATE_PublicConstructor","assign","__PRIVATE_PublicFirestore","__PRIVATE_PublicTransaction","__PRIVATE_PublicWriteBatch","__PRIVATE_PublicDocumentReference","__PRIVATE_PublicDocumentSnapshot","__PRIVATE_PublicQueryDocumentSnapshot","__PRIVATE_PublicQuery","__PRIVATE_PublicQuerySnapshot","__PRIVATE_PublicCollectionReference","__PRIVATE_PublicFieldValue","__PRIVATE_PublicBlob","__PRIVATE_firestoreNamespace","FieldValue","__PRIVATE_NoopConnectivityMonitor","__PRIVATE_BrowserConnectivityMonitor","__PRIVATE_onNetworkAvailable","__PRIVATE_onNetworkUnavailable","__PRIVATE_configureNetworkMonitoring","__PRIVATE_networkAvailableListener","__PRIVATE_networkUnavailableListener","__PRIVATE_StreamBridge","__PRIVATE_sendFn","__PRIVATE_closeFn","__PRIVATE_wrappedOnOpen","__PRIVATE_wrappedOnClose","__PRIVATE_wrappedOnMessage","__PRIVATE_RPC_NAME_REST_MAPPING","BatchGetDocuments","Commit","__PRIVATE_X_GOOG_API_CLIENT_VALUE","__PRIVATE_WebChannelConnection","info","__PRIVATE_baseUrl","__PRIVATE_header","url","__PRIVATE_makeUrl","__PRIVATE_xhr","XhrIo","listenOnce","EventType","COMPLETE","getLastErrorCode","ErrorCode","NO_ERROR","json","getResponseJson","TIMEOUT","HTTP_ERROR","getStatus","getResponseText","__PRIVATE_responseError","__PRIVATE_firestoreErrorCode","__PRIVATE_serverError","toLowerCase","__PRIVATE_mapCodeFromHttpResponseErrorStatus","__PRIVATE_jsonObj","__PRIVATE_requestString","Content-Type","__PRIVATE_modifyHeadersForRequest","__PRIVATE_urlParts","__PRIVATE_webchannelTransport","createWebChannelTransport","httpSessionIdParam","initMessageHeaders","messageUrlParams","sendRawJson","supportsCrossDomainXhr","internalChannelParams","forwardChannelRequestTimeoutMs","isMobileCordova","isReactNative","isElectron","isIE","isUWP","isBrowserExtension","httpHeadersOverwriteParam","channel","createWebChannel","__PRIVATE_opened","closed","__PRIVATE_streamBridge","zf","open","Yf","__PRIVATE_unguardedEventListen","param","WebChannel","OPEN","CLOSE","__PRIVATE_callOnClose","MESSAGE","__PRIVATE_msgData","__PRIVATE_msgDataOrError","__PRIVATE_mapCodeFromRpcStatus","__PRIVATE_callOnMessage","__PRIVATE_callOnOpen","__PRIVATE_urlRpcName","__PRIVATE_setPlatform","__PRIVATE_isAvailable","hi","__PRIVATE_encoded","raw","__PRIVATE_nBytes","crypto","self","getRandomValues","__PRIVATE_registerFirestore","__PRIVATE_firestoreFactory","registerComponent","Component","container","getProvider","setServiceProps","__PRIVATE_configureForFirebase","registerVersion"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;sDAoBO,OAAMA,IAAcC,EAASD;;;;;;;;;;;;;;;;;;;;;UCCvBE;IASXC,YAAqBC;QAAAC,WAAAD;;IAErBD;QACE,OAAmB,QAAZE,KAAKD;;;;;WAOdD;QACE,OAAIE,KAAKC,MACA,SAASD,KAAKD,MAEd;;IAIXD,QAAQI;QACN,OAAOA,EAAUH,QAAQC,KAAKD;;;;8BA1BhBF,qBAAkB,IAAIA,EAAK;;;AAI3CA,MAAqC,IAAIA,EAAK,2BAC9CA,MAA8B,IAAIA,EAAK;;;;;;;;;;;;;;;;;;ACHlC,MAAMM,IAAO;;;;IAIlBC,IAAI;;IAGJC,WAAW;;IAGXC,SAAS;;;;;;;IAQTC,kBAAkB;;;;;;;;IASlBC,mBAAmB;;IAGnBC,WAAW;;;;;IAMXC,gBAAgB;;;;;;;;IAShBC,mBAAmB;;;;;IAMnBC,iBAAiB;;;;;IAMjBC,oBAAoB;;;;;;;;;;;;;;;;;;;;;IAsBpBC,qBAAqB;;;;;;;;IASrBC,SAAS;;;;;;;;;;;;;;;;IAiBTC,cAAc;;IAGdC,eAAe;;;;;IAMfC,UAAU;;;;;;;;IASVC,aAAa;;IAGbC,WAAW;;;;;;;;UASAC,UAAuBC;IAIlCxB,YAAqByB,GAAqBC;QACxCC,MAAMD,IADaxB,YAAAuB,GAAqBvB,eAAAwB,GAH1CxB,YAAO;;;;QASLA,KAAK0B,WAAW,MAAM,GAAG1B,KAAK2B,eAAe3B,KAAKuB,UAAUvB,KAAKwB;;;;;;;;;;;;;;;;;;;UC9GxDI;IAGX9B,YAAY+B,GAAsBC;QAAA9B,YAAA8B,GAFlC9B,YAAO,SAGLA,KAAK+B,IAAc;;QAEnB/B,KAAK+B,EAA2B,gBAAI,UAAUF;;;;sEAqCrCG;IAAblC;;;;;;QAMEE,SAA0D;;IAE1DF;QACE,OAAOmC,QAAQC,QAAsB;;IAGvCpC;IAEAA,EAAkBqC;QAKhBnC,KAAKmC,IAAiBA;;QAEtBA,EAAetC,EAAKe;;IAGtBd;QAKEE,KAAKmC,IAAiB;;;;MAIbC;IAwBXtC,YAAYuC;;;;;QAnBZrC,SAAiE;;QAGzDA,mBAAoBH,EAAKe,iBACjCZ,UAAuC;;;;;QAMvCA,SAAuB;;QAGvBA,SAA0D,MAElDA,qBAAe,GAKrBA,KAAKsC,IAAgB;YACnBtC,KAAKuC,KACLvC,KAAKwC,cAAcxC,KAAKyC,KACxBzC,KAAK0C,KAAsB,GACvB1C,KAAKmC,KACPnC,KAAKmC,EAAenC,KAAKwC;WAI7BxC,KAAKuC,IAAe,GAEpBvC,KAAK2C,OAAON,EAAaO,aAAa;YAAEC,WAAU;YAE9C7C,KAAK2C,OACP3C,KAAK2C,KAAKG,qBAAqB9C,KAAmB;;QAGlDA,KAAKsC,EAAc,OACnBD,EAAaU,MAAMC,KACjBL;YACE3C,KAAK2C,OAAOA,GACR3C,KAAKsC;;YAEPtC,KAAK2C,KAAKG,qBAAqB9C,KAAKsC;WAGxC;;IAONxC;;;;QASE,MAAMmD,IAAsBjD,KAAKuC,GAC3BW,IAAelD,KAAKkD;QAG1B,OAFAlD,KAAKkD,gBAAe,GAEflD,KAAK2C,OAIH3C,KAAK2C,KAAKQ,SAASD,GAAcF,KAAKI;;;;YAI3C,IAAIpD,KAAKuC,MAAiBU,GACxB,MAAM,IAAI5B,EACRlB,EAAKY,SACL;YAGF,OAAIqC,KACFC,GACmC,mBAA1BD,EAAUE,cAGZ,IAAI1B,EAAWwB,EAAUE,aAAatD,KAAKwC,gBAE3C;aApBJP,QAAQC,QAAQ;;IA0B3BpC;QACEE,KAAKkD,gBAAe;;IAGtBpD,EAAkBqC;QAKhBnC,KAAKmC,IAAiBA;;QAGlBnC,KAAK0C,KACPP,EAAenC,KAAKwC;;IAIxB1C;QAUME,KAAK2C,QACP3C,KAAK2C,KAAKY,wBAAwBvD,KAAmB,IAEvDA,KAAKsC,IAAgB,MACrBtC,KAAKmC,IAAiB;;;;;;IAOhBrC;QACN,MAAM0D,IAAaxD,KAAK2C,QAAQ3C,KAAK2C,KAAKc;QAK1C,OAJAJ,GACiB,SAAfG,KAA6C,mBAAfA,IAGzB,IAAI3D,EAAK2D;;;;;;;;;;UAoBPE;IAIX5D,YAAoB6D,GAAoBC;iBAApBD,YAAoBC,GAHxC5D,YAAO,cACPA,YAAOH,EAAKgE;;IAIZC;QACE,MAAMC,IAAwC;YAC5CC,mBAAmBhE,KAAK4D;WAEpBK,IAAajE,KAAK2D,EAAKhB,KAAKuB,EAAgC;QAIlE,OAHID,MACFF,EAAuB,gBAAIE,IAEtBF;;;;;;;;UASEI;IACXrE,YAAoB6D,GAAoBC;iBAApBD,YAAoBC;;IAExC9D;QACE,OAAOmC,QAAQC,QAAQ,IAAIwB,EAAgB1D,KAAK2D,GAAM3D,KAAK4D;;IAG7D9D,EAAkBqC;;QAEhBA,EAAetC,EAAKgE;;IAGtB/D;IAEAA;;;;;;UCxTWsE;IAeXtE,YAAqBuE,GAA0BC;QAC7C,IADmBtE,eAAAqE,GAA0BrE,mBAAAsE,GACzCA,IAAc,GAChB,MAAM,IAAIjD,EACRlB,EAAKI,kBACL,yCAAyC+D;QAG7C,IAAIA,KAAe,KACjB,MAAM,IAAIjD,EACRlB,EAAKI,kBACL,yCAAyC+D;QAG7C,IAAID,KA9BY,aA+Bd,MAAM,IAAIhD,EACRlB,EAAKI,kBACL,qCAAqC8D;;gBAIzC,IAAIA,KAAW,cACb,MAAM,IAAIhD,EACRlB,EAAKI,kBACL,qCAAqC8D;;IArC3CvE;QACE,OAAOsE,EAAUG,WAAWC,KAAKC;;IAGnC3E,gBAAgB4E;QACd,OAAON,EAAUG,WAAWG,EAAKC;;IAGnC7E,kBAAkB8E;QAChB,MAAMP,IAAUQ,KAAKC,MAAMF,IAAe;QAE1C,OAAO,IAAIR,EAAUC,GAD2B,OAAjCO,IAAyB,MAAVP;;IAgChCvE;QACE,OAAO,IAAI0E,KAAKxE,KAAK+E;;IAGvBjF;QACE,OAAsB,MAAfE,KAAKqE,UAAiBrE,KAAKsE,cAAc;;IAGlDxE,EAAWkF;QACT,OAAIhF,KAAKqE,YAAYW,EAAMX,UAClBY,GAAoBjF,KAAKsE,aAAaU,EAAMV,eAE9CW,GAAoBjF,KAAKqE,SAASW,EAAMX;;IAGjDvE,QAAQkF;QACN,OACEA,EAAMX,YAAYrE,KAAKqE,WAAWW,EAAMV,gBAAgBtE,KAAKsE;;IAIjExE;QACE,OACE,uBACAE,KAAKqE,UACL,mBACArE,KAAKsE,cACL;;IAIJxE;;;;;;;QAOE,MAAMoF,IAAkBlF,KAAKqE,WAnFb;;gBAuFhB,OAFyBc,OAAOD,GAAiBE,SAAS,IAAI,OAEpC,MADGD,OAAOnF,KAAKsE,aAAac,SAAS,GAAG;;;;;;;;;;;;;;;;;;;;;;;UCpFzDC;IASXvF,YAA4BwF;QAAAtF,iBAAAsF;;IAR5BxF,SAAqB+B;QACnB,OAAO,IAAIwD,EAAgBxD;;IAG7B/B;QACE,OAAO,IAAIuF,EAAgB,IAAIjB,EAAU,GAAG;;IAK9CtE,EAAUkF;QACR,OAAOhF,KAAKsF,UAAUC,EAAWP,EAAMM;;IAGzCxF,QAAQkF;QACN,OAAOhF,KAAKsF,UAAUE,QAAQR,EAAMM;;oFAItCxF;;QAEE,OAAgC,MAAzBE,KAAKsF,UAAUjB,UAAgBrE,KAAKsF,UAAUhB,cAAc;;IAGrExE;QACE,OAAO,qBAAqBE,KAAKsF,UAAU5D,aAAa;;IAG1D5B;QACE,OAAOE,KAAKsF;;;;;;;;;;;;;;;;;;;;;;;AC5BhB,MAAeG;IAKb3F,YAAY4F,GAAoBC,GAAiBC;aAChCC,MAAXF,IACFA,IAAS,IACAA,IAASD,EAASE,UAC3BE,WAGaD,MAAXD,IACFA,IAASF,EAASE,SAASD,IAClBC,IAASF,EAASE,SAASD,KACpCG;QAEF9F,KAAK0F,WAAWA,GAChB1F,KAAK2F,SAASA,GACd3F,KAAK+F,IAAMH;;IAqBbA;QACE,OAAO5F,KAAK+F;;IAGdjG,QAAQkF;QACN,OAA4C,MAArCS,EAASO,EAAWhG,MAAMgF;;IAGnClF,MAAMmG;QACJ,MAAMP,IAAW1F,KAAK0F,SAASQ,MAAMlG,KAAK2F,QAAQ3F,KAAKmG;QAQvD,OAPIF,aAAsBR,IACxBQ,EAAWG,QAAQC;YACjBX,EAASY,KAAKD;aAGhBX,EAASY,KAAKL,IAETjG,KAAKuG,EAAUb;;kEAIhB5F;QACN,OAAOE,KAAK2F,SAAS3F,KAAK4F;;IAG5B9F,EAAS0G;QAMP,OALAA,SAAgBX,MAATW,IAAqB,IAAIA,GAKzBxG,KAAKuG,EACVvG,KAAK0F,UACL1F,KAAK2F,SAASa,GACdxG,KAAK4F,SAASY;;IAIlB1G;QAEE,OAAOE,KAAKuG,EAAUvG,KAAK0F,UAAU1F,KAAK2F,QAAQ3F,KAAK4F,SAAS;;IAGlE9F;QAEE,OAAOE,KAAK0F,SAAS1F,KAAK2F;;IAG5B7F;QACE,OAAOE,KAAK+C,IAAI/C,KAAK4F,SAAS;;IAGhC9F,IAAI2G;QAEF,OAAOzG,KAAK0F,SAAS1F,KAAK2F,SAASc;;IAGrC3G;QACE,OAAuB,MAAhBE,KAAK4F;;IAGd9F,EAAWkF;QACT,IAAIA,EAAMY,SAAS5F,KAAK4F,QACtB,QAAO;QAGT,KAAK,IAAIc,IAAI,GAAGA,IAAI1G,KAAK4F,QAAQc,KAC/B,IAAI1G,KAAK+C,IAAI2D,OAAO1B,EAAMjC,IAAI2D,IAC5B,QAAO;QAIX,QAAO;;IAGT5G,EAAoB6G;QAClB,IAAI3G,KAAK4F,SAAS,MAAMe,EAAef,QACrC,QAAO;QAGT,KAAK,IAAIc,IAAI,GAAGA,IAAI1G,KAAK4F,QAAQc,KAC/B,IAAI1G,KAAK+C,IAAI2D,OAAOC,EAAe5D,IAAI2D,IACrC,QAAO;QAIX,QAAO;;IAGT5G,QAAQ8G;QACN,KAAK,IAAIF,IAAI1G,KAAK2F,QAAQkB,IAAM7G,KAAKmG,SAASO,IAAIG,GAAKH,KACrDE,EAAG5G,KAAK0F,SAASgB;;IAIrB5G;QACE,OAAOE,KAAK0F,SAASQ,MAAMlG,KAAK2F,QAAQ3F,KAAKmG;;IAG/CrG,SACEgH,GACAC;QAEA,MAAMhB,IAAMlB,KAAKmC,IAAIF,EAAGlB,QAAQmB,EAAGnB;QACnC,KAAK,IAAIc,IAAI,GAAGA,IAAIX,GAAKW,KAAK;YAC5B,MAAMO,IAAOH,EAAG/D,IAAI2D,IACdQ,IAAQH,EAAGhE,IAAI2D;YACrB,IAAIO,IAAOC,GACT,QAAQ;YAEV,IAAID,IAAOC,GACT,OAAO;;QAGX,OAAIJ,EAAGlB,SAASmB,EAAGnB,UACT,IAENkB,EAAGlB,SAASmB,EAAGnB,SACV,IAEF;;;;;;;UAQEuB,UAAqB1B;IACtB3F,EACR4F,GACAC,GACAC;QAEA,OAAO,IAAIuB,EAAazB,GAAUC,GAAQC;;IAG5C9F;;;;QAKE,OAAOE,KAAKoH,IAAUC,KAAK;;IAG7BvH;QACE,OAAOE,KAAKsH;;;;WAMdxH,SAAkByH;;;;QAKhB,IAAIA,EAAKC,QAAQ,SAAS,GACxB,MAAM,IAAInG,EACRlB,EAAKI,kBACL,iBAAiBgH;;;gBAMrB,MAAM7B,IAAW6B,EAAKE,MAAM,KAAKC,OAAOrB,KAAWA,EAAQT,SAAS;QAEpE,OAAO,IAAIuB,EAAazB;;;;AAGnByB,MAAa,IAAIA,EAAa;;AAGvC,MAAMQ,IAAmB;;gFAGZC,UAAkBnC;IACnB3F,EACR4F,GACAC,GACAC;QAEA,OAAO,IAAIgC,EAAUlC,GAAUC,GAAQC;;;;;WAOjC9F,SAAyBuG;QAC/B,OAAOsB,EAAiBE,KAAKxB;;IAG/BvG;QACE,OAAOE,KAAKoH,IACTU,IAAIC,MACHA,IAAMA,EAAIC,QAAQ,MAAM,QAAQA,QAAQ,KAAK,QACxCJ,EAAUK,EAAkBF,OAC/BA,IAAM,MAAMA,IAAM;QAEbA,IAERV,KAAK;;IAGVvH;QACE,OAAOE,KAAKsH;;;;WAMdxH;QACE,OAAuB,MAAhBE,KAAK4F,UAnQiB,eAmQD5F,KAAK+C,IAAI;;;;WAMvCjD;QACE,OAAO,IAAI8H,EAAU,EA1QQ;;;;;;;;;;;WAuR/B9H,SAAwByH;QACtB,MAAM7B,IAAqB;QAC3B,IAAIwC,IAAU,IACVxB,IAAI;QAER,MAAMyB,IAAoB;YACxB,IAAuB,MAAnBD,EAAQtC,QACV,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,uBAAuBgH,wCACrB;YAGN7B,EAASY,KAAK4B,IACdA,IAAU;;QAGZ,IAAIE,KAAc;QAElB,MAAO1B,IAAIa,EAAK3B,UAAQ;YACtB,MAAMyC,IAAId,EAAKb;YACf,IAAU,SAAN2B,GAAY;gBACd,IAAI3B,IAAI,MAAMa,EAAK3B,QACjB,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,yCAAyCgH;gBAG7C,MAAMe,IAAOf,EAAKb,IAAI;gBACtB,IAAe,SAAT4B,KAA0B,QAATA,KAAyB,QAATA,GACrC,MAAM,IAAIjH,EACRlB,EAAKI,kBACL,uCAAuCgH;gBAG3CW,KAAWI,GACX5B,KAAK;mBACU,QAAN2B,KACTD,KAAeA,GACf1B,OACe,QAAN2B,KAAcD,KAIvBF,KAAWG,GACX3B,QAJAyB,KACAzB;;QAQJ,IAFAyB,KAEIC,GACF,MAAM,IAAI/G,EACRlB,EAAKI,kBACL,6BAA6BgH;QAIjC,OAAO,IAAIK,EAAUlC;;;;AAGhBkC,MAAa,IAAIA,EAAU;;;;;;;;;;;;;;;;;;MClVvBW;IACXzI,YAAqByH;QAAAvH,YAAAuH;;IAQrBzH,SAAgB6B;QACd,OAAO,IAAI4G,EAAYpB,EAAaqB,EAAW7G,GAAM8G,EAAS;;6EAIhE3I,GAAgB4I;QACd,OACE1I,KAAKuH,KAAK3B,UAAU,KACpB5F,KAAKuH,KAAKxE,IAAI/C,KAAKuH,KAAK3B,SAAS,OAAO8C;;IAI5C5I,QAAQkF;QACN,OACY,SAAVA,KAAqE,MAAnDmC,EAAanB,EAAWhG,KAAKuH,MAAMvC,EAAMuC;;IAI/DzH;QACE,OAAOE,KAAKuH,KAAK7F;;IAKnB5B,SAAkB6I,GAAiBC;QACjC,OAAOzB,EAAanB,EAAW2C,EAAGpB,MAAMqB,EAAGrB;;IAG7CzH,UAAqByH;QACnB,OAAOA,EAAK3B,SAAS,KAAM;;;;;;;WAS7B9F,UAAoB4F;QAClB,OAAO,IAAI6C,EAAY,IAAIpB,EAAazB,EAASQ;;;;;;;;;;;;;;;;;;;;SC9CrC2C,EAAcC;IAC5B,IAAIC,IAAQ;IACZ,KAAK,MAAMC,KAAOF,GACZG,OAAOC,UAAUC,eAAeC,KAAKN,GAAKE,MAC5CD;IAGJ,OAAOA;;;SAGO3C,EACd0C,GACAlC;IAEA,KAAK,MAAMoC,KAAOF,GACZG,OAAOC,UAAUC,eAAeC,KAAKN,GAAKE,MAC5CpC,EAAGoC,GAAKF,EAAIE;;;SAKFK,EAAWP;IAKzB,KAAK,MAAME,KAAOF,GAChB,IAAIG,OAAOC,UAAUC,eAAeC,KAAKN,GAAKE,IAC5C,QAAO;IAGX,QAAO;;;;;;;;;;;;;;;;;;;;;;;;;;GDFAT,WAAQ,IAAIA,EAAY,IAAIpB,EAAa;;MExBrCmC;IAGXxJ,YAAqCyJ;kBAAAA;;IAErCzJ,wBAAwB0J;QACtB,MAAMD,IAAeE,GAAgBC,KAAcC,KAAKH;QACxD,OAAO,IAAIF,EAAWC;;IAGxBzJ,sBAAsB8J;QACpB,MAAML;;;;iBA4BiCK;YACzC,IAAIL,IAAe;YACnB,KAAK,IAAI7C,IAAI,GAAGA,IAAIkD,EAAMhE,UAAUc,GAClC6C,KAAgBpE,OAAO0E,aAAaD,EAAMlD;YAE5C,OAAO6C;;;;GAjCgBO,EAA2BF;QAChD,OAAO,IAAIN,EAAWC;;IAGxBzJ;QACE,OAAO2J,GAAgBC,KAAcK,KAAK/J,KAAKuJ;;IAGjDzJ;QACE,gBA8BuCyJ;YACzC,MAAMS,IAAS,IAAIC,WAAWV,EAAa3D;YAC3C,KAAK,IAAIc,IAAI,GAAGA,IAAI6C,EAAa3D,QAAQc,KACvCsD,EAAOtD,KAAK6C,EAAaW,WAAWxD;YAEtC,OAAOsD;;;;;;;;;;;;;;;;;;;;GAnCEG,EAA2BnK,KAAKuJ;;IAGzCzJ;QACE,OAAkC,IAA3BE,KAAKuJ,GAAa3D;;IAG3B9F,EAAUkF;QACR,OAAOC,GAAoBjF,KAAKuJ,IAAcvE,EAAMuE;;IAGtDzJ,QAAQkF;QACN,OAAOhF,KAAKuJ,OAAiBvE,EAAMuE;;;;SCnCvBa,EAAkBvI;IAChC,OAAOA,QAAAA;;;yDAIOwI,EAAexI;;;IAG7B,QAAkB,MAAXA,KAAgB,IAAIA,MAAU,IAAA;;;;;;GDJrCyH,QAAoC,IAAIA,EAAW;;SEiBrCgB,EAAkBzI;;IAEhC,OAPgC,sDAMlBA,QAAAA,aAAAA,EAAO0I,uCAAUC,WAAU,IAAY,uCAAGC;;;;;;;;;SAkD1CC,EAAkB7I;IAChC,MAAM8I,IAAiBC,EACrB/I,EAAM0I,SAAUC,OAA4B,qBAAiB;IAE/D,OAAO,IAAIpG,EAAUuG,EAAetG,SAASsG,EAAeE;;;;;;;;;;;;;;;;;;;oECnE9D;MAAMC,IAAwB,IAAIC,OAChC;;0EAIcC,EAAUnJ;IACxB,OAAI,eAAeA,wBAER,kBAAkBA,2BAElB,kBAAkBA,KAAS,iBAAiBA,0BAE5C,oBAAoBA,6BAEpB,iBAAiBA,0BAEjB,gBAAgBA,wBAEhB,oBAAoBA,uBAEpB,mBAAmBA,4BAEnB,gBAAgBA,yBAEhB,cAAcA,IACnByI,EAAkBzI,2DAnCSiE;;;sFA6CnBmF,EAAYhE,GAAiBC;IAC3C,MAAMgE,IAAWF,EAAU/D;IAE3B,IAAIiE,MADcF,EAAU9D,IAE1B,QAAO;IAGT,QAAQgE;MACN;QACE,QAAO;;MACT;QACE,OAAOjE,EAAKkE,iBAAiBjE,EAAMiE;;MACrC;QACE,OAAOT,EAAkBzD,GAAMzB,QAAQkF,EAAkBxD;;MAC3D;QACE,OAwBN,SAAyBD,GAAiBC;YACxC,IACiC,mBAAxBD,EAAKmE,kBACoB,mBAAzBlE,EAAMkE,kBACbnE,EAAKmE,eAAexF,WAAWsB,EAAMkE,eAAexF;;YAGpD,OAAOqB,EAAKmE,mBAAmBlE,EAAMkE;YAGvC,MAAMC,IAAgBT,EAAmB3D,EAAoB,iBACvDqE,IAAiBV,EAAmB1D,EAAqB;YAC/D,OACEmE,EAAchH,YAAYiH,EAAejH,WACzCgH,EAAcR,UAAUS,EAAeT;SAtC9BU,CAAgBtE,GAAMC;;MAC/B;QACE,OAAOD,EAAKwD,gBAAgBvD,EAAMuD;;MACpC;QACE,OA+CN,SAAoBxD,GAAiBC;YACnC,OAAOsE,EAAoBvE,EAAgB,YAAEzB,QAC3CgG,EAAoBtE,EAAiB;SAjD5BuE,CAAWxE,GAAMC;;MAC1B;QACE,OAAOD,EAAKyE,mBAAmBxE,EAAMwE;;MACvC;QACE,OAkCN,SAAwBzE,GAAiBC;YACvC,OACEyE,EAAgB1E,EAAK2E,cAAeC,cAClCF,EAAgBzE,EAAM0E,cAAeC,aACvCF,EAAgB1E,EAAK2E,cAAeE,eAClCH,EAAgBzE,EAAM0E,cAAeE;SAvC9BC,CAAe9E,GAAMC;;MAC9B;QACE,gBA+CuBD,GAAiBC;YAC5C,IAAI,kBAAkBD,KAAQ,kBAAkBC,GAC9C,OACEyE,EAAgB1E,EAAK+E,kBAAkBL,EAAgBzE,EAAM8E;YAE1D,IAAI,iBAAiB/E,KAAQ,iBAAiBC,GAAO;gBAC1D,MAAM+E,IAAKN,EAAgB1E,EAAiB,cACtCiF,IAAKP,EAAgBzE,EAAkB;gBAE7C,OAAI+E,MAAOC,IACF7B,EAAe4B,OAAQ5B,EAAe6B,KAEtCC,MAAMF,MAAOE,MAAMD;;YAI9B,QAAO;SA/DIE,CAAanF,GAAMC;;MAC5B;QACE,OAAOmF,GACLpF,EAAKqF,WAAYC,UAAU,IAC3BrF,EAAMoF,WAAYC,UAAU,IAC5BtB;;MAEJ;QACE,OA0DN,SAAsBhE,GAAiBC;YACrC,MAAMsF,IAAUvF,EAAKsD,SAAUC,UAAU,IACnCiC,IAAWvF,EAAMqD,SAAUC,UAAU;YAE3C,IAAI3B,EAAW2D,OAAa3D,EAAW4D,IACrC,QAAO;YAGT,KAAK,MAAMzD,KAAOwD,GAChB,IAAIA,EAAQrD,eAAeH,YAELnD,MAAlB4G,EAASzD,OACRiC,EAAYuB,EAAQxD,IAAMyD,EAASzD,MAEpC,QAAO;YAIb,QAAO;;6EA5EI0D,EAAazF,GAAMC;;MAC5B;QACE,OAhF6BpB;;;;SA8JnB6G,EACdC,GACAC;IAEA,YACgEhH,OAA7D+G,EAASL,UAAU,IAAIO,KAAKC,KAAK9B,EAAY8B,GAAGF;;;SAIrCG,EAAa/F,GAAiBC;IAC5C,MAAMgE,IAAWF,EAAU/D,IACrBgG,IAAYjC,EAAU9D;IAE5B,IAAIgE,MAAa+B,GACf,OAAOhI,GAAoBiG,GAAU+B;IAGvC,QAAQ/B;MACN;QACE,OAAO;;MACT;QACE,OAAOjG,GAAoBgC,EAAkB,cAAEC,EAAmB;;MACpE;QACE,OAyBN,SAAwBD,GAAiBC;YACvC,MAAMgG,IAAavB,EAAgB1E,EAAK+E,gBAAgB/E,EAAKkG,cACvDC,IAAczB,EAAgBzE,EAAM8E,gBAAgB9E,EAAMiG;YAEhE,OAAID,IAAaE,KACP,IACCF,IAAaE,IACf,IACEF,MAAeE,IACjB;;YAGHjB,MAAMe,KACDf,MAAMiB,KAAe,KAAK,IAE1B;SAxCAC,CAAepG,GAAMC;;MAC9B;QACE,OAAOoG,EAAkBrG,EAAoB,gBAAEC,EAAqB;;MACtE;QACE,OAAOoG,EACL5C,EAAkBzD,IAClByD,EAAkBxD;;MAEtB;QACE,OAAOjC,GAAoBgC,EAAiB,aAAEC,EAAkB;;MAClE;QACE,OAkFN,SACED,GACAC;YAEA,MAAMqG,IAAY/B,EAAoBvE,IAChCuG,IAAahC,EAAoBtE;YACvC,OAAOqG,EAAUE,EAAUD;SAxFhBE,CAAazG,EAAgB,YAAEC,EAAiB;;MACzD;QACE,OAsDN,SAA2ByG,GAAkBC;YAC3C,MAAMC,IAAeF,EAASlG,MAAM,MAC9BqG,IAAgBF,EAAUnG,MAAM;YACtC,KAAK,IAAIf,IAAI,GAAGA,IAAImH,EAAajI,UAAUc,IAAIoH,EAAclI,QAAQc,KAAK;gBACxE,MAAMqH,IAAa9I,GAAoB4I,EAAanH,IAAIoH,EAAcpH;gBACtE,IAAmB,MAAfqH,GACF,OAAOA;;YAGX,OAAO9I,GAAoB4I,EAAajI,QAAQkI,EAAclI;SA/DnDoI,CAAkB/G,EAAoB,gBAAEC,EAAqB;;MACtE;QACE,OAgEN,SAA0BD,GAAkBC;YAC1C,MAAM6G,IAAa9I,GACjB0G,EAAgB1E,EAAK4E,WACrBF,EAAgBzE,EAAM2E;YAExB,IAAmB,MAAfkC,GACF,OAAOA;YAET,OAAO9I,GACL0G,EAAgB1E,EAAK6E,YACrBH,EAAgBzE,EAAM4E;SA1EbmC,CAAiBhH,EAAmB,eAAEC,EAAoB;;MACnE;QACE,OAqFN,SAAuBD,GAAsBC;YAC3C,MAAMgH,IAAYjH,EAAKsF,UAAU,IAC3B4B,IAAajH,EAAMqF,UAAU;YAEnC,KAAK,IAAI7F,IAAI,GAAGA,IAAIwH,EAAUtI,UAAUc,IAAIyH,EAAWvI,UAAUc,GAAG;gBAClE,MAAM0H,IAAUpB,EAAakB,EAAUxH,IAAIyH,EAAWzH;gBACtD,IAAI0H,GACF,OAAOA;;YAGX,OAAOnJ,GAAoBiJ,EAAUtI,QAAQuI,EAAWvI;SA/F7CyI,CAAcpH,EAAgB,YAAEC,EAAiB;;MAC1D;QACE,OAgGN,SAAqBD,GAAoBC;YACvC,MAAMsF,IAAUvF,EAAKuD,UAAU,IACzB8D,IAAWrF,OAAOsF,KAAK/B,IACvBC,IAAWvF,EAAMsD,UAAU,IAC3BgE,IAAYvF,OAAOsF,KAAK9B;;;;;YAM9B6B,EAASG,QACTD,EAAUC;YAEV,KAAK,IAAI/H,IAAI,GAAGA,IAAI4H,EAAS1I,UAAUc,IAAI8H,EAAU5I,UAAUc,GAAG;gBAChE,MAAMgI,IAAazJ,GAAoBqJ,EAAS5H,IAAI8H,EAAU9H;gBAC9D,IAAmB,MAAfgI,GACF,OAAOA;gBAET,MAAMN,IAAUpB,EAAaR,EAAQ8B,EAAS5H,KAAK+F,EAAS+B,EAAU9H;gBACtE,IAAgB,MAAZ0H,GACF,OAAOA;;YAIX,OAAOnJ,GAAoBqJ,EAAS1I,QAAQ4I,EAAU5I;;;;;GAxH3C+I,EAAY1H,EAAc,UAAEC,EAAe;;MACpD;QACE,MA1M6BpB;;;;AAkOnC,SAASwH,EAAkBrG,GAAqBC;IAC9C,IACkB,mBAATD,KACU,mBAAVC,KACPD,EAAKrB,WAAWsB,EAAMtB,QAEtB,OAAOX,GAAoBgC,GAAMC;IAGnC,MAAMmE,IAAgBT,EAAmB3D,IACnCqE,IAAiBV,EAAmB1D,IAEpC6G,IAAa9I,GACjBoG,EAAchH,SACdiH,EAAejH;IAEjB,OAAmB,MAAf0J,IACKA,IAEF9I,GAAoBoG,EAAcR,OAAOS,EAAeT;;;SAkFjD+D,EAAY/M;IAC1B,OAAOgN,EAAchN;;;AAGvB,SAASgN,EAAchN;IACrB,OAAI,eAAeA,IACV,SACE,kBAAkBA,IACpB,KAAKA,EAAMsJ,eACT,kBAAkBtJ,IACpB,KAAKA,EAAMmK,eACT,iBAAiBnK,IACnB,KAAKA,EAAMsL,cACT,oBAAoBtL,IAuBjC,SAA2ByD;QACzB,MAAMwJ,IAAsBlE,EAAmBtF;QAC/C,OAAO,QAAQwJ,EAAoBzK,WAAWyK,EAAoBjE;KAxBzDkE,CAAkBlN,EAAqB,kBACrC,iBAAiBA,IACnBA,EAAM4I,cACJ,gBAAgB5I,IAgBpB2J,EAfqB3J,EAAiB,YAeNmN,aAd5B,oBAAoBnN,KA0BN6J,IAzBE7J,EAAqB;IA0BzC0G,EAAY0G,EAASvD,GAAgBhK,cAzBjC,mBAAmBG,IAqBvB,QADiBqN,IAnBErN,EAAoB,eAoBvBgK,YAAYqD,EAASpD,eAnBjC,gBAAgBjK,IA4C7B,SAAuByK;QACrB,IAAI6C,IAAS,KACTC,KAAQ;QACZ,KAAK,MAAMvN,KAASyK,EAAWC,UAAU,IAClC6C,IAGHA,KAAQ,IAFRD,KAAU,KAIZA,KAAUN,EAAchN;QAE1B,OAAOsN,IAAS;;;;;GAtDPE,EAAcxN,EAAiB,cAC7B,cAAcA,IAwB3B,SAAqB0I;;;QAGnB,MAAM+E,IAAarG,OAAOsF,KAAKhE,EAASC,UAAU,IAAIiE;QAEtD,IAAIU,IAAS,KACTC,KAAQ;QACZ,KAAK,MAAMpG,KAAOsG,GACXF,IAGHA,KAAQ,IAFRD,KAAU,KAIZA,KAAU,GAAGnG,KAAO6F,EAActE,EAASC,OAAQxB;QAErD,OAAOmG,IAAS;KAtCPI,CAAY1N,EAAe,YAjWHiE;IAgXnC,IAA0BoJ,GAICxD;;;SAiGXd,EACdlG;;;;IAOA,IAzcoDrB,KAocvCqB,IAKO,mBAATA,GAAmB;;;;QAK5B,IAAImG,IAAQ;QACZ,MAAM2E,IAAW1E,EAAsB2E,KAAK/K;QAE5C,IAjdkDrB,KAgdrCmM,IACTA,EAAS,IAAI;;YAEf,IAAIE,IAAUF,EAAS;YACvBE,KAAWA,IAAU,aAAaC,OAAO,GAAG,IAC5C9E,IAAQ+E,OAAOF;;;gBAIjB,MAAMG,IAAa,IAAIrL,KAAKE;QAG5B,OAAO;YAAEL,SAFOQ,KAAKC,MAAM+K,EAAWlL,YAAY;YAEhCkG,OAAAA;;;IAOlB,OAAO;QAAExG,SAFOsH,EAAgBjH,EAAKL;QAEnBwG,OADJc,EAAgBjH,EAAKmG;;;;;;;aASvBc,EAAgB9J;;IAE9B,OAAqB,mBAAVA,IACFA,IACmB,mBAAVA,IACT+N,OAAO/N,KAEP;;;+EAKK2J,EAAoBsE;IAClC,OAAoB,mBAATA,IACFxG,EAAWyG,iBAAiBD,KAE5BxG,EAAW0G,eAAeF;;;6EAKrBG,EAASC,GAAwBlH;IAC/C,OAAO;QACL0C,gBAAgB,YAAYwE,EAAWC,uBACrCD,EAAWE,sBACCpH,EAAIzB,KAAKD;;;;6DAKX+I,EACdxO;IAEA,SAASA,KAAS,kBAAkBA;;;;;SAgBtByO,EACdzO;IAEA,SAASA,KAAS,gBAAgBA;;;wDAWpB0O,EACd1O;IAEA,SAASA,KAAS,eAAeA;;;gDAInB2O,GACd3O;IAEA,SAASA,KAAS,iBAAiBA,KAASsK,MAAMyD,OAAO/N,EAAMsL;;;uDAIjDsD,GACd5O;IAEA,SAASA,KAAS,cAAcA;;;;;;;;;;;;;;;;;;;mEC9hBrB6O;IACX5Q;IAGAA,GACE6Q,GACAhG;QAEA,gBF1BFA,GACAgG;YAEA,MAAMpG,IAAyB;gBAC7BC,QAAQ;oBACNoG,UAAY;wBACVnG,aApB0B;;oBAsB5BoG,sBAAwB;wBACtBzF,gBAAgB;4BACd/G,SAASsG,EAAetG;4BACxBwG,OAAOF,EAAerG;;;;;YAU9B,OAJIqM,MACFpG,EAASC,OAA0B,qBAAImG,IAGlC;gBAAEpG,UAAAA;;;;;;;;GEKAuG,KAAiCH;;IAG1C7Q,GACE6Q,GACAI;QAEA,OAAOA;;IAGTjR,GAAiB6Q;QACf,OAAO;;;IAGT7Q,QAAQkF;QACN,OAAOA,aAAiB0L;;;;AArBnBA,cAAW,IAAIA;;;MA0BXM;IACXlR,YAAqBmR;QAAAjR,gBAAAiR;;IAErBnR,GACE6Q,GACAhG;QAEA,OAAO3K,KAAKkR,MAAMP;;IAGpB7Q,GACE6Q,GACAI;;;;QAKA,OAAO/Q,KAAKkR,MAAMP;;IAGZ7Q,MAAM6Q;QACZ,MAAMpE,IAAS4E,GAAwBR;QACvC,KAAK,MAAMS,KAAWpR,KAAKiR,UACpB1E,EAAO8E,KAAKC,KAAWrG,EAAYqG,GAASF,OAC/C7E,EAAOjG,KAAK8K;QAGhB,OAAO;YAAE9E,YAAY;gBAAEC,QAAAA;;;;IAGzBzM,GAAiB6Q;QACf,OAAO;;;IAGT7Q,QAAQkF;QACN,OACEA,aAAiBgM,MACjB3E,GAAYrM,KAAKiR,UAAUjM,EAAMiM,UAAUhG;;;;+DAMpCsG;IACXzR,YAAqBmR;QAAAjR,gBAAAiR;;IAErBnR,GACE6Q,GACAhG;QAEA,OAAO3K,KAAKkR,MAAMP;;IAGpB7Q,GACE6Q,GACAI;;;;QAKA,OAAO/Q,KAAKkR,MAAMP;;IAGZ7Q,MAAM6Q;QACZ,IAAIpE,IAAS4E,GAAwBR;QACrC,KAAK,MAAMa,KAAYxR,KAAKiR,UAC1B1E,IAASA,EAAO7E,OAAO4J,MAAYrG,EAAYqG,GAASE;QAE1D,OAAO;YAAElF,YAAY;gBAAEC,QAAAA;;;;IAGzBzM,GAAiB6Q;QACf,OAAO;;;IAGT7Q,QAAQkF;QACN,OACEA,aAAiBuM,MACjBlF,GAAYrM,KAAKiR,UAAUjM,EAAMiM,UAAUhG;;;;;;;;;UAWpCwG;IACX3R,YACmB4R,GACRC;QADQ3R,kBAAA0R,aACRC;;IAQX7R,GACE6Q,GACAhG;;;;QAKA,MAAMiH,IAAY5R,KAAK6R,GAAiBlB,IAClCmB,IAAM9R,KAAK+R,SAASH,KAAa5R,KAAK+R,SAAS/R,KAAK2R;QAC1D,OAAItB,EAAUuB,MAAcvB,EAAUrQ,KAAK2R,MAClC3R,KAAK0R,WAAWM,GAAUF,KAE1B9R,KAAK0R,WAAWO,GAASH;;IAIpChS,GACE6Q,GACAI;QAMA,OAAOA;;;;;WAOTjR,GAAiB6Q;QACf,OD2VKN,EADgBxO,IC1VL8O,eDoVlB9O;YAEA,SAASA,KAAS,iBAAiBA;;kFAKRqQ,EAASrQ,KC3VD8O,IAAiB;YAAE3E,cAAc;;YD0V7CnK;;ICvVvB/B,QAAQkF;QACN,OACEA,aAAiByM,MACjBxG,EAAYjL,KAAK2R,IAAS3M,EAAM2M;;IAI5B7R,SAAS+B;QACf,OAAO8J,EAAgB9J,EAAMmK,gBAAgBnK,EAAMsL;;;;AAIvD,SAASgE,GAAwBtP;IAC/B,OAAOyO,EAAQzO,MAAUA,EAAMyK,WAAWC,SACtC1K,EAAMyK,WAAWC,OAAOrG,UACxB;;;;;;;;;;;;;;;;;;;;;;;;;;;;UC9MOiM;IACXrS,YAAqB0K;QAAAxK,cAAAwK;;;QAGnBA,EAAOiE,KAAK7G,EAAU5B;;;;;;;WAcxBlG,GAAOsS;QACL,KAAK,MAAMC,KAAiBrS,KAAKwK,QAC/B,IAAI6H,EAAcC,EAAWF,IAC3B,QAAO;QAGX,QAAO;;IAGTtS,QAAQkF;QACN,OAAOqH,GAAYrM,KAAKwK,QAAQxF,EAAMwF,QAAQ,CAAC+H,GAAGC,MAAMD,EAAE/M,QAAQgN;;;;yEAKzDC;IACX3S,YACW4S,GACAC;QADA3S,aAAA0S,GACA1S,iBAAA2S;;IAGX7S,QAAQkF;QACN,OACEhF,KAAK0S,MAAMlN,QAAQR,EAAM0N,UAAU1S,KAAK2S,UAAUnN,QAAQR,EAAM2N;;;;4EAMzDC;IACX9S;;;;;;;;;;;IAWW+S;;;;;;;;IAQAC;QARA9S,eAAA6S,GAQA7S,wBAAA8S;;;;;;;;UAiBAC;IACXjT,YACWkT,GACAC;QADAjT,kBAAAgT,GACAhT,cAAAiT;;gDASXnT;QACE,OAAO,IAAIiT;;8DAIbjT,cAAcmT;QACZ,OAAO,IAAIF,QAAalN,GAAWoN;;kFAIrCnT,kBAAkB+S;QAChB,OAAO,IAAIE,GAAaF;;0DAI1BK;QACE,YAA2BrN,MAApB7F,KAAKgT,mBAA4CnN,MAAhB7F,KAAKiT;;;;;WAO/CnT,GAAWqT;QACT,YAAwBtN,MAApB7F,KAAKgT,aAELG,aAAoBC,MACpBD,EAASN,QAAQrN,QAAQxF,KAAKgT,mBAEPnN,MAAhB7F,KAAKiT,UACPjT,KAAKiT,WAAWE,aAAoBC;;IAO/CtT,QAAQkF;QACN,OACEhF,KAAKiT,WAAWjO,EAAMiO,WACrBjT,KAAKgT,eACAhO,EAAMgO,cAAchT,KAAKgT,WAAWxN,QAAQR,EAAMgO,eACnDhO,EAAMgO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;UAsDKK;IAgEVvT,GAAiBqT;;;;;;WAejBrT,UACRqT;QAEA,OAAIA,aAAoBC,KACfD,EAASN,UAETxN,EAAgB2B;;;;;;;UAShBsM,WAAoBD;IAC/BvT,YACWkJ,GACAnH,GACA0R;QAET9R,SAJSzB,WAAAgJ,GACAhJ,aAAA6B,aACA0R,GAKFvT;;IAETF,GACEqT,GACAK;QAEAxT,KAAKyT,GAAiBN;;;;QAWtB,MAAMN,IAAUW,EAAeX;QAC/B,OAAO,IAAIO,GAASpT,KAAKgJ,KAAK6J,GAAS7S,KAAK6B,OAAO;YACjD6R,wBAAuB;;;IAI3B5T,GACEqT,GACAQ,GACAhJ;QAIA,IAFA3K,KAAKyT,GAAiBN,KAEjBnT,KAAKuT,GAAaK,GAAWT,IAChC,OAAOA;QAGT,MAAMN,IAAUQ,GAASQ,GAAuBV;QAChD,OAAO,IAAIC,GAASpT,KAAKgJ,KAAK6J,GAAS7S,KAAK6B,OAAO;YACjDiS,KAAmB;;;IAIvBhU,GAAiBqT;QACf,OAAO;;IAGTrT,QAAQkF;QACN,OACEA,aAAiBsO,MACjBtT,KAAKgJ,IAAIxD,QAAQR,EAAMgE,QACvBhJ,KAAK6B,MAAM2D,QAAQR,EAAMnD,UACzB7B,KAAKuT,GAAa/N,QAAQR,EAAMuO;;;;;;;;;;;;;;;;UAkBzBQ,WAAsBV;IACjCvT,YACWkJ,GACAgL,GACAC,GACAV;QAET9R,SALSzB,WAAAgJ,GACAhJ,YAAAgU,aACAC,aACAV,GAKFvT;;IAETF,GACEqT,GACAK;QASA,IAPAxT,KAAKyT,GAAiBN,KAOjBnT,KAAKuT,GAAaK,GAAWT;;;;;QAKhC,OAAO,IAAIe,GAAgBlU,KAAKgJ,KAAKwK,EAAeX;QAGtD,MAAMsB,IAAUnU,KAAKoU,GAAcjB;QACnC,OAAO,IAAIC,GAASpT,KAAKgJ,KAAKwK,EAAeX,SAASsB,GAAS;YAC7DT,wBAAuB;;;IAI3B5T,GACEqT,GACAQ,GACAhJ;QAIA,IAFA3K,KAAKyT,GAAiBN,KAEjBnT,KAAKuT,GAAaK,GAAWT,IAChC,OAAOA;QAGT,MAAMN,IAAUQ,GAASQ,GAAuBV,IAC1CgB,IAAUnU,KAAKoU,GAAcjB;QACnC,OAAO,IAAIC,GAASpT,KAAKgJ,KAAK6J,GAASsB,GAAS;YAC9CL,KAAmB;;;IAIvBhU,GAAiBqT;QACf,OAAO;;IAGTrT,QAAQkF;QACN,OACEA,aAAiB+O,MACjB/T,KAAKgJ,IAAIxD,QAAQR,EAAMgE,QACvBhJ,KAAKiU,GAAUzO,QAAQR,EAAMiP,OAC7BjU,KAAKuT,GAAa/N,QAAQR,EAAMuO;;;;;;WAS5BzT,GAAcqT;QACpB,IAAIa;QAMJ,OAJEA,IADEb,aAAoBC,KACfD,EAASa,SAETK,GAAYC,SAEdtU,KAAKuU,GAAYP;;IAGlBlU,GAAYkU;QAClB,MAAMQ,IAAU,IAAIC,GAAmBT;QAWvC,OAVAhU,KAAKiU,GAAUzJ,OAAOpE,QAAQgM;YAC5B,KAAKA,EAAU/I,KAAW;gBACxB,MAAMqL,IAAW1U,KAAKgU,KAAKtB,MAAMN;gBAChB,SAAbsC,IACFF,EAAQG,IAAIvC,GAAWsC,KAEvBF,EAAQI,OAAOxC;;YAIdoC,EAAQK;;;;;;;;;;;;UAaNC,WAA0BzB;IAQrCvT,YACWkJ,GACA+L;QAETtT,SAHSzB,WAAAgJ,GACAhJ,uBAAA+U,GATF/U;;;;QAKTA,UAAwB+S,GAAaE,QAAO;;IAS5CnT,GACEqT,GACAK;QASA,IAPAxT,KAAKyT,GAAiBN,IAEtB9P,GACqC,QAAnCmQ,EAAeV,oBAIZ9S,KAAKuT,GAAaK,GAAWT;;;;;QAKhC,OAAO,IAAIe,GAAgBlU,KAAKgJ,KAAKwK,EAAeX;QAGtD,MAAMmC,IAAMhV,KAAKiV,GAAgB9B,IAC3BL,IAAmB9S,KAAKkV,GAC5B/B,GACAK,EAAgC,mBAG5BX,IAAUW,EAAeX,SACzBsB,IAAUnU,KAAKmV,GAAgBH,EAAIhB,QAAQlB;QACjD,OAAO,IAAIM,GAASpT,KAAKgJ,KAAK6J,GAASsB,GAAS;YAC9CT,wBAAuB;;;IAI3B5T,GACEqT,GACAQ,GACAhJ;QAIA,IAFA3K,KAAKyT,GAAiBN,KAEjBnT,KAAKuT,GAAaK,GAAWT,IAChC,OAAOA;QAGT,MAAM6B,IAAMhV,KAAKiV,GAAgB9B,IAC3BL,IAAmB9S,KAAKoV,GAC5BzK,GACAwI,GACAQ,IAEIQ,IAAUnU,KAAKmV,GAAgBH,EAAIhB,QAAQlB;QACjD,OAAO,IAAIM,GAASpT,KAAKgJ,KAAKgM,EAAInC,SAASsB,GAAS;YAClDL,KAAmB;;;IAIvBhU,GAAiBqT;QACf,IAAIkC,IAAwC;QAC5C,KAAK,MAAMC,KAAkBtV,KAAK+U,iBAAiB;YACjD,MAAMQ,IACJpC,aAAoBC,KAChBD,EAAST,MAAM4C,EAAe5C,cAC9B7M,GACA2P,IAAeF,EAAe3C,UAAUd,GAC5C0D,KAAiB;YAGC,QAAhBC,MAEAH,IADgB,QAAdA,KACW,IAAIZ,IAAqBE,IACpCW,EAAe5C,OACf8C,KAGWH,EAAWV,IAAIW,EAAe5C,OAAO8C;;QAIxD,OAAOH,IAAaA,EAAWR,OAAU;;IAG3C/U,QAAQkF;QACN,OACEA,aAAiB8P,MACjB9U,KAAKgJ,IAAIxD,QAAQR,EAAMgE,QACvBqD,GAAYrM,KAAK+U,iBAAiB/P,EAAM+P,iBAAiB,CAACxC,GAAGC,MAC3DD,EAAE/M,QAAQgN,OAEZxS,KAAKuT,GAAa/N,QAAQR,EAAMuO;;;;;;;WAU5BzT,GAAgBqT;QAStB,OAAOA;;;;;;;;;;WAYDrT,GACN6T,GACAuB;QAEA,MAAMpC,IAAgC;QA/lBjCzP,GAimBHrD,KAAK+U,gBAAgBnP,WAAWsP,EAAuBtP;QAKzD,KAAK,IAAIc,IAAI,GAAGA,IAAIwO,EAAuBtP,QAAQc,KAAK;YACtD,MAAM4O,IAAiBtV,KAAK+U,gBAAgBrO,IACtCiM,IAAY2C,EAAe3C;YACjC,IAAIhC,IAAkC;YAClCgD,aAAmBP,OACrBzC,IAAgBgD,EAAQjB,MAAM4C,EAAe5C,SAE/CI,EAAiBxM,KACfqM,EAAU8C,GACR9E,GACAuE,EAAuBxO;;QAI7B,OAAOoM;;;;;;;;;;;;;WAeDhT,GACN6K,GACAwI,GACAQ;QAEA,MAAMb,IAAgC;QACtC,KAAK,MAAMwC,KAAkBtV,KAAK+U,iBAAiB;YACjD,MAAMpC,IAAY2C,EAAe3C;YAEjC,IAAIhC,IAAkC;YAClCwC,aAAoBC,OACtBzC,IAAgBwC,EAAST,MAAM4C,EAAe5C,SAG1B,SAAlB/B,KAA0BgD,aAAmBP;;;;;YAK/CzC,IAAgBgD,EAAQjB,MAAM4C,EAAe5C,SAG/CI,EAAiBxM,KACfqM,EAAU+C,GAAiB/E,GAAehG;;QAG9C,OAAOmI;;IAGDhT,GACNkU,GACAlB;QAOA,MAAM0B,IAAU,IAAIC,GAAmBT;QACvC,KAAK,IAAItN,IAAI,GAAGA,IAAI1G,KAAK+U,gBAAgBnP,QAAQc,KAAK;YACpD,MACM0L,IADiBpS,KAAK+U,gBAAgBrO,GACXgM;YACjC8B,EAAQG,IAAIvC,GAAWU,EAAiBpM;;QAE1C,OAAO8N,EAAQK;;;;oEAKNc,WAAuBtC;IAClCvT,YAAqBkJ,GAA2BuK;QAC9C9R,SADmBzB,WAAAgJ,aAA2BuK,GAIvCvT;;IAETF,GACEqT,GACAK;;;;QAaA,OAXAxT,KAAKyT,GAAiBN,IAWf,IAAIyC,GAAW5V,KAAKgJ,KAAKwK,EAAeX,SAAS;YACtDa,wBAAuB;;;IAI3B5T,GACEqT,GACAQ,GACAhJ;QAIA,OAFA3K,KAAKyT,GAAiBN,IAEjBnT,KAAKuT,GAAaK,GAAWT,KAU3B,IAAIyC,GAAW5V,KAAKgJ,KAAK3D,EAAgB2B,SATvCmM;;IAYXrT,GAAiBqT;QACf,OAAO;;IAGTrT,QAAQkF;QACN,OACEA,aAAiB2Q,MACjB3V,KAAKgJ,IAAIxD,QAAQR,EAAMgE,QACvBhJ,KAAKuT,GAAa/N,QAAQR,EAAMuO;;;;;;;;;;UAYzBsC,WAAuBxC;IAClCvT,YAAqBkJ,GAA2BuK;QAC9C9R,SADmBzB,WAAAgJ,aAA2BuK,GAIvCvT;;IAETF,GACEqT,GACAK;QAEA1N;;IAGFhG,GACEqT,GACAQ,GACAhJ;QAEA7E;;IAGFhG,GAAiBqT;QACfrN;;IAGFhG,QAAQkF;QACN,OACEA,aAAiB6Q,MACjB7V,KAAKgJ,IAAIxD,QAAQR,EAAMgE,QACvBhJ,KAAKuT,GAAa/N,QAAQR,EAAMuO;;;;;;;;;;;;;;;;;;;;;;;UChxBzBc;IACXvU,YAA4BgW;QAAA9V,aAAA8V;;IAO5BhW;QACE,OAAO,IAAIuU,GAAY;YAAE9J,UAAU;;;;;;;;WASrCzK,MAAMyH;QACJ,IAAIA,EAAK8B,KACP,OAAOrJ,KAAK8V;QACP;YACL,IAAIjU,IAAmB7B,KAAK8V;YAC5B,KAAK,IAAIpP,IAAI,GAAGA,IAAIa,EAAK3B,SAAS,KAAKc,GAAG;gBACxC,KAAK7E,EAAM0I,SAAUC,QACnB,OAAO;gBAGT,IADA3I,IAAQA,EAAM0I,SAAUC,OAAOjD,EAAKxE,IAAI2D,MACnC+J,GAAW5O,IACd,OAAO;;YAKX,OADAA,KAASA,EAAM0I,SAAUC,UAAU,IAAIjD,EAAKwO,MACrClU,KAAS;;;IAIpB/B,QAAQkF;QACN,OAAOiG,EAAYjL,KAAK8V,OAAO9Q,EAAM8Q;;;;;;;UAe5BrB;;;;IAOX3U,YAA6BuV,IAA0BhB,GAAYC;kBAAtCe;;QAL7BrV,UAAqB,IAAIgW;;;;;;;;WAczBlW,IAAIyH,GAAiB1F;QAMnB,OADA7B,KAAKiW,GAAW1O,GAAM1F,IACf7B;;;;;;;;WAUTF,OAAOyH;QAML,OADAvH,KAAKiW,GAAW1O,GAAM,OACfvH;;;;;WAODF,GAAWyH,GAAiB1F;QAClC,IAAIqU,IAAelW,KAAKmW;QAExB,KAAK,IAAIzP,IAAI,GAAGA,IAAIa,EAAK3B,SAAS,KAAKc,GAAG;YACxC,MAAM0P,IAAiB7O,EAAKxE,IAAI2D;YAChC,IAAI2P,IAAeH,EAAanT,IAAIqT;YAEhCC,aAAwBL;;YAE1BE,IAAeG,IAEfA,8BACArL,EAAUqL;;YAGVA,IAAe,IAAIL,IACjB/M,OAAOqN,QAAQD,EAAa9L,SAAUC,UAAU,MAElD0L,EAAavB,IAAIyB,GAAgBC,IACjCH,IAAeG;;YAGfA,IAAe,IAAIL,KACnBE,EAAavB,IAAIyB,GAAgBC,IACjCH,IAAeG;;QAInBH,EAAavB,IAAIpN,EAAKwO,KAAelU;;iEAIvC/B;QACE,MAAMyW,IAAevW,KAAKwW,GACxB5O,EAAU6O,GACVzW,KAAKmW;QAEP,OAAoB,QAAhBI,IACK,IAAIlC,GAAYkC,KAEhBvW,KAAKqV;;;;;;;;;;;;;WAgBRvV,GACN4W,GACAC;QAEA,IAAIC,KAAW;QAEf,MAAMrB,IAAgBvV,KAAKqV,GAAW3C,MAAMgE,IACtCG,IAAepG,GAAW8E;0BAGvBA,EAAchL,SAASC,UAC5B;QAkBJ,OAhBAmM,EAAgBvQ,QAAQ,CAACvE,GAAOiV;YAC9B,IAAIjV,aAAiBmU,KAAK;gBACxB,MAAMe,IAAS/W,KAAKwW,GAAaE,EAAYM,MAAMF,IAAcjV;gBACnD,QAAVkV,MACFF,EAAaC,KAAeC,GAC5BH,KAAW;mBAEM,SAAV/U,KACTgV,EAAaC,KAAejV,GAC5B+U,KAAW,KACFC,EAAa1N,eAAe2N,cAC9BD,EAAaC,IACpBF,KAAW;YAIRA,IAAW;YAAErM,UAAU;gBAAEC,QAAQqM;;YAAmB;;;;;;aAO/CI,GAAiBpV;IAC/B,MAAM2I,IAAsB;IAsB5B,OArBApE,EAAQvE,EAAO2I,UAAU,IAAI,CAACxB,GAAKnH;QACjC,MAAM6U,IAAc,IAAI9O,EAAU,EAACoB;QACnC,IAAIyH,GAAW5O,IAAQ;YACrB,MACMqV,IADaD,GAAiBpV,EAAe,UACnB2I;YAChC,IAA4B,MAAxB0M,EAAatR;;YAEf4E,EAAOlE,KAAKoQ;;;YAIZ,KAAK,MAAMS,KAAcD,GACvB1M,EAAOlE,KAAKoQ,EAAYM,MAAMG;;;;QAMlC3M,EAAOlE,KAAKoQ;QAGT,IAAIvE,GAAU3H;;;;;;;;;;;;;;;;;;;;;;UCpOD4M;IACpBtX,YAAqBkJ,GAA2B6J;QAA3B7S,WAAAgJ,GAA2BhJ,eAAA6S;;;;;;;UAiBrCO,WAAiBgE;IAI5BtX,YACEkJ,GACA6J,GACiBwE,GACjBC;QAEA7V,MAAMuH,GAAK6J,cAHMwE,GAIjBrX,KAAKuX,OAAsBD,EAAQC,IACnCvX,KAAK0T,0BAA0B4D,EAAQ5D;;IAGzC5T,MAAMyH;QACJ,OAAOvH,KAAKqX,GAAY3E,MAAMnL;;IAGhCzH;QACE,OAAOE,KAAKqX;;IAGdvX;QACE,OAAOE,KAAKqX,GAAYvB;;IAG1BhW,QAAQkF;QACN,OACEA,aAAiBoO,MACjBpT,KAAKgJ,IAAIxD,QAAQR,EAAMgE,QACvBhJ,KAAK6S,QAAQrN,QAAQR,EAAM6N,YAC3B7S,KAAKuX,OAAsBvS,EAAMuS,MACjCvX,KAAK0T,0BAA0B1O,EAAM0O,yBACrC1T,KAAKqX,GAAY7R,QAAQR,EAAMqS;;IAInCvX;QACE,OACE,YAAYE,KAAKgJ,QACfhJ,KAAK6S,YACF7S,KAAKqX,GAAY3V,iBACtB,uBAAuB1B,KAAKuX,WAC5B,2BAA2BvX,KAAK0T;;IAIpC8D;QACE,OAAOxX,KAAKuX,MAAqBvX,KAAK0T;;;;;;;;;;;;;MA2B7BkC,WAAmBwB;IAG9BtX,YACEkJ,GACA6J,GACAyE;QAEA7V,MAAMuH,GAAK6J,IACX7S,KAAK0T,2BAA2B4D,MAAWA,EAAQ5D;;IAGrD5T;QACE,OAAO,cAAcE,KAAKgJ,QAAQhJ,KAAK6S;;IAGzC2E;QACE,OAAOxX,KAAK0T;;IAGd5T,QAAQkF;QACN,OACEA,aAAiB4Q,MACjB5Q,EAAM0O,0BAA0B1T,KAAK0T,yBACrC1O,EAAM6N,QAAQrN,QAAQxF,KAAK6S,YAC3B7N,EAAMgE,IAAIxD,QAAQxF,KAAKgJ;;;;;;;UAShBkL,WAAwBkD;IACnCtX;QACE,OAAO,mBAAmBE,KAAKgJ,QAAQhJ,KAAK6S;;IAG9C2E;QACE,QAAO;;IAGT1X,QAAQkF;QACN,OACEA,aAAiBkP,MACjBlP,EAAM6N,QAAQrN,QAAQxF,KAAK6S,YAC3B7N,EAAMgE,IAAIxD,QAAQxF,KAAKgJ;;;;;;;;;;;;;;;;;;;;;;;;;;UCpJhByO;;;;;;;;;IAWX3X,YACWyH,GACAmQ,IAAiC,MACjCC,IAAqB,IACrBC,IAAoB,IACpBzR,IAAuB,MACvB0R,IAAwB,MACxBC,IAAsB;QANtB9X,YAAAuH,GACAvH,uBAAA0X,GACA1X,eAAA2X,GACA3X,eAAA4X,GACA5X,aAAAmG;QACAnG,eAAA6X,GACA7X,aAAA8X,GAjBX9X,UAA6C;;IAoB7CF;QACE,IAAiC,SAA7BE,KAAK+X,IAA8B;YACrC,IAAInJ,IAAc5O,KAAKuH,KAAKD;YACC,SAAzBtH,KAAK0X,oBACP9I,KAAe,SAAS5O,KAAK0X,kBAE/B9I,KAAe;YACfA,KAAe5O,KAAK4X,QAAQ9P,IAAIkQ,KAAKA,EAAEpJ,eAAevH,KAAK,MAC3DuH,KAAe,QACfA,KAAe5O,KAAK2X,QAAQ7P,IAAIhE,KAAKA,EAAE8K,eAAevH,KAAK;YAEtD+C,EAAkBpK,KAAKmG,WAC1ByI,KAAe,OACfA,KAAe5O,KAAKmG,QAElBnG,KAAK6X,YACPjJ,KAAe,QACfA,KAAe5O,KAAK6X,QAAQjJ;YAE1B5O,KAAK8X,UACPlJ,KAAe,QACfA,KAAe5O,KAAK8X,MAAMlJ,gBAE5B5O,KAAK+X,KAAsBnJ;;QAE7B,OAAO5O,KAAK+X;;IAGdjY;QACE,IAAIiI,IAAM/H,KAAKuH,KAAKD;QAmBpB,OAlB6B,SAAzBtH,KAAK0X,oBACP3P,KAAO,sBAAsB/H,KAAK0X;QAEhC1X,KAAK4X,QAAQhS,SAAS,MACxBmC,KAAO,eAAe/H,KAAK4X,QAAQvQ,KAAK,WAErC+C,EAAkBpK,KAAKmG,WAC1B4B,KAAO,cAAc/H,KAAKmG;QAExBnG,KAAK2X,QAAQ/R,SAAS,MACxBmC,KAAO,eAAe/H,KAAK2X,QAAQtQ,KAAK,WAEtCrH,KAAK6X,YACP9P,KAAO,gBAAgB/H,KAAK6X,QAAQjJ;QAElC5O,KAAK8X,UACP/P,KAAO,cAAc/H,KAAK8X,MAAMlJ,gBAE3B,UAAU7G;;IAGnBjI,QAAQkF;QACN,IAAIhF,KAAKmG,UAAUnB,EAAMmB,OACvB,QAAO;QAGT,IAAInG,KAAK2X,QAAQ/R,WAAWZ,EAAM2S,QAAQ/R,QACxC,QAAO;QAGT,KAAK,IAAIc,IAAI,GAAGA,IAAI1G,KAAK2X,QAAQ/R,QAAQc,KACvC,KAAK1G,KAAK2X,QAAQjR,GAAGlB,QAAQR,EAAM2S,QAAQjR,KACzC,QAAO;QAIX,IAAI1G,KAAK4X,QAAQhS,WAAWZ,EAAM4S,QAAQhS,QACxC,QAAO;QAGT,KAAK,IAAIc,IAAI,GAAGA,IAAI1G,KAAK4X,QAAQhS,QAAQc,KACvC,KAAK1G,KAAK4X,QAAQlR,GAAGlB,QAAQR,EAAM4S,QAAQlR,KACzC,QAAO;QAIX,OAAI1G,KAAK0X,oBAAoB1S,EAAM0S,sBAI9B1X,KAAKuH,KAAK/B,QAAQR,EAAMuC,aAKV,SAAjBvH,KAAK6X,UACA7X,KAAK6X,QAAQrS,QAAQR,EAAM6S,WACV,SAAlB7S,EAAM6S,aAKU,SAAf7X,KAAK8X,QACR9X,KAAK8X,MAAMtS,QAAQR,EAAM8S,SACT,SAAhB9S,EAAM8S;;IAGZhY;QACE,OACEyI,EAAY0P,GAAcjY,KAAKuH,SACN,SAAzBvH,KAAK0X,mBACmB,MAAxB1X,KAAK4X,QAAQhS;;;;;;;;;;;;;;;;;;;;;;;;UCvGNsS;;;;;IAcXpY,YACWyH,GACAmQ,IAAiC,MACjCS,IAA6B,IAC7BP,IAAoB,IACpBzR,IAAuB,MACvBiS,sBACAP,IAAwB,MACxBC,IAAsB;QAPtB9X,YAAAuH,GACAvH,uBAAA0X,aACAS,GACAnY,eAAA4X,GACA5X,aAAAmG;kBACAiS,GACApY,eAAA6X,GACA7X,aAAA8X,GAjBX9X,UAA4C;;QAG5CA,UAAwC,MAgBlCA,KAAK6X,WACP7X,KAAKqY,GAAiBrY,KAAK6X,UAEzB7X,KAAK8X,SACP9X,KAAKqY,GAAiBrY,KAAK8X;;IA3B/BhY,UAAcyH;QACZ,OAAO,IAAI2Q,GAAM3Q;;IA8BnBoQ;QACE,IAA6B,SAAzB3X,KAAKsY,IAA0B;YACjCtY,KAAKsY,KAAkB;YAEvB,MAAMC,IAAkBvY,KAAKwY,MACvBC,IAAoBzY,KAAK0Y;YAC/B,IAAwB,SAApBH,KAAkD,SAAtBE;;;;YAIzBF,EAAgBI,OACnB3Y,KAAKsY,GAAgBhS,KAAK,IAAIsS,GAAQL,KAExCvY,KAAKsY,GAAgBhS,KACnB,IAAIsS,GAAQhR,EAAUiR,mCAEnB;gBAOL,IAAIC,KAAmB;gBACvB,KAAK,MAAMnB,KAAW3X,KAAKmY,IACzBnY,KAAKsY,GAAgBhS,KAAKqR,IACtBA,EAAQjF,MAAMiG,QAChBG,KAAmB;gBAGvB,KAAKA,GAAkB;;;oBAGrB,MAAMC,IACJ/Y,KAAKmY,GAAgBvS,SAAS,IAC1B5F,KAAKmY,GAAgBnY,KAAKmY,GAAgBvS,SAAS,GAAGoT;oBAE5DhZ,KAAKsY,GAAgBhS,KACnB,IAAIsS,GAAQhR,EAAUiR,KAAYE;;;;QAK1C,OAAO/Y,KAAKsY;;IAGdxY,GAAU4H;QAcR,MAAMuR,IAAajZ,KAAK4X,QAAQsB,OAAO,EAACxR;QACxC,OAAO,IAAIwQ,GACTlY,KAAKuH,MACLvH,KAAK0X,iBACL1X,KAAKmY,GAAgBjS,SACrB+S,GACAjZ,KAAKmG,OACLnG,KAAKoY,IACLpY,KAAK6X,SACL7X,KAAK8X;;IAIThY,GAAW6X;;QAMT,MAAMwB,IAAanZ,KAAKmY,GAAgBe,OAAO,EAACvB;QAChD,OAAO,IAAIO,GACTlY,KAAKuH,MACLvH,KAAK0X,iBACLyB,GACAnZ,KAAK4X,QAAQ1R,SACblG,KAAKmG,OACLnG,KAAKoY,IACLpY,KAAK6X,SACL7X,KAAK8X;;IAIThY,GAAiBqG;QACf,OAAO,IAAI+R,GACTlY,KAAKuH,MACLvH,KAAK0X,iBACL1X,KAAKmY,GAAgBjS,SACrBlG,KAAK4X,QAAQ1R,SACbC,qBAEAnG,KAAK6X,SACL7X,KAAK8X;;IAIThY,GAAgBqG;QACd,OAAO,IAAI+R,GACTlY,KAAKuH,MACLvH,KAAK0X,iBACL1X,KAAKmY,GAAgBjS,SACrBlG,KAAK4X,QAAQ1R,SACbC,oBAEAnG,KAAK6X,SACL7X,KAAK8X;;IAIThY,GAAYsZ;QACV,OAAO,IAAIlB,GACTlY,KAAKuH,MACLvH,KAAK0X,iBACL1X,KAAKmY,GAAgBjS,SACrBlG,KAAK4X,QAAQ1R,SACblG,KAAKmG,OACLnG,KAAKoY,IACLgB,GACApZ,KAAK8X;;IAIThY,GAAUsZ;QACR,OAAO,IAAIlB,GACTlY,KAAKuH,MACLvH,KAAK0X,iBACL1X,KAAKmY,GAAgBjS,SACrBlG,KAAK4X,QAAQ1R,SACblG,KAAKmG,OACLnG,KAAKoY,IACLpY,KAAK6X,SACLuB;;;;;;;WAUJtZ,GAAwByH;QACtB,OAAO,IAAI2Q,GACT3Q;6BACqB,MACrBvH,KAAKmY,GAAgBjS,SACrBlG,KAAK4X,QAAQ1R,SACblG,KAAKmG,OACLnG,KAAKoY,IACLpY,KAAK6X,SACL7X,KAAK8X;;;;;WAQThY;QACE,OAC0B,MAAxBE,KAAK4X,QAAQhS,UACE,SAAf5F,KAAKmG,SACW,QAAhBnG,KAAK6X,WACS,QAAd7X,KAAK8X,UAC4B,MAAhC9X,KAAKmY,GAAgBvS,UACa,MAAhC5F,KAAKmY,GAAgBvS,UACpB5F,KAAKmY,GAAgB,GAAGzF,MAAMiG;;;;;IAOtC7Y;QACE,OAAO,GAAGE,KAAKqZ,KAAWzK,oBAAoB5O,KAAKoY;;IAGrDtY;QACE,OAAO,gBAAgBE,KAAKqZ,KAAW3X,yBACrC1B,KAAKoY;;IAITtY,QAAQkF;QACN,OACEhF,KAAKqZ,KAAW7T,QAAQR,EAAMqU,SAC9BrZ,KAAKoY,OAAcpT,EAAMoT;;IAI7BtY,GAAcwZ,GAAcC;QAC1B,IAAIC,KAAqB;QACzB,KAAK,MAAM7B,KAAW3X,KAAK2X,SAAS;YAClC,MAAM8B,IAAO9B,EAAQvJ,QAAQkL,GAAIC;YACjC,IAAa,MAATE,GACF,OAAOA;YAETD,IAAqBA,KAAsB7B,EAAQjF,MAAMiG;;QAO3D,OAAO;;IAGT7Y,QAAQkV;QACN,OACEhV,KAAK0Z,GAA8B1E,MACnChV,KAAK2Z,GAAe3E,MACpBhV,KAAK4Z,GAAe5E,MACpBhV,KAAK6Z,GAAc7E;;IAIvBlV;QACE,QAAQsK,EAAkBpK,KAAKmG,8BAAUnG,KAAKoY;;IAGhDtY;QACE,QAAQsK,EAAkBpK,KAAKmG,6BAAUnG,KAAKoY;;IAGhDtY;QACE,OAAOE,KAAKmY,GAAgBvS,SAAS,IACjC5F,KAAKmY,GAAgB,GAAGzF,QACxB;;IAGN5S;QACE,KAAK,MAAM4H,KAAU1H,KAAK4X,SACxB,IAAIlQ,aAAkBoS,MAAepS,EAAOqS,MAC1C,OAAOrS,EAAOgL;QAGlB,OAAO;;;;IAKT5S,GAAmBka;QACjB,KAAK,MAAMtS,KAAU1H,KAAK4X,SACxB,IAAIlQ,aAAkBoS,MAChBE,EAAUxS,QAAQE,EAAOuS,OAAO,GAClC,OAAOvS,EAAOuS;QAIpB,OAAO;;IAGTna;QACE,OAAOE,KAAKqZ,KAAWa;;IAGzBpa;QACE,OAAgC,SAAzBE,KAAK0X;;;;;WAOd5X;QACE,KAAKE,KAAKma,IACR,wBAAIna,KAAKoY,IACPpY,KAAKma,KAAiB,IAAI1C,GACxBzX,KAAKuH,MACLvH,KAAK0X,iBACL1X,KAAK2X,SACL3X,KAAK4X,SACL5X,KAAKmG,OACLnG,KAAK6X,SACL7X,KAAK8X,aAEF;;YAEL,MAAMsC,IAAW;YACjB,KAAK,MAAMzC,KAAW3X,KAAK2X,SAAS;gBAClC,MAAMqB,gCACJrB,EAAQqB;gBAGVoB,EAAS9T,KAAK,IAAIsS,GAAQjB,EAAQjF,OAAOsG;;;wBAI3C,MAAMnB,IAAU7X,KAAK8X,QACjB,IAAIuC,GAAMra,KAAK8X,MAAMwC,WAAWta,KAAK8X,MAAMyC,UAC3C,MACEzC,IAAQ9X,KAAK6X,UACf,IAAIwC,GAAMra,KAAK6X,QAAQyC,WAAWta,KAAK6X,QAAQ0C,UAC/C;;YAGJva,KAAKma,KAAiB,IAAI1C,GACxBzX,KAAKuH,MACLvH,KAAK0X,iBACL0C,GACApa,KAAK4X,SACL5X,KAAKmG,OACL0R,GACAC;;QAIN,OAAO9X,KAAKma;;IAGNra,GAA8BkV;QACpC,MAAMwF,IAAUxF,EAAIhM,IAAIzB;QACxB,OAA6B,SAAzBvH,KAAK0X,kBAIL1C,EAAIhM,IAAIyR,GAAgBza,KAAK0X,oBAC7B1X,KAAKuH,KAAK+K,EAAWkI,KAEdjS,EAAY0P,GAAcjY,KAAKuH,QAEjCvH,KAAKuH,KAAK/B,QAAQgV,KAGlBxa,KAAKuH,KAAKmT,EAAoBF;;;;;WAQjC1a,GAAekV;QACrB,KAAK,MAAM2C,KAAW3X,KAAKmY;;QAEzB,KAAKR,EAAQjF,MAAMiG,OAA6C,SAA7B3D,EAAItC,MAAMiF,EAAQjF,QACnD,QAAO;QAGX,QAAO;;IAGD5S,GAAekV;QACrB,KAAK,MAAMtN,KAAU1H,KAAK4X,SACxB,KAAKlQ,EAAOiT,QAAQ3F,IAClB,QAAO;QAGX,QAAO;;;;WAMDlV,GAAckV;QACpB,SAAIhV,KAAK6X,YAAY7X,KAAK6X,QAAQ+C,GAAoB5a,KAAK2X,SAAS3C,SAGhEhV,KAAK8X,UAAS9X,KAAK8X,MAAM8C,GAAoB5a,KAAK2X,SAAS3C;;IAMzDlV,GAAiBsZ;;;MAyBdU;IACXha,YACS4S,GACAuH,GACApY;QAEPJ,SAJOzB,aAAA0S,GACA1S,UAAAia,GACAja,aAAA6B;;;;WAQT/B,cAAc4S,GAAkBuH,GAAcpY;QAC5C,IAAI6Q,EAAMiG,KACR,yBAAIsB,IASK,IAAIY,GAAiBnI,GAAO7Q,KAU5B,IAAIiZ,GAAepI,GAAOuH,GAAIpY;QAElC,IAAI0O,EAAY1O,IAAQ;YAC7B,yBAAIoY,GACF,MAAM,IAAI5Y,EACRlB,EAAKI,kBACL;YAGJ,OAAO,IAAIuZ,GAAYpH,GAAOuH,GAAIpY;;QAC7B,IAAI2O,GAAW3O,IAAQ;YAC5B,yBAAIoY,GACF,MAAM,IAAI5Y,EACRlB,EAAKI,kBACL;YAGJ,OAAO,IAAIuZ,GAAYpH,GAAOuH,GAAIpY;;QAC7B,iDAAIoY,IACF,IAAIc,GAAoBrI,GAAO7Q,uBAC7BoY,IAKF,IAAIe,GAAStI,GAAO7Q,uDAClBoY,IAKF,IAAIgB,GAAuBvI,GAAO7Q,KAElC,IAAIiY,GAAYpH,GAAOuH,GAAIpY;;IAItC/B,QAAQkV;QACN,MAAMhQ,IAAQgQ,EAAItC,MAAM1S,KAAK0S;;gBAG7B,OACY,SAAV1N,KACAgG,EAAUhL,KAAK6B,WAAWmJ,EAAUhG,MACpChF,KAAKkb,GAAkBlO,EAAahI,GAAOhF,KAAK6B;;IAI1C/B,GAAkBiO;QAC1B,QAAQ/N,KAAKia;UACX;YACE,OAAOlM,IAAa;;UACtB;YACE,OAAOA,KAAc;;UACvB;YACE,OAAsB,MAAfA;;UACT;YACE,OAAOA,IAAa;;UACtB;YACE,OAAOA,KAAc;;UACvB;YACE,OA7hBDjI;;;IAiiBLhG;QACE,OACE,oHAKE0H,QAAQxH,KAAKia,OAAO;;IAI1Bna;;;;QAIE,OACEE,KAAK0S,MAAMpL,MACXtH,KAAKia,GAAGvY,aACRkN,EAAY5O,KAAK6B;;IAIrB/B,QAAQkF;QACN,OAAIA,aAAiB8U,OAEjB9Z,KAAKia,OAAOjV,EAAMiV,MAClBja,KAAK0S,MAAMlN,QAAQR,EAAM0N,UACzBzH,EAAYjL,KAAK6B,OAAOmD,EAAMnD;;IAOpC/B;QACE,OAAO,GAAGE,KAAK0S,MAAMpL,OAAqBtH,KAAKia,MAAMrL,EACnD5O,KAAK6B;;;;kEAMEiZ,WAAuBhB;IAGlCha,YAAY4S,GAAkBuH,GAAcpY;QAC1CJ,MAAMiR,GAAOuH,GAAIpY,IAKjB7B,KAAKgJ,MAAMT,EAAY0G,EAASpN,EAAM6J;;IAGxC5L,QAAQkV;QACN,MAAMjH,IAAaxF,EAAYvC,EAAWgP,EAAIhM,KAAKhJ,KAAKgJ;QACxD,OAAOhJ,KAAKkb,GAAkBnN;;;;gEAKrB8M,WAAyBf;IAGpCha,YAAY4S,GAAkB7Q;QAC5BJ,MAAMiR,mBAAoB7Q,IAE1B7B,KAAKuO,QAAQ1M,EAAMyK,WAAWC,UAAU,IAAIzE,IAAIiF,KAKvCxE,EAAY0G,EAASlC,EAAErB;;IAIlC5L,QAAQkV;QACN,OAAOhV,KAAKuO,KAAK8C,KAAKrI,KAAOA,EAAIxD,QAAQwP,EAAIhM;;;;mEAKpC+R,WAA4BjB;IACvCha,YAAY4S,GAAkB7Q;QAC5BJ,MAAMiR,2CAAgC7Q;;IAGxC/B,QAAQkV;QACN,MAAMhQ,IAAQgQ,EAAItC,MAAM1S,KAAK0S;QAC7B,OAAOpC,EAAQtL,MAAU2H,EAAmB3H,EAAMsH,YAAYtM,KAAK6B;;;;uDAK1DmZ,WAAiBlB;IAC5Bha,YAAY4S,GAAkB7Q;QAC5BJ,MAAMiR,mBAAoB7Q;;IAI5B/B,QAAQkV;QACN,MAAMhQ,IAAQgQ,EAAItC,MAAM1S,KAAK0S;QAC7B,OAAiB,SAAV1N,KAAkB2H,EAAmB3M,KAAK6B,MAAiB,YAAEmD;;;;uEAK3DiW,WAA+BnB;IAC1Cha,YAAY4S,GAAkB7Q;QAC5BJ,MAAMiR,mDAAoC7Q;;IAI5C/B,QAAQkV;QACN,MAAMhQ,IAAQgQ,EAAItC,MAAM1S,KAAK0S;QAC7B,UAAKpC,EAAQtL,OAAWA,EAAMsH,WAAWC,WAGlCvH,EAAMsH,WAAWC,OAAO8E,KAAK8J,KAClCxO,EAAmB3M,KAAK6B,MAAiB,YAAEsZ;;;;;;;;;;;;;;;;;UA2BpCd;IACXva,YAAqBwa,GAAgCC;QAAhCva,gBAAAsa,GAAgCta,cAAAua;;IAErDza;;QAEE,OAAO,GAAGE,KAAKua,SAAS,MAAM,OAAOva,KAAKsa,SACvCxS,IAAIsT,KAAKxM,EAAYwM,IACrB/T,KAAK;;;;;WAOVvH,GAAoB6X,GAAoB3C;QAKtC,IAAIjH,IAAa;QACjB,KAAK,IAAIrH,IAAI,GAAGA,IAAI1G,KAAKsa,SAAS1U,QAAQc,KAAK;YAC7C,MAAM2U,IAAmB1D,EAAQjR,IAC3B4U,IAAYtb,KAAKsa,SAAS5T;YAChC,IAAI2U,EAAiB3I,MAAMiG,KAKzB5K,IAAaxF,EAAYvC,EACvBuC,EAAY0G,EAASqM,EAAU5P,iBAC/BsJ,EAAIhM,WAED;gBAML+E,IAAaf,EAAasO,GALTtG,EAAItC,MAAM2I,EAAiB3I;;YAU9C,gCAHI2I,EAAiBrC,QACnBjL,MAA2B,IAEV,MAAfA,GACF;;QAGJ,OAAO/N,KAAKua,SAASxM,KAAc,IAAIA,IAAa;;IAGtDjO,QAAQkF;QACN,IAAc,SAAVA,GACF,QAAO;QAET,IACEhF,KAAKua,WAAWvV,EAAMuV,UACtBva,KAAKsa,SAAS1U,WAAWZ,EAAMsV,SAAS1U,QAExC,QAAO;QAET,KAAK,IAAIc,IAAI,GAAGA,IAAI1G,KAAKsa,SAAS1U,QAAQc,KAAK;YAG7C,KAAKuE,EAFgBjL,KAAKsa,SAAS5T,IACb1B,EAAMsV,SAAS5T,KAEnC,QAAO;;QAGX,QAAO;;;;;;UAOEkS;IAIX9Y,YAAqB4S,GAAkBsG;QAAlBhZ,aAAA0S,QACP7M,MAARmT,MACFA,4BAEFhZ,KAAKgZ,MAAMA,GACXhZ,KAAKub,KAAe7I,EAAMiG;;IAG5B7Y,QAAQwZ,GAAcC;QACpB,MAAMxL,IAAa/N,KAAKub,KACpBhT,EAAYvC,EAAWsT,EAAGtQ,KAAKuQ,EAAGvQ,gBF1rBxC0J,GACA4G,GACAC;YAEA,MAAMiC,IAAKlC,EAAG5G,MAAMA,IACd+I,IAAKlC,EAAG7G,MAAMA;YACpB,OAAW,SAAP8I,KAAsB,SAAPC,IACVzO,EAAawO,GAAIC,KA5FnB3V;SEgxBD4V,CAAwB1b,KAAK0S,OAAO4G,GAAIC;QAC5C,QAAQvZ,KAAKgZ;UACX;YACE,OAAOjL;;UACT;YACE,QAAQ,IAAIA;;UACd;YACE,OAnxBDjI;;;IAuxBLhG;;QAEE,OAAOE,KAAK0S,MAAMpL,MAAoBtH,KAAKgZ,IAAItX;;IAGjD5B;QACE,OAAO,GAAGE,KAAK0S,MAAMpL,QAAsBtH,KAAKgZ;;IAGlDlZ,QAAQkF;QACN,OAAOhF,KAAKgZ,QAAQhU,EAAMgU,OAAOhZ,KAAK0S,MAAMlN,QAAQR,EAAM0N;;;;;;;;;;;;;;;;;;;;;;UCzxBjDiJ;IACX7b;;IAEW8b;;;;;IAKAC;;IAEAC;;;;;IAKAC;;IAEAC,IAAmC3W,EAAgB2B;;;;UAKnDiV,IAAgD5W,EAAgB2B;;;;;;UAOhEkV,IAA0B5S,EAAW6S;QA1BrCnc,cAAA4b,GAKA5b,gBAAA6b,aAEAC,GAKA9b,sBAAA+b,aAEAC;QAKAhc,oCAAAic,GAOAjc,mBAAAkc;;kFAIXpc,GAAmBic;QACjB,OAAO,IAAIJ,GACT3b,KAAK4b,QACL5b,KAAK6b,UACL7b,KAAK8b,IACLC,GACA/b,KAAKgc,IACLhc,KAAKic,8BACLjc,KAAKkc;;;;;WAQTpc,GACEoc,GACAF;QAEA,OAAO,IAAIL,GACT3b,KAAK4b,QACL5b,KAAK6b,UACL7b,KAAK8b,IACL9b,KAAK+b,gBACLC,GACAhc,KAAKic,8BACLC;;;;;WAQJpc,GACEmc;QAEA,OAAO,IAAIN,GACT3b,KAAK4b,QACL5b,KAAK6b,UACL7b,KAAK8b,IACL9b,KAAK+b,gBACL/b,KAAKgc,IACLC,GACAjc,KAAKkc;;;;;;;;;;;;;;;;;;;UCpGEE;;IAEXtc,YAAmBiJ;QAAA/I,aAAA+I;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GCYrB,KAAKsT;;;;;;;;SA0BWC,GAAiB/a;IAC/B,QAAQA;MACN,KAAKpB,EAAKC;QACR,OAnCwF0F;;MAoC1F,KAAK3F,EAAKE;MACV,KAAKF,EAAKG;MACV,KAAKH,EAAKK;MACV,KAAKL,EAAKU;MACV,KAAKV,EAAKe;MACV,KAAKf,EAAKgB;;;cAGV,KAAKhB,EAAKS;QACR,QAAO;;MACT,KAAKT,EAAKI;MACV,KAAKJ,EAAKM;MACV,KAAKN,EAAKO;MACV,KAAKP,EAAKQ;MACV,KAAKR,EAAKW;;;;cAIV,KAAKX,EAAKY;MACV,KAAKZ,EAAKa;MACV,KAAKb,EAAKc;MACV,KAAKd,EAAKiB;QACR,QAAO;;MACT;QACE,OA5DwF0E;;;;;;;;;;;;;;;;;;;;;;;SAwG9EyW,GAAmBhb;IACjC,SAAasE,MAATtE;;;IAIF,OADAib,GAAS,4BACFrc,EAAKG;IAGd,QAAQiB;MACN,KAAK8a,GAAQjc;QACX,OAAOD,EAAKC;;MACd,KAAKic,GAAQhc;QACX,OAAOF,EAAKE;;MACd,KAAKgc,GAAQ/b;QACX,OAAOH,EAAKG;;MACd,KAAK+b,GAAQ7b;QACX,OAAOL,EAAKK;;MACd,KAAK6b,GAAQxb;QACX,OAAOV,EAAKU;;MACd,KAAKwb,GAAQnb;QACX,OAAOf,EAAKe;;MACd,KAAKmb,GAAQlb;QACX,OAAOhB,EAAKgB;;MACd,KAAKkb,GAAQzb;QACX,OAAOT,EAAKS;;MACd,KAAKyb,GAAQ9b;QACX,OAAOJ,EAAKI;;MACd,KAAK8b,GAAQ5b;QACX,OAAON,EAAKM;;MACd,KAAK4b,GAAQ3b;QACX,OAAOP,EAAKO;;MACd,KAAK2b,GAAQ1b;QACX,OAAOR,EAAKQ;;MACd,KAAK0b,GAAQvb;QACX,OAAOX,EAAKW;;MACd,KAAKub,GAAQtb;QACX,OAAOZ,EAAKY;;MACd,KAAKsb,GAAQrb;QACX,OAAOb,EAAKa;;MACd,KAAKqb,GAAQpb;QACX,OAAOd,EAAKc;;MACd,KAAKob,GAAQjb;QACX,OAAOjB,EAAKiB;;MACd;QACE,OApJwF0E;;;;;;;;;;;UAMzFuW,OAAAA,6BAEHI;AACAA,gCACAA;AACAA,oDACAA;AACAA,8CACAA;AACAA,iDACAA;AACAA,wDACAA;AACAA,2CACAA;AACAA,mCACAA,yCACAA;;;;;;;;;;;;;;;;;;;;MCNWC;IAIX5c,YACSkG,GACP2W;iBADO3W,GAGPhG,KAAK2c,OAAOA,KAAcC,GAASC;;;IAIrC/c,GAAOkJ,GAAQnH;QACb,OAAO,IAAI6a,GACT1c,KAAKgG,GACLhG,KAAK2c,KACFG,GAAO9T,GAAKnH,GAAO7B,KAAKgG,GACxB+W,GAAK,MAAM,MAAMH,GAASI,IAAO,MAAM;;;IAK9Cld,OAAOkJ;QACL,OAAO,IAAI0T,GACT1c,KAAKgG,GACLhG,KAAK2c,KACFM,OAAOjU,GAAKhJ,KAAKgG,GACjB+W,GAAK,MAAM,MAAMH,GAASI,IAAO,MAAM;;;IAK9Cld,IAAIkJ;QACF,IAAIkU,IAAOld,KAAK2c;QAChB,OAAQO,EAAK7T,OAAW;YACtB,MAAM8T,IAAMnd,KAAKgG,EAAWgD,GAAKkU,EAAKlU;YACtC,IAAY,MAARmU,GACF,OAAOD,EAAKrb;YACHsb,IAAM,IACfD,IAAOA,EAAKjW,OACHkW,IAAM,MACfD,IAAOA,EAAKhW;;QAGhB,OAAO;;;;IAKTpH,QAAQkJ;;QAEN,IAAIoU,IAAc,GACdF,IAAOld,KAAK2c;QAChB,OAAQO,EAAK7T,OAAW;YACtB,MAAM8T,IAAMnd,KAAKgG,EAAWgD,GAAKkU,EAAKlU;YACtC,IAAY,MAARmU,GACF,OAAOC,IAAcF,EAAKjW,KAAKT;YACtB2W,IAAM,IACfD,IAAOA,EAAKjW;;YAGZmW,KAAeF,EAAKjW,KAAKT,OAAO,GAChC0W,IAAOA,EAAKhW;;;gBAIhB,QAAQ;;IAGVpH;QACE,OAAOE,KAAK2c,KAAKtT;;;IAInB7C;QACE,OAAOxG,KAAK2c,KAAKnW;;;IAInB1G;QACE,OAAOE,KAAK2c,KAAKU;;;IAInBvd;QACE,OAAOE,KAAK2c,KAAKW;;;;;;IAOnBxd,GAAoByd;QAClB,OAAQvd,KAAK2c,KAAwBa,GAAiBD;;IAGxDzd,QAAQ8G;QACN5G,KAAKwd,GAAiB,CAACC,GAAG1Q,OACxBnG,EAAG6W,GAAG1Q,KACC;;IAIXjN;QACE,MAAM4d,IAAyB;QAK/B,OAJA1d,KAAKwd,GAAiB,CAACC,GAAG1Q,OACxB2Q,EAAapX,KAAK,GAAGmX,KAAK1Q,OACnB,KAEF,IAAI2Q,EAAarW,KAAK;;;;;;;IAQ/BvH,GAAoByd;QAClB,OAAQvd,KAAK2c,KAAwBgB,GAAiBJ;;;IAIxDzd;QACE,OAAO,IAAI8d,GAAwB5d,KAAK2c,MAAM,MAAM3c,KAAKgG,IAAY;;IAGvElG,GAAgBkJ;QACd,OAAO,IAAI4U,GAAwB5d,KAAK2c,MAAM3T,GAAKhJ,KAAKgG,IAAY;;IAGtElG;QACE,OAAO,IAAI8d,GAAwB5d,KAAK2c,MAAM,MAAM3c,KAAKgG,IAAY;;IAGvElG,GAAuBkJ;QACrB,OAAO,IAAI4U,GAAwB5d,KAAK2c,MAAM3T,GAAKhJ,KAAKgG,IAAY;;;;;;MAK3D4X;IAIX9d,YACEod,GACAW,GACA7X,GACA8X;QAEA9d,KAAK8d,KAAYA,GACjB9d,KAAK+d,KAAY;QAEjB,IAAIZ,IAAM;QACV,OAAQD,EAAK7T,OAOX,IANA8T,IAAMU,IAAW7X,EAAWkX,EAAKlU,KAAK6U,KAAY;;QAE9CC,MACFX,MAAQ,IAGNA,IAAM;;QAGND,IADEld,KAAK8d,KACAZ,EAAKjW,OAELiW,EAAKhW,YAET;YAAA,IAAY,MAARiW,GAAW;;;gBAGpBnd,KAAK+d,GAAUzX,KAAK4W;gBACpB;;;;YAIAld,KAAK+d,GAAUzX,KAAK4W,IAElBA,IADEld,KAAK8d,KACAZ,EAAKhW,QAELgW,EAAKjW;;;IAMpBnH;QAME,IAAIod,IAAOld,KAAK+d,GAAUC;QAC1B,MAAM7O,IAAS;YAAEnG,KAAKkU,EAAKlU;YAAKnH,OAAOqb,EAAKrb;;QAE5C,IAAI7B,KAAK8d,IAEP,KADAZ,IAAOA,EAAKjW,OACJiW,EAAK7T,OACXrJ,KAAK+d,GAAUzX,KAAK4W,IACpBA,IAAOA,EAAKhW,YAId,KADAgW,IAAOA,EAAKhW,QACJgW,EAAK7T,OACXrJ,KAAK+d,GAAUzX,KAAK4W;QACpBA,IAAOA,EAAKjW;QAIhB,OAAOkI;;IAGTrP;QACE,OAAOE,KAAK+d,GAAUnY,SAAS;;IAGjC9F;QACE,IAA8B,MAA1BE,KAAK+d,GAAUnY,QACjB,OAAO;QAGT,MAAMsX,IAAOld,KAAK+d,GAAU/d,KAAK+d,GAAUnY,SAAS;QACpD,OAAO;YAAEoD,KAAKkU,EAAKlU;YAAKnH,OAAOqb,EAAKrb;;;;;;;MAK3B+a;IAaX9c,YACSkJ,GACAnH,GACPoc,GACAhX,GACAC;QAJOlH,WAAAgJ,GACAhJ,aAAA6B,GAKP7B,KAAKie,QAAiB,QAATA,IAAgBA,IAAQrB,GAASsB,KAC9Cle,KAAKiH,OAAe,QAARA,IAAeA,IAAO2V,GAASC;QAC3C7c,KAAKkH,QAAiB,QAATA,IAAgBA,IAAQ0V,GAASC,OAC9C7c,KAAKwG,OAAOxG,KAAKiH,KAAKT,OAAO,IAAIxG,KAAKkH,MAAMV;;;IAI9C1G,GACEkJ,GACAnH,GACAoc,GACAhX,GACAC;QAEA,OAAO,IAAI0V,GACF,QAAP5T,IAAcA,IAAMhJ,KAAKgJ,KAChB,QAATnH,IAAgBA,IAAQ7B,KAAK6B,OACpB,QAAToc,IAAgBA,IAAQje,KAAKie,OACrB,QAARhX,IAAeA,IAAOjH,KAAKiH,MAClB,QAATC,IAAgBA,IAAQlH,KAAKkH;;IAIjCpH;QACE,QAAO;;;;;;IAOTA,GAAoByd;QAClB,OACGvd,KAAKiH,KAAwBuW,GAAiBD,MAC/CA,EAAOvd,KAAKgJ,KAAKhJ,KAAK6B,UACrB7B,KAAKkH,MAAyBsW,GAAiBD;;;;;;IAQpDzd,GAAoByd;QAClB,OACGvd,KAAKkH,MAAyByW,GAAiBJ,MAChDA,EAAOvd,KAAKgJ,KAAKhJ,KAAK6B,UACrB7B,KAAKiH,KAAwB0W,GAAiBJ;;;IAK3Czd;QACN,OAAIE,KAAKiH,KAAKoC,MACLrJ,OAECA,KAAKiH,KAAwBD;;;IAKzClH;QACE,OAAOE,KAAKgH,MAAMgC;;;IAIpBlJ;QACE,OAAIE,KAAKkH,MAAMmC,MACNrJ,KAAKgJ,MAELhJ,KAAKkH,MAAMoW;;;IAKtBxd,GAAOkJ,GAAQnH,GAAUmE;QACvB,IAAImY,IAAoBne;QACxB,MAAMmd,IAAMnX,EAAWgD,GAAKmV,EAAEnV;QAc9B,OAZEmV,IADEhB,IAAM,IACJgB,EAAEpB,GAAK,MAAM,MAAM,MAAMoB,EAAElX,KAAK6V,GAAO9T,GAAKnH,GAAOmE,IAAa,QACnD,MAARmX,IACLgB,EAAEpB,GAAK,MAAMlb,GAAO,MAAM,MAAM,QAEhCsc,EAAEpB,GACJ,MACA,MACA,MACA,MACAoB,EAAEjX,MAAM4V,GAAO9T,GAAKnH,GAAOmE;QAGxBmY,EAAEC;;IAGHte;QACN,IAAIE,KAAKiH,KAAKoC,KACZ,OAAOuT,GAASC;QAElB,IAAIsB,IAAoBne;QAKxB,OAJKme,EAAElX,KAAKoX,QAAYF,EAAElX,KAAKA,KAAKoX,SAClCF,IAAIA,EAAEG,OAERH,IAAIA,EAAEpB,GAAK,MAAM,MAAM,MAAOoB,EAAElX,KAAwBsX,MAAa;QAC9DJ,EAAEC;;;IAIXte,OACEkJ,GACAhD;QAEA,IAAIwY,GACAL,IAAoBne;QACxB,IAAIgG,EAAWgD,GAAKmV,EAAEnV,OAAO,GACtBmV,EAAElX,KAAKoC,OAAc8U,EAAElX,KAAKoX,QAAYF,EAAElX,KAAKA,KAAKoX,SACvDF,IAAIA,EAAEG;QAERH,IAAIA,EAAEpB,GAAK,MAAM,MAAM,MAAMoB,EAAElX,KAAKgW,OAAOjU,GAAKhD,IAAa,YACxD;YAOL,IANImY,EAAElX,KAAKoX,SACTF,IAAIA,EAAEM,OAEHN,EAAEjX,MAAMmC,OAAc8U,EAAEjX,MAAMmX,QAAYF,EAAEjX,MAAMD,KAAKoX,SAC1DF,IAAIA,EAAEO;YAEuB,MAA3B1Y,EAAWgD,GAAKmV,EAAEnV,MAAY;gBAChC,IAAImV,EAAEjX,MAAMmC,KACV,OAAOuT,GAASC;gBAEhB2B,IAAYL,EAAEjX,MAAyBF,OACvCmX,IAAIA,EAAEpB,GACJyB,EAASxV,KACTwV,EAAS3c,OACT,MACA,MACCsc,EAAEjX,MAAyBqX;;YAIlCJ,IAAIA,EAAEpB,GAAK,MAAM,MAAM,MAAM,MAAMoB,EAAEjX,MAAM+V,OAAOjU,GAAKhD;;QAEzD,OAAOmY,EAAEC;;IAGXte;QACE,OAAOE,KAAKie;;;IAINne;QACN,IAAIqe,IAAoBne;QAUxB,OATIme,EAAEjX,MAAMmX,SAAYF,EAAElX,KAAKoX,SAC7BF,IAAIA,EAAEQ,OAEJR,EAAElX,KAAKoX,QAAWF,EAAElX,KAAKA,KAAKoX,SAChCF,IAAIA,EAAEM;QAEJN,EAAElX,KAAKoX,QAAWF,EAAEjX,MAAMmX,SAC5BF,IAAIA,EAAES,OAEDT;;IAGDre;QACN,IAAIqe,IAAIne,KAAK4e;QAYb,OAXIT,EAAEjX,MAAMD,KAAKoX,SACfF,IAAIA,EAAEpB,GACJ,MACA,MACA,MACA,MACCoB,EAAEjX,MAAyBuX,OAE9BN,IAAIA,EAAEQ;QACNR,IAAIA,EAAES,OAEDT;;IAGDre;QACN,IAAIqe,IAAIne,KAAK4e;QAKb,OAJIT,EAAElX,KAAKA,KAAKoX,SACdF,IAAIA,EAAEM,MACNN,IAAIA,EAAES,OAEDT;;IAGDre;QACN,MAAM+e,IAAK7e,KAAK+c,GAAK,MAAM,MAAMH,GAASsB,KAAK,MAAMle,KAAKkH,MAAMD;QAChE,OAAQjH,KAAKkH,MAAyB6V,GACpC,MACA,MACA/c,KAAKie,OACLY,GACA;;IAII/e;QACN,MAAMgf,IAAK9e,KAAK+c,GAAK,MAAM,MAAMH,GAASsB,KAAKle,KAAKiH,KAAKC,OAAO;QAChE,OAAQlH,KAAKiH,KAAwB8V,GAAK,MAAM,MAAM/c,KAAKie,OAAO,MAAMa;;IAGlEhf;QACN,MAAMmH,IAAOjH,KAAKiH,KAAK8V,GAAK,MAAM,OAAO/c,KAAKiH,KAAKgX,OAAO,MAAM,OAC1D/W,IAAQlH,KAAKkH,MAAM6V,GAAK,MAAM,OAAO/c,KAAKkH,MAAM+W,OAAO,MAAM;QACnE,OAAOje,KAAK+c,GAAK,MAAM,OAAO/c,KAAKie,OAAOhX,GAAMC;;;IAIlDpH;QACE,MAAMif,IAAa/e,KAAKgf;QACxB,OAAIna,KAAKoa,IAAI,GAAKF,MAAe/e,KAAKwG,OAAO;;;;IASrC1G;QACR,IAAIE,KAAKqe,QAAWre,KAAKiH,KAAKoX,MAC5B,MAveevY;QAyejB,IAAI9F,KAAKkH,MAAMmX,MACb,MA1eevY;QA4ejB,MAAMiZ,IAAc/e,KAAKiH,KAAwB+X;QACjD,IAAID,MAAgB/e,KAAKkH,MAAyB8X,MAChD,MA9eelZ;QAgff,OAAOiZ,KAAc/e,KAAKqe,OAAU,IAAI;;;;;;8DArPrCzB;WAAiC,MAEjCA,UAAM,GACNA,SAAQ;;;AAiUjBA,GAASC,QAAQ;;;IAzEjB/c;QAgBEE,YAAO;;IAfPgJ;QACE,MAxfiBlD;;IA0fnBjE;QACE,MA3fiBiE;;IA6fnBmY;QACE,MA9fiBnY;;IAggBnBmB;QACE,MAjgBiBnB;;IAmgBnBoB;QACE,MApgBiBpB;;;IAygBnBhG,GACEkJ,GACAnH,GACAoc,GACAhX,GACAC;QAEA,OAAOlH;;;IAITF,GAAOkJ,GAAQnH,GAAUmE;QACvB,OAAO,IAAI4W,GAAe5T,GAAKnH;;;IAIjC/B,OAAOkJ,GAAQhD;QACb,OAAOhG;;IAGTF;QACE,QAAO;;IAGTA,GAAiByd;QACf,QAAO;;IAGTzd,GAAiByd;QACf,QAAO;;IAGTzd;QACE,OAAO;;IAGTA;QACE,OAAO;;IAGTA;QACE,QAAO;;;IAITA;QACE,QAAO;;IAGCA;QACR,OAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;MC3jBEof;IAGXpf,YAAoBkG;iBAAAA,GAClBhG,KAAKgU,OAAO,IAAI0I,GAAsB1c,KAAKgG;;IAG7ClG,IAAIqf;QACF,OAA+B,SAAxBnf,KAAKgU,KAAKjR,IAAIoc;;IAGvBrf;QACE,OAAOE,KAAKgU,KAAKqJ;;IAGnBvd;QACE,OAAOE,KAAKgU,KAAKsJ;;IAGnB9W;QACE,OAAOxG,KAAKgU,KAAKxN;;IAGnB1G,QAAQqf;QACN,OAAOnf,KAAKgU,KAAKxM,QAAQ2X;;iEAI3Brf,QAAQsf;QACNpf,KAAKgU,KAAKwJ,GAAiB,CAACC,GAAM1Q,OAChCqS,EAAG3B,KACI;;4EAKX3d,GAAeuf,GAAeD;QAC5B,MAAME,IAAOtf,KAAKgU,KAAKuL,GAAgBF,EAAM;QAC7C,MAAOC,EAAKE,QAAW;YACrB,MAAML,IAAOG,EAAKG;YAClB,IAAIzf,KAAKgG,EAAWmZ,EAAKnW,KAAKqW,EAAM,OAAO,GACzC;YAEFD,EAAGD,EAAKnW;;;;;WAOZlJ,GAAasf,GAA0BM;QACrC,IAAIJ;QAMJ,KAJEA,SADYzZ,MAAV6Z,IACK1f,KAAKgU,KAAKuL,GAAgBG,KAE1B1f,KAAKgU,KAAK2L,MAEZL,EAAKE,QAAW;YAGrB,KADeJ,EADFE,EAAKG,KACKzW,MAErB;;;uEAMNlJ,GAAkBqf;QAChB,MAAMG,IAAOtf,KAAKgU,KAAKuL,GAAgBJ;QACvC,OAAOG,EAAKE,OAAYF,EAAKG,KAAUzW,MAAM;;IAG/ClJ;QACE,OAAO,IAAI8f,GAAqB5f,KAAKgU,KAAK2L;;IAG5C7f,GAAgBkJ;QACd,OAAO,IAAI4W,GAAqB5f,KAAKgU,KAAKuL,GAAgBvW;;4CAI5DlJ,IAAIqf;QACF,OAAOnf,KAAK+c,GAAK/c,KAAKgU,KAAKiJ,OAAOkC,GAAMrC,GAAOqC,IAAM;;iCAIvDrf,OAAOqf;QACL,OAAKnf,KAAK6f,IAAIV,KAGPnf,KAAK+c,GAAK/c,KAAKgU,KAAKiJ,OAAOkC,MAFzBnf;;IAKXF;QACE,OAAOE,KAAKgU,KAAK3K;;IAGnBvJ,GAAUkF;QACR,IAAImK,IAAuBnP;;gBAW3B,OARImP,EAAO3I,OAAOxB,EAAMwB,SACtB2I,IAASnK,GACTA,IAAQhF,OAGVgF,EAAMoB,QAAQ+Y;YACZhQ,IAASA,EAAO2Q,IAAIX;YAEfhQ;;IAGTrP,QAAQkF;QACN,MAAMA,aAAiBka,KACrB,QAAO;QAET,IAAIlf,KAAKwG,SAASxB,EAAMwB,MACtB,QAAO;QAGT,MAAMuZ,IAAS/f,KAAKgU,KAAK2L,MACnBK,IAAUhb,EAAMgP,KAAK2L;QAC3B,MAAOI,EAAOP,QAAW;YACvB,MAAMS,IAAWF,EAAON,KAAUzW,KAC5BkX,IAAYF,EAAQP,KAAUzW;YACpC,IAA6C,MAAzChJ,KAAKgG,EAAWia,GAAUC,IAC5B,QAAO;;QAGX,QAAO;;IAGTpgB;QACE,MAAMqgB,IAAW;QAIjB,OAHAngB,KAAKoG,QAAQyV;YACXsE,EAAI7Z,KAAKuV;YAEJsE;;IAGTrgB;QACE,MAAMqP,IAAc;QAEpB,OADAnP,KAAKoG,QAAQ+Y,KAAQhQ,EAAO7I,KAAK6Y,KAC1B,eAAehQ,EAAOzN,aAAa;;IAGpC5B,GAAKkU;QACX,MAAM7E,IAAS,IAAI+P,GAAUlf,KAAKgG;QAElC,OADAmJ,EAAO6E,OAAOA,GACP7E;;;;MAIEyQ;IACX9f,YAAoBwf;kBAAAA;;IAEpBxf;QACE,OAAOE,KAAKsf,GAAKG,KAAUzW;;IAG7BlJ;QACE,OAAOE,KAAKsf,GAAKE;;;;;;;;;;;;;;;;;;;GC1JrB,OAAMY,KAA2B,IAAI1D,GACnCnU,EAAYvC;;SAEEqa;IACd,OAAOD;;;SAQOE;IACd,OAAOD;;;AAST,MAAME,KAAqB,IAAI7D,GAC7BnU,EAAYvC;;SAEEwa;IACd,OAAOD;;;AAIT,MAAME,KAA6B,IAAI/D,GACrCnU,EAAYvC;;SAEE0a;IACd,OAAOD;;;AAIT,MAAME,KAAyB,IAAIzB,GAAU3W,EAAYvC;;SACzC4a,MAAkBrS;IAChC,IAAIoG,IAAMgM;IACV,KAAK,MAAM3X,KAAOuF,GAChBoG,IAAMA,EAAImL,IAAI9W;IAEhB,OAAO2L;;;AAIT,MAAMkM,KAAsB,IAAI3B,GAAoBja;;SACpC6b;IACd,OAAOD;;;;;;;;;;;;;;;;;;;;;;;;UCpDIE;;IAcXjhB,YAAY2Z;;;QAIRzZ,KAAKgG,IADHyT,IACgB,CAACH,GAAcC,MAC/BE,EAAKH,GAAIC,MAAOhR,EAAYvC,EAAWsT,EAAGtQ,KAAKuQ,EAAGvQ,OAElC,CAACsQ,GAAcC,MAC/BhR,EAAYvC,EAAWsT,EAAGtQ,KAAKuQ,EAAGvQ;QAGtChJ,KAAKghB,KAAWR,MAChBxgB,KAAKihB,KAAY,IAAIvE,GAA0B1c,KAAKgG;;;;;WArBtDlG,UAAgBohB;QACd,OAAO,IAAIH,GAAYG,EAAOlb;;IAuBhClG,IAAIkJ;QACF,OAAiC,QAA1BhJ,KAAKghB,GAASje,IAAIiG;;IAG3BlJ,IAAIkJ;QACF,OAAOhJ,KAAKghB,GAASje,IAAIiG;;IAG3BlJ;QACE,OAAOE,KAAKihB,GAAU5D;;IAGxBvd;QACE,OAAOE,KAAKihB,GAAU3D;;IAGxBxd;QACE,OAAOE,KAAKihB,GAAU5X;;;;;WAOxBvJ,QAAQkJ;QACN,MAAMgM,IAAMhV,KAAKghB,GAASje,IAAIiG;QAC9B,OAAOgM,IAAMhV,KAAKihB,GAAUzZ,QAAQwN,MAAQ;;IAG9CxO;QACE,OAAOxG,KAAKihB,GAAUza;;kEAIxB1G,QAAQsf;QACNpf,KAAKihB,GAAUzD,GAAiB,CAACC,GAAG1Q,OAClCqS,EAAG3B,KACI;;8DAKX3d,IAAIkV;;QAEF,MAAML,IAAM3U,KAAK4U,OAAOI,EAAIhM;QAC5B,OAAO2L,EAAIoI,GACTpI,EAAIqM,GAASlE,GAAO9H,EAAIhM,KAAKgM,IAC7BL,EAAIsM,GAAUnE,GAAO9H,GAAK;;kDAK9BlV,OAAOkJ;QACL,MAAMgM,IAAMhV,KAAK+C,IAAIiG;QACrB,OAAKgM,IAIEhV,KAAK+c,GAAK/c,KAAKghB,GAAS/D,OAAOjU,IAAMhJ,KAAKihB,GAAUhE,OAAOjI,MAHzDhV;;IAMXF,QAAQkF;QACN,MAAMA,aAAiB+b,KACrB,QAAO;QAET,IAAI/gB,KAAKwG,SAASxB,EAAMwB,MACtB,QAAO;QAGT,MAAMuZ,IAAS/f,KAAKihB,GAAUtB,MACxBK,IAAUhb,EAAMic,GAAUtB;QAChC,MAAOI,EAAOP,QAAW;YACvB,MAAM2B,IAAUpB,EAAON,KAAUzW,KAC3BoY,IAAWpB,EAAQP,KAAUzW;YACnC,KAAKmY,EAAQ3b,QAAQ4b,IACnB,QAAO;;QAGX,QAAO;;IAGTthB;QACE,MAAMuhB,IAAuB;QAI7B,OAHArhB,KAAKoG,QAAQ4O;YACXqM,EAAW/a,KAAK0O,EAAItT;YAEI,MAAtB2f,EAAWzb,SACN,mBAEA,sBAAsByb,EAAWha,KAAK,UAAU;;IAInDvH,GACNkhB,GACAC;QAEA,MAAMK,IAAS,IAAIP;QAInB,OAHAO,EAAOtb,IAAahG,KAAKgG,GACzBsb,EAAON,KAAWA,GAClBM,EAAOL,KAAYA,GACZK;;;;;;;;;;;;;;;;;;;;;;;UClHEC;IAAbzhB;QACEE,UAAoB,IAAI0c,GACtBnU,EAAYvC;;IAGdlG,MAAM0hB;QACJ,MAAMxY,IAAMwY,EAAOxM,IAAIhM,KACjByY,IAAYzhB,KAAK0hB,GAAU3e,IAAIiG;QAChCyY;;0BAOHD,EAAOG,6BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAKwY,0BAE5CA,EAAOG,4BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAK;YAC1C2Y,MAAMF,EAAUE;YAChB3M,KAAKwM,EAAOxM;kCAGdwM,EAAOG,6BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAK;YAC1C2Y;YACA3M,KAAKwM,EAAOxM;kCAGdwM,EAAOG,0BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAK;YAC1C2Y;YACA3M,KAAKwM,EAAOxM;iCAGdwM,EAAOG,0BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAUzE,OAAOjU,yBAEvCwY,EAAOG,6BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAK;YAC1C2Y;YACA3M,KAAKyM,EAAUzM;+BAGjBwM,EAAOG,4BACPF,EAAUE,OAEV3hB,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAK;YAC1C2Y;YACA3M,KAAKwM,EAAOxM;;;;;;;;;QAUdlP,OA/DA9F,KAAK0hB,KAAY1hB,KAAK0hB,GAAU5E,GAAO9T,GAAKwY;;IAwEhD1hB;QACE,MAAM8hB,IAAgC;QAMtC,OALA5hB,KAAK0hB,GAAUlE,GACb,CAACxU,GAAkBwY;YACjBI,EAAQtb,KAAKkb;YAGVI;;;;MAIEC;IACX/hB,YACWgiB,GACAC,GACAC,GACAC,GACAC,GACAC,GACAC,GACAC;QAPAriB,aAAA8hB,GACA9hB,YAAA+hB,aACAC,GACAhiB,kBAAAiiB,aACAC,GACAliB,iBAAAmiB;kBACAC,aACAC;;sFAIXviB,UACEgiB,GACAQ,GACAJ,GACAC;QAEA,MAAMP,IAAgC;QAKtC,OAJAU,EAAUlc,QAAQ4O;YAChB4M,EAAQtb,KAAK;gBAAEqb;gBAAwB3M,KAAAA;;YAGlC,IAAI6M,GACTC,GACAQ,GACAvB,GAAYwB,GAASD,IACrBV,GACAM,GACAC;iCACwB;wCACO;;IAInC3K;QACE,QAAQxX,KAAKkiB,GAAY7Y;;IAG3BvJ,QAAQkF;QACN,MACEhF,KAAKmiB,cAAcnd,EAAMmd,aACzBniB,KAAKoiB,OAAqBpd,EAAMod,MAC/BpiB,KAAKkiB,GAAY1c,QAAQR,EAAMkd,OAC/BliB,KAAK8hB,MAAMtc,QAAQR,EAAM8c,UACzB9hB,KAAK+hB,KAAKvc,QAAQR,EAAM+c,SACxB/hB,KAAKgiB,GAAQxc,QAAQR,EAAMgd,MAE5B,QAAO;QAET,MAAMJ,IAAgC5hB,KAAKiiB,YACrCO,IAAqCxd,EAAMid;QACjD,IAAIL,EAAQhc,WAAW4c,EAAa5c,QAClC,QAAO;QAET,KAAK,IAAIc,IAAI,GAAGA,IAAIkb,EAAQhc,QAAQc,KAClC,IACEkb,EAAQlb,GAAGib,SAASa,EAAa9b,GAAGib,SACnCC,EAAQlb,GAAGsO,IAAIxP,QAAQgd,EAAa9b,GAAGsO,MAExC,QAAO;QAGX,QAAO;;;;;;;;;;;;;;;;;;;;;;;;UCzKEyN;IACX3iB;;;;IAIWkc;;;;IAIA0G;;;;;IAKAC;;;;;IAKAC;;;;IAIAC;kBAlBA7G,aAIA0G,aAKAC,aAKAC,aAIAC;;;;;;;;;IAUX/iB,UACE+b,GACA3T;QAEA,MAAMwa,IAAgB,IAAI1M;QAQ1B,OAPA0M,EAAc/N,IACZkH,GACAiH,GAAaC,GACXlH,GACA3T,KAGG,IAAIua,GACTpd,EAAgB2B,OAChB0b,GACA5B,MACAT,MACAO;;;;;;;;;;;UAaOkC;IACXhjB;;;;;;;IAOWoc;;;;;;IAMAhU;;;;;IAKA8a;;;;;IAKAC;;;;;IAKAC;QArBAljB,mBAAAkc,aAMAhU,aAKA8a,aAKAC,aAKAC;;;;;;WAQXpjB,UACE+b,GACA3T;QAEA,OAAO,IAAI4a,GACTxZ,EAAW6S,IACXjU,GACA0Y,MACAA,MACAA;;;;;;;;;;;;;;;;;;;;;;;;;UC3FOuC;IACXrjB;;IAESsjB;;IAEAC;;IAEAra;;;;;IAKAsa;kBATAF,GAEApjB,wBAAAqjB,GAEArjB,WAAAgJ,aAKAsa;;;;MAIEC;IACXzjB,YACS+b,GACA2H;QADAxjB,gBAAA6b,aACA2H;;;;MAYEC;IACX3jB;;IAES4jB;;IAEAC;;;;;;;IAOAzH,IAA0B5S,EAAW6S;2DAErCyH,IAA+B;QAX/B5jB,aAAA0jB,GAEA1jB,iBAAA2jB,GAOA3jB,mBAAAkc,GAEAlc,aAAA4jB;;;;mDAKX,OAAMC;IAAN/jB;;;;;QAKEE,UAA2B;;;;;;;QAQ3BA,UAGI8jB;;QAGJ9jB,UAAmCsJ,EAAW6S,IAC9Cnc,WAAmB;;;;;;QAOnBA,WAA6B;;;;;;;;;WAU7B+jB;QACE,OAAO/jB,KAAKgkB;;gEAId9H;QACE,OAAOlc,KAAKikB;;6EAIdC;QACE,OAAiC,MAA1BlkB,KAAKmkB;;iFAIdC;QACE,OAAOpkB,KAAKqkB;;;;;WAOdvkB,GAAkBoc;QACZA,EAAYoI,OAAwB,MACtCtkB,KAAKqkB,MAAqB,GAC1BrkB,KAAKikB,KAAe/H;;;;;;;WAUxBpc;QACE,IAAIkjB,IAAiBpC,MACjBqC,IAAoBrC,MACpBsC,IAAmBtC;QAkBvB,OAhBA5gB,KAAKukB,GAAgBne,QAAQ,CAAC4C,GAAKwb;YACjC,QAAQA;cACN;gBACExB,IAAiBA,EAAelD,IAAI9W;gBACpC;;cACF;gBACEia,IAAoBA,EAAkBnD,IAAI9W;gBAC1C;;cACF;gBACEka,IAAmBA,EAAiBpD,IAAI9W;gBACxC;;cACF;gBACElD;;YAIC,IAAIgd,GACT9iB,KAAKikB,IACLjkB,KAAKgkB,IACLhB,GACAC,GACAC;;;;WAOJpjB;QACEE,KAAKqkB,MAAqB,GAC1BrkB,KAAKukB,KAAkBT;;IAGzBhkB,GAAkBkJ,GAAkBwb;QAClCxkB,KAAKqkB,MAAqB,GAC1BrkB,KAAKukB,KAAkBvkB,KAAKukB,GAAgBzH,GAAO9T,GAAKwb;;IAG1D1kB,GAAqBkJ;QACnBhJ,KAAKqkB,MAAqB,GAC1BrkB,KAAKukB,KAAkBvkB,KAAKukB,GAAgBtH,OAAOjU;;IAGrDlJ;QACEE,KAAKmkB,MAAoB;;IAG3BrkB;QACEE,KAAKmkB,MAAoB;;IAG3BrkB;QACEE,KAAKqkB,MAAqB,GAC1BrkB,KAAKgkB,MAAW;;;;;;;MA2BPS;IACX3kB,YAAoB4kB;kBAAAA;;QAGpB1kB,UAAuB,IAAIgW;;QAG3BhW,UAAiCqgB;;QAGjCrgB,UAAuC2kB;;;;;;QAOvC3kB,UAA8B,IAAIkf,GAAoBja;;;;WAKtDnF,GAAqB8kB;QACnB,KAAK,MAAM/I,KAAY+I,EAAUxB,IAC3BwB,EAAUtB,cAAkBlQ,KAC9BpT,KAAK6kB,GAAoBhJ,GAAU+I,EAAUtB,MACpCsB,EAAUtB,cAAkB1N,MACrC5V,KAAK8kB,GACHjJ,GACA+I,EAAU5b,KACV4b,EAAUtB;QAKhB,KAAK,MAAMzH,KAAY+I,EAAUvB,kBAC/BrjB,KAAK8kB,GAAyBjJ,GAAU+I,EAAU5b,KAAK4b,EAAUtB;;sFAKrExjB,GAAmBilB;QACjB/kB,KAAKglB,GAAcD,GAAclJ;YAC/B,MAAMoJ,IAAcjlB,KAAKklB,GAAkBrJ;YAC3C,QAAQkJ,EAAarB;cACnB;gBACM1jB,KAAKmlB,GAAetJ,MACtBoJ,EAAYG,GAAkBL,EAAa7I;gBAE7C;;cACF;;;gBAGE+I,EAAYI,MACPJ,EAAYK;;;;gBAIfL,EAAYM,MAEdN,EAAYG,GAAkBL,EAAa7I;gBAC3C;;cACF;;;;;gBAKE+I,EAAYI,MACPJ,EAAYK,MACftlB,KAAKwlB,aAAa3J;gBAMpB;;cACF;gBACM7b,KAAKmlB,GAAetJ,OACtBoJ,EAAYQ,MACZR,EAAYG,GAAkBL,EAAa7I;gBAE7C;;cACF;gBACMlc,KAAKmlB,GAAetJ;;;;gBAItB7b,KAAK0lB,GAAY7J,IACjBoJ,EAAYG,GAAkBL,EAAa7I;gBAE7C;;cACF;gBACEpW;;;;;;;;WAURhG,GACEilB,GACAne;QAEIme,EAAapB,UAAU/d,SAAS,IAClCmf,EAAapB,UAAUvd,QAAQQ,KAE/B5G,KAAK2lB,GAAavf,QAAQ,CAACwf,GAAG/J;YACxB7b,KAAKmlB,GAAetJ,MACtBjV,EAAGiV;;;;;;;WAWX/b,GAAsB+lB;QACpB,MAAMhK,IAAWgK,EAAYhK,UACvBiK,IAAgBD,EAAYrC,GAAgBza,OAE5Cgd,IAAa/lB,KAAKgmB,GAA0BnK;QAClD,IAAIkK,GAAY;YACd,MAAMnK,IAASmK,EAAWnK;YAC1B,IAAIA,EAAO1B,MACT,IAAsB,MAAlB4L,GAAqB;;;;;;;gBAOvB,MAAM9c,IAAM,IAAIT,EAAYqT,EAAOrU;gBACnCvH,KAAK8kB,GACHjJ,GACA7S,GACA,IAAI4M,GAAW5M,GAAK3D,EAAgB2B;mBAtWpC3D,GA0WkB,MAAlByiB,SAIC;gBACe9lB,KAAKimB,GAAiCpK,OACtCiK;;;gBAGlB9lB,KAAK0lB,GAAY7J,IACjB7b,KAAKkmB,KAAsBlmB,KAAKkmB,GAAoBpG,IAAIjE;;;;;;;WAUhE/b,GAAkBkc;QAChB,MAAM0G,IAAgB,IAAI1M;QAE1BhW,KAAK2lB,GAAavf,QAAQ,CAAC6e,GAAapJ;YACtC,MAAMkK,IAAa/lB,KAAKgmB,GAA0BnK;YAClD,IAAIkK,GAAY;gBACd,IAAId,EAAY/c,MAAW6d,EAAWnK,OAAO1B,MAAmB;;;;;;;;;oBAU9D,MAAMlR,IAAM,IAAIT,EAAYwd,EAAWnK,OAAOrU;oBAEH,SAAzCvH,KAAKmmB,GAAuBpjB,IAAIiG,MAC/BhJ,KAAKomB,GAAuBvK,GAAU7S,MAEvChJ,KAAK8kB,GACHjJ,GACA7S,GACA,IAAI4M,GAAW5M,GAAKgT;;gBAKtBiJ,EAAYoB,OACd3D,EAAc/N,IAAIkH,GAAUoJ,EAAYqB,OACxCrB,EAAYM;;;QAKlB,IAAI1C,IAAyBjC;;;;;;gBAO7B5gB,KAAKumB,GAA6BngB,QAAQ,CAAC4C,GAAKwd;YAC9C,IAAIC,KAAoB;YAExBD,EAAQE,GAAa7K;gBACnB,MAAMkK,IAAa/lB,KAAKgmB,GAA0BnK;gBAClD,QACEkK,iCACAA,EAAWjK,OAEX2K,KAAoB,IACb;gBAMPA,MACF5D,IAAyBA,EAAuB/C,IAAI9W;;QAIxD,MAAM2d,IAAc,IAAIlE,GACtBzG,GACA0G,GACA1iB,KAAKkmB,IACLlmB,KAAKmmB,IACLtD;QAOF,OAJA7iB,KAAKmmB,KAAyB9F,MAC9BrgB,KAAKumB,KAA+B5B,MACpC3kB,KAAKkmB,KAAsB,IAAIhH,GAAoBja,KAE5C0hB;;;;;;;IAQT7mB,GAAoB+b,GAAoB+K;QACtC,KAAK5mB,KAAKmlB,GAAetJ,IACvB;QAGF,MAAM2I,IAAaxkB,KAAKomB,GAAuBvK,GAAU+K,EAAS5d;QAI9ChJ,KAAKklB,GAAkBrJ,GAC/BgL,GAAkBD,EAAS5d,KAAKwb,IAE5CxkB,KAAKmmB,KAAyBnmB,KAAKmmB,GAAuBrJ,GACxD8J,EAAS5d,KACT4d,IAGF5mB,KAAKumB,KAA+BvmB,KAAKumB,GAA6BzJ,GACpE8J,EAAS5d,KACThJ,KAAK8mB,GAA4BF,EAAS5d,KAAK8W,IAAIjE;;;;;;;;;;IAYvD/b,GACE+b,GACA7S,GACA+d;QAEA,KAAK/mB,KAAKmlB,GAAetJ,IACvB;QAGF,MAAMoJ,IAAcjlB,KAAKklB,GAAkBrJ;QACvC7b,KAAKomB,GAAuBvK,GAAU7S,KACxCic,EAAY4B,GAAkB7d;;;QAI9Bic,EAAY+B,GAAqBhe,IAGnChJ,KAAKumB,KAA+BvmB,KAAKumB,GAA6BzJ,GACpE9T,GACAhJ,KAAK8mB,GAA4B9d,GAAK4L,OAAOiH,KAG3CkL,MACF/mB,KAAKmmB,KAAyBnmB,KAAKmmB,GAAuBrJ,GACxD9T,GACA+d;;IAKNjnB,aAAa+b;QACX7b,KAAK2lB,GAAa/Q,OAAOiH;;;;;;WAQnB/b,GAAiC+b;QACvC,MACMkJ,IADc/kB,KAAKklB,GAAkBrJ,GACVyK;QACjC,OACEtmB,KAAK0kB,GAAiBuC,GAAuBpL,GAAUrV,OACvDue,EAAa/B,GAAexc,OAC5Bue,EAAa7B,GAAiB1c;;;;;WAQlC1G,GAA2B+b;QAEL7b,KAAKklB,GAAkBrJ,GAC/BqL;;IAGNpnB,GAAkB+b;QACxB,IAAI1M,IAASnP,KAAK2lB,GAAa5iB,IAAI8Y;QAKnC,OAJK1M,MACHA,IAAS,IAAI0U,IACb7jB,KAAK2lB,GAAahR,IAAIkH,GAAU1M,KAE3BA;;IAGDrP,GAA4BkJ;QAClC,IAAIme,IAAgBnnB,KAAKumB,GAA6BxjB,IAAIiG;QAU1D,OARKme,MACHA,IAAgB,IAAIjI,GAAoBja,KACxCjF,KAAKumB,KAA+BvmB,KAAKumB,GAA6BzJ,GACpE9T,GACAme,KAIGA;;;;;;WAQCrnB,GAAe+b;QACvB,MAAMuL,IAA4D,SAA7CpnB,KAAKgmB,GAA0BnK;QAIpD,OAHKuL,KACHC,GAxXU,yBAwXQ,4BAA4BxL,IAEzCuL;;;;;WAOCtnB,GAA0B+b;QAClC,MAAMoJ,IAAcjlB,KAAK2lB,GAAa5iB,IAAI8Y;QAC1C,OAAOoJ,KAAeA,EAAYK,KAC9B,OACAtlB,KAAK0kB,GAAiB4C,GAAuBzL;;;;;;WAQ3C/b,GAAY+b;QAKlB7b,KAAK2lB,GAAahR,IAAIkH,GAAU,IAAIgI,KAKf7jB,KAAK0kB,GAAiBuC,GAAuBpL,GACrDzV,QAAQ4C;YACnBhJ,KAAK8kB,GAAyBjJ,GAAU7S,wBAA0B;;;;;;WAO9DlJ,GACN+b,GACA7S;QAGA,OADqBhJ,KAAK0kB,GAAiBuC,GAAuBpL,GAC9CgE,IAAI7W;;;;AAI5B,SAAS2b;IACP,OAAO,IAAIjI,GACTnU,EAAYvC;;;AAIhB,SAAS8d;IACP,OAAO,IAAIpH,GAAmCnU,EAAYvC;;;;;;;;;;;;;;;;;;GClmB5D,OAAMuhB,KAAa;IACjB,MAAMC,IAA8C;QACpDC,KAA4B;QAC5BC,MAA6B;;IAC7B,OAAOF;EAJU,IAObG,KAAY;IAChB,MAAMC,IAA2C;QACjDC,KAA0B;QAC1BC,MAAmC;QACnCC,KAA6B;QAC7BC,MAAsC;QACtCC,MAAsB;QACtBC,kBAA+B;QAC/BC,IAAmB;QACnBC,sBAAmC;;IACnC,OAAOR;EAVS;;;;;;;;MAkCLS;IACXvoB,YACUoQ,GACAoH;kBADApH,GACAlQ,eAAAsX;;IAGVxX,GAAcwoB;QACZ,MAAM/mB,SACYsE,MAAhByiB,EAAO/mB,OACHpB,EAAKG,UACLic,GAAmB+L,EAAO/mB;QAChC,OAAO,IAAIF,EAAeE,GAAM+mB,EAAO9mB,WAAW;;;;;;;;;WAW5C1B,GAAaqb;QACnB,OAAInb,KAAKsX,QAAQiR,MAAiBne,EAAkB+Q,KAC3CA,IAEA;YAAEtZ,OAAOsZ;;;;;WAOZrb,GACNqb;QAEA,IAAIhM;QAMJ,OAJEA,IADiB,mBAARgM,IACAA,EAAItZ,QAEJsZ,GAEJ/Q,EAAkB+E,KAAU,OAAOA;;;;WAM5CrP,GAAU+B;QACR,OAAO;YAAEmK,cAAc,KAAKnK;;;;;;WAO9B/B,GAAS+B;QACP,IAAI7B,KAAKsX,QAAQiR,IAAe;YAC9B,IAAIpc,MAAMtK,IACR,OAAO;gBAAEsL,aAAa;;YACjB,IAAItL,MAAU2mB,IAAAA,GACnB,OAAO;gBAAErb,aAAa;;YACjB,IAAItL,OAAW2mB,IAAAA,GACpB,OAAO;gBAAErb,aAAa;;;QAG1B,OAAO;YAAEA,aAAa9C,EAAexI,KAAS,OAAOA;;;;;;;WAQvD/B,GAAS+B;QACP,gBnBzJ0BA;YAC5B,OACmB,mBAAVA,KACP+N,OAAOS,UAAUxO,OAChBwI,EAAexI,MAChBA,KAAS+N,OAAO6Y,oBAChB5mB,KAAS+N,OAAO8Y;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GmBmJTC,EAAc9mB,KAAS7B,KAAKgS,GAAUnQ,KAAS7B,KAAKiS,GAASpQ;;;;WAMtE/B,EAAYwF;QACV,IAAItF,KAAKsX,QAAQiR,IAAe;YAU9B,OAAO,GANW,IAAI/jB,KAAyB,MAApBc,EAAUjB,SAAgBukB,cAEnB5gB,QAAQ,SAAS,IAAIA,QAAQ,KAAK,QAEnD,cAAc1C,EAAUhB,aAAa4B,OAAO;;QAI7D,OAAO;YACL7B,SAAS,KAAKiB,EAAUjB;YACxBwG,OAAOvF,EAAUhB;;;IAMfxE,EAAc4E;QACpB,MAAMY,IAAYsF,EAAmBlG;QACrC,OAAO,IAAIN,EAAUkB,EAAUjB,SAASiB,EAAUuF;;;;;;WAQpD/K,GAAQ+oB;QACN,OAAI7oB,KAAKsX,QAAQiR,KACRM,EAAM7Z,aAEN6Z,EAAMC;;;;WAOjBhpB,GAAU+B;QACR,OAAI7B,KAAKsX,QAAQiR,MACfllB,QACYwC,MAAVhE,KAAwC,mBAAVA,IAGzByH,EAAWyG,iBAAiBlO,KAAgB,QAEnDwB,QACYwC,MAAVhE,KAAuBA,aAAiBoI;QAGnCX,EAAW0G,eAAenO,KAAgB,IAAIoI;;IAIzDnK,UAAU+S;QACR,OAAO7S,KAAK+oB,EAAYlW,EAAQkW;;IAGlCjpB,YAAY+S;QAEV,OAnNaxP,KAkNAwP,IACNxN,EAAgB2jB,EAAchpB,KAAKgpB,EAAcnW;;IAG1D/S,GAAeyH,GAAoB2I;QACjC,OAAOlQ,KAAKipB,GAAyB/Y,KAAclQ,KAAKkQ,IACrD8G,MAAM,aACNA,MAAMzP,GACND;;IAGLxH,GAAiB6B;QACf,MAAMunB,IAAW/hB,EAAaqB,EAAW7G;QAKzC,OAnOa0B,GAgOX8lB,GAAoBD,KAGfA;;IAGTppB,GAAOkJ;QACL,OAAOhJ,KAAKopB,GAAepgB,EAAIzB;;IAGjCzH,EAAS6B;QACP,MAAMunB,IAAWlpB,KAAKqpB,GAAiB1nB;QAgBvC,OA3Pa0B,GA6OX6lB,EAASnmB,IAAI,OAAO/C,KAAKkQ,GAAWC,YA7OzB9M,IAoPT6lB,EAASnmB,IAAI,OAAO/C,KAAKkQ,GAAWE,YACpC8Y,EAASnmB,IAAI,OAAO/C,KAAKkQ,GAAWE;QAMjC,IAAI7H,EAAYvI,KAAKspB,GAAiCJ;;IAG/DppB,GAAYyH;QACV,OAAOvH,KAAKopB,GAAe7hB;;IAG7BzH,GAAc6B;QACZ,MAAM4nB,IAAevpB,KAAKqpB,GAAiB1nB;;;;;gBAK3C,OAA4B,MAAxB4nB,EAAa3jB,SACRuB,EAAasP,IAEfzW,KAAKspB,GAAiCC;;IAG/CC;QAOE,OANa,IAAIriB,EAAa,EAC5B,YACAnH,KAAKkQ,GAAWC,WAChB,aACAnQ,KAAKkQ,GAAWE,YAEN9I;;IAGNxH,GAAyBoQ;QAC/B,OAAO,IAAI/I,EAAa,EACtB,YACA+I,EAAWC,WACX,aACAD,EAAWE;;IAIPtQ,GACNypB;QAMA,OAxSalmB,GAqSXkmB,EAAa3jB,SAAS,KAA6B,gBAAxB2jB,EAAaxmB,IAAI,KAGvCwmB,EAAa9gB,EAAS;;sFAI/B3I,GAAmBkJ,GAAkBwB;QACnC,OAAO;YACL7I,MAAM3B,KAAKypB,GAAOzgB;YAClBwB,QAAQA,EAAOsL,MAAMvL,SAASC;;;IAIlC1K,GAAW8mB;QAKT,OAAO;YACLjlB,MAAM3B,KAAKypB,GAAO7C,EAAS5d;YAC3BwB,QAAQoc,EAAS8C,KAAUnf,SAASC;YACpCwI,YAAYhT,KAAK+oB,EAAYnC,EAAS/T,QAAQkW;;;IAIlDjpB,GACE8mB,GACAlT;QAEA,MAAM1K,IAAMhJ,KAAKiP,EAAS2X,EAAc,OAClC/T,IAAU7S,KAAK2pB,YAAY/C,EAAS5T,aACpCgB,IAAO,IAAIK,GAAY;YAAE9J,UAAU;gBAAEC,QAAQoc,EAASpc;;;QAC5D,OAAO,IAAI4I,GAASpK,GAAK6J,GAASmB,GAAM;YACtCN,yBAAyBA;;;IAIrB5T,GAAUkV;QAChB3R,KACI2R,EAAI4U,QAGM5U,EAAI4U,MAAMjoB,MACVqT,EAAI4U,MAAM5W;QACxB,MAAMhK,IAAMhJ,KAAKiP,EAAS+F,EAAI4U,MAAMjoB,OAC9BkR,IAAU7S,KAAK2pB,YAAY3U,EAAI4U,MAAM5W,aACrCgB,IAAO,IAAIK,GAAY;YAAE9J,UAAU;gBAAEC,QAAQwK,EAAI4U,MAAMpf;;;QAC7D,OAAO,IAAI4I,GAASpK,GAAK6J,GAASmB,GAAM;;IAGlClU,GAAYqP;QAClB9L,KACI8L,EAAO0a,UAGXxmB,KACI8L,EAAO2a;QAGX,MAAM9gB,IAAMhJ,KAAKiP,EAASE,EAAO0a,UAC3BhX,IAAU7S,KAAK2pB,YAAYxa,EAAO2a;QACxC,OAAO,IAAIlU,GAAW5M,GAAK6J;;IAG7B/S,GAAkBqP;QAChB,OAAI,WAAWA,IACNnP,KAAK+pB,GAAU5a,KACb,aAAaA,IACfnP,KAAKgqB,GAAY7a,KA/XdrJ;;IAoYdhG,GAAgB0hB;QACd,IAAIqE;QACJ,IAAI,kBAAkBrE,GAAQ;YACdA,EAAOuD;;;YAGrB,MAAMrB,IAAQ1jB,KAAKiqB,GACjBzI,EAAOuD,aAAamF,oBAAoB,cAEpCvG,IAAwBnC,EAAOuD,aAAapB,aAAa,IAEzDzH,IAAclc,KAAKmqB,GAAU3I,EAAOuD,aAAa7I,cACjDkO,IAAa5I,EAAOuD,aAAcnB,OAClCA,IAAQwG,KAAcpqB,KAAKqqB,GAAcD;YAC/CvE,IAAc,IAAIpC,GAChBC,GACAC,GACAzH,GACA0H,KAAS;eAEN,IAAI,oBAAoBpC,GAAQ;YACvBA,EAAO8I;YACrB,MAAMC,IAAe/I,EAAO8I;YACdC,EAAa3D,UACb2D,EAAa3D,SAASjlB,MAElC4oB,EAAa3D,SAAS5T;YAGxB,MAAMhK,IAAMhJ,KAAKiP,EAASsb,EAAa3D,SAASjlB,OAC1CkR,IAAU7S,KAAK2pB,YAAYY,EAAa3D,SAAS5T,aACjDgB,IAAO,IAAIK,GAAY;gBAC3B9J,UAAU;oBAAEC,QAAQ+f,EAAa3D,SAASpc;;gBAEtCwK,IAAM,IAAI5B,GAASpK,GAAK6J,GAASmB,GAAM,KACvCoP,IAAmBmH,EAAa5G,aAAa,IAC7CN,IAAmBkH,EAAalH,oBAAoB;YAC1DwC,IAAc,IAAI1C,GAChBC,GACAC,GACArO,EAAIhM,KACJgM;eAEG,IAAI,oBAAoBwM,GAAQ;YACvBA,EAAOgJ;YACrB,MAAMC,IAAYjJ,EAAOgJ;YACXC,EAAU7D;YACxB,MAAM5d,IAAMhJ,KAAKiP,EAASwb,EAAU7D,WAC9B/T,IAAU4X,EAAUX,WACtB9pB,KAAK2pB,YAAYc,EAAUX,YAC3BzkB,EAAgB2B,OACdgO,IAAM,IAAIY,GAAW5M,GAAK6J,IAC1BwQ,IAAmBoH,EAAUpH,oBAAoB;YACvDwC,IAAc,IAAI1C,GAAoB,IAAIE,GAAkBrO,EAAIhM,KAAKgM;eAChE,IAAI,oBAAoBwM,GAAQ;YACvBA,EAAOkJ;YACrB,MAAMC,IAAYnJ,EAAOkJ;YACXC,EAAU/D;YACxB,MAAM5d,IAAMhJ,KAAKiP,EAAS0b,EAAU/D,WAC9BvD,IAAmBsH,EAAUtH,oBAAoB;YACvDwC,IAAc,IAAI1C,GAAoB,IAAIE,GAAkBra,GAAK;eAC5D;YAAA,MAAI,YAAYwY,IAUrB,OA3cU1b;YAicmB;gBAEf0b,EAAO9Z;gBACrB,MAAMA,IAAS8Z,EAAO9Z;gBACRA,EAAOmU;gBACrB,MAAM9S,IAAQrB,EAAOqB,SAAS,GACxBya,IAAkB,IAAIpH,GAAgBrT,IACtC8S,IAAWnU,EAAOmU;gBACxBgK,IAAc,IAAItC,GAAsB1H,GAAU2H;;;QAIpD,OAAOqC;;IAGT/lB,GACE4jB;QAEA,OAAc,gBAAVA,uBAEiB,UAAVA,oBAEU,aAAVA,sBAEU,cAAVA,sBAEU,YAAVA,oBA3dC5d;;IAkedhG,GAA0B0hB;;;;QAIxB,MAAM,kBAAkBA,IACtB,OAAOnc,EAAgB2B;QAEzB,MAAM+d,IAAevD,EAAoB;QACzC,OAAIuD,EAAapB,aAAaoB,EAAapB,UAAU/d,SAC5CP,EAAgB2B,QAEpB+d,EAAa+E,WAGX9pB,KAAK2pB,YAAY5E,EAAa+E,YAF5BzkB,EAAgB2B;;IAK3BlH,GAAW8qB;QACT,IAAIzb;QACJ,IAAIyb,aAAoBtX,IACtBnE,IAAS;YACP0b,QAAQ7qB,KAAK8qB,GAAmBF,EAAS5hB,KAAK4hB,EAAS/oB;gBAEpD,IAAI+oB,aAAoBjV,IAC7BxG,IAAS;YAAEyF,QAAQ5U,KAAKypB,GAAOmB,EAAS5hB;gBACnC,IAAI4hB,aAAoB7W,IAC7B5E,IAAS;YACP0b,QAAQ7qB,KAAK8qB,GAAmBF,EAAS5hB,KAAK4hB,EAAS5W;YACvD+W,YAAY/qB,KAAKgrB,GAAeJ,EAAS3W;gBAEtC,IAAI2W,aAAoB9V,IAC7B3F,IAAS;YACPwD,WAAW;gBACTiU,UAAU5mB,KAAKypB,GAAOmB,EAAS5hB;gBAC/B+L,iBAAiB6V,EAAS7V,gBAAgBjN,IAAI6K,KAC5C3S,KAAKirB,GAAiBtY;;gBAIvB;YAAA,MAAIiY,aAAoB/U,KAK7B,OA9gBU/P;YA0gBVqJ,IAAS;gBACP+b,QAAQlrB,KAAKypB,GAAOmB,EAAS5hB;;;QAUjC,OAJK4hB,EAASrX,GAAa4X,OACzBhc,EAAOic,kBAAkBprB,KAAKqrB,GAAeT,EAASrX,MAGjDpE;;IAGTrP,GAAagW;QACX,MAAMvC,IAAeuC,EAAMsV,kBACvBprB,KAAKsrB,GAAiBxV,EAAMsV,mBAC5BrY,GAAawY;QAEjB,IAAIzV,EAAM+U,QAAQ;YACF/U,EAAM+U,OAAOlpB;YAC3B,MAAMqH,IAAMhJ,KAAKiP,EAAS6G,EAAM+U,OAAOlpB,OACjCE,IAAQ,IAAIwS,GAAY;gBAC5B9J,UAAU;oBAAEC,QAAQsL,EAAM+U,OAAOrgB;;;YAEnC,IAAIsL,EAAMiV,YAAY;gBACpB,MAAM9W,IAAYjU,KAAKwrB,GAAiB1V,EAAMiV;gBAC9C,OAAO,IAAIhX,GAAc/K,GAAKnH,GAAOoS,GAAWV;;YAEhD,OAAO,IAAID,GAAYtK,GAAKnH,GAAO0R;;QAEhC,IAAIuC,EAAMlB,QAAQ;YACvB,MAAM5L,IAAMhJ,KAAKiP,EAAS6G,EAAMlB;YAChC,OAAO,IAAIe,GAAe3M,GAAKuK;;QAC1B,IAAIuC,EAAMnD,WAAW;YAC1B,MAAM3J,IAAMhJ,KAAKiP,EAAS6G,EAAMnD,UAAmB,WAC7CoC,IAAkBe,EAAMnD,UAAUoC,gBAAiBjN,IAAI6K,KAC3D3S,KAAKyrB,GAAmB9Y;YAM1B,OAJAtP,IAC0B,MAAxBkQ,EAAaN,SAGR,IAAI6B,GAAkB9L,GAAK+L;;QAC7B,IAAIe,EAAMoV,QAAQ;YACvB,MAAMliB,IAAMhJ,KAAKiP,EAAS6G,EAAMoV;YAChC,OAAO,IAAIrV,GAAe7M,GAAKuK;;QAE/B,OA1jBUzN;;IA8jBNhG,GAAeyT;QAErB,YAAgC1N,MAA5B0N,EAAaP,aACR;YACLA,YAAYhT,KAAK0rB,UAAUnY,EAAaP;iBAETnN,MAAxB0N,EAAaN,SACf;YAAEA,QAAQM,EAAaN;YArkBpBnN;;IA2kBNhG,GAAiByT;QACvB,YAAgC1N,MAA5B0N,EAAaP,aACRD,GAAaC,WAAWhT,KAAK2pB,YAAYpW,EAAaP,oBAC5BnN,MAAxB0N,EAAaN,SACfF,GAAaE,OAAOM,EAAaN,UAEjCF,GAAawY;;IAIhBzrB,GACNgW,GACA6V;;QAGA,IAAI9Y,IAAUiD,EAAM9C,aAChBhT,KAAK2pB,YAAY7T,EAAM9C,cACvBhT,KAAK2pB,YAAYgC;QAEjB9Y,EAAQrN,QAAQH,EAAgB2B;;;;;;QAMlC6L,IAAU7S,KAAK2pB,YAAYgC;QAG7B,IAAI7Y,IAAuC;QAI3C,OAHIgD,EAAMhD,oBAAoBgD,EAAMhD,iBAAiBlN,SAAS,MAC5DkN,IAAmBgD,EAAMhD;QAEpB,IAAIF,GAAeC,GAASC;;IAGrChT,GACE8rB,GACAD;QAEA,OAAIC,KAAUA,EAAOhmB,SAAS,KA7lBjBvC,QA+lBMwC,MAAf8lB,IAGKC,EAAO9jB,IAAIgO,KAAS9V,KAAK6rB,GAAgB/V,GAAO6V,OAEhD;;IAIH7rB,GAAiBwV;QACvB,MAAM3C,IAAY2C,EAAe3C;QACjC,IAAIA,aAAqBjC,IACvB,OAAO;YACL0B,WAAWkD,EAAe5C,MAAMpL;YAChCwkB,kBAAkB;;QAEf,IAAInZ,aAAqB3B,IAC9B,OAAO;YACLoB,WAAWkD,EAAe5C,MAAMpL;YAChCykB,uBAAuB;gBACrBxf,QAAQoG,EAAU1B;;;QAGjB,IAAI0B,aAAqBpB,IAC9B,OAAO;YACLa,WAAWkD,EAAe5C,MAAMpL;YAChC0kB,oBAAoB;gBAClBzf,QAAQoG,EAAU1B;;;QAGjB,IAAI0B,aAAqBlB,IAC9B,OAAO;YACLW,WAAWkD,EAAe5C,MAAMpL;YAChC2kB,WAAWtZ,EAAUhB;;QAGvB,MAxpBU7L;;IA4pBNhG,GAAmBgW;QACzB,IAAInD,IAAuC;QAC3C,IAAI,sBAAsBmD,GACxBzS,GAC6B,mBAA3ByS,EAAMgW,mBAGRnZ,IAAYjC,GAAyBwb,eAChC,IAAI,2BAA2BpW,GAAO;YAC3C,MAAMvJ,IAASuJ,EAAMiW,sBAAuBxf,UAAU;YACtDoG,IAAY,IAAI3B,GAA6BzE;eACxC,IAAI,wBAAwBuJ,GAAO;YACxC,MAAMvJ,IAASuJ,EAAMkW,mBAAoBzf,UAAU;YACnDoG,IAAY,IAAIpB,GAA8BhF;eACrC,eAAeuJ,IACxBnD,IAAY,IAAIlB,GACdzR,MACA8V,EAAgB,aAGlBhQ;QAEF,MAAMsM,IAAYxK,EAAUukB,EAAiBrW,EAAgB;QAC7D,OAAO,IAAIrD,GAAeL,GAAWO;;IAGvC7S,GAAkB8b;QAChB,OAAO;YAAE0G,WAAW,EAACtiB,KAAKosB,GAAYxQ,EAAOrU;;;IAG/CzH,GAAoBusB;QArqBLhpB,GAwqBD,MAFEgpB,EAAgB/J,UAAW1c;QAKzC,MAAMjE,IAAO0qB,EAAgB/J,UAAW;QACxC,OAAOpK,GAAMoU,GAAOtsB,KAAKusB,GAAc5qB,IAAO0X;;IAGhDvZ,GAAc8b;;QAEZ,MAAMzM,IAA0B;YAAEqd,iBAAiB;WAC7CjlB,IAAOqU,EAAOrU;QACW,SAA3BqU,EAAOlE,mBAKTvI,EAAOsd,SAASzsB,KAAKosB,GAAY7kB,IACjC4H,EAAOqd,gBAAiBE,OAAO,EAC7B;YACEhkB,cAAckT,EAAOlE;YACrBiV,iBAAgB;gBAQpBxd,EAAOsd,SAASzsB,KAAKosB,GAAY7kB,EAAKqlB,MACtCzd,EAAOqd,gBAAiBE,OAAO,EAAC;YAAEhkB,cAAcnB,EAAKwO;;QAGvD,MAAM8W,IAAQ7sB,KAAK8sB,GAASlR,EAAOhE;QAC/BiV,MACF1d,EAAOqd,gBAAiBK,QAAQA;QAGlC,MAAMlV,IAAU3X,KAAK+sB,GAAQnR,EAAOjE;QAChCA,MACFxI,EAAOqd,gBAAiB7U,UAAUA;QAGpC,MAAMxR,IAAQnG,KAAKgtB,GAAapR,EAAOzV;QAYvC,OAXc,SAAVA,MACFgJ,EAAOqd,gBAAiBrmB,QAAQA,IAG9ByV,EAAO/D,YACT1I,EAAOqd,gBAAiB3U,UAAU7X,KAAKitB,GAASrR,EAAO/D;QAErD+D,EAAO9D,UACT3I,EAAOqd,gBAAiB1U,QAAQ9X,KAAKitB,GAASrR,EAAO9D,SAGhD3I;;IAGTrP,GAAgB8b;QACd,IAAIrU,IAAOvH,KAAKusB,GAAc3Q,EAAc;QAE5C,MAAMkG,IAAQlG,EAAO4Q,iBACfU,IAAYpL,EAAM4K,OAAO5K,EAAM4K,KAAK9mB,SAAS;QACnD,IAAI8R,IAAiC;QACrC,IAAIwV,IAAY,GAAG;YAvuBN7pB,GAyuBK,MAAd6pB;YAGF,MAAMR,IAAO5K,EAAM4K,KAAM;YACrBA,EAAKC,iBACPjV,IAAkBgV,EAAKhkB,eAEvBnB,IAAOA,EAAKyP,MAAM0V,EAAKhkB;;QAI3B,IAAIykB,IAAqB;QACrBrL,EAAM+K,UACRM,IAAWntB,KAAKotB,GAAWtL,EAAM+K;QAGnC,IAAIlV,IAAqB;QACrBmK,EAAMnK,YACRA,IAAU3X,KAAKqtB,GAAUvL,EAAMnK;QAGjC,IAAIxR,IAAuB;QACvB2b,EAAM3b,UACRA,IAAQnG,KAAKstB,GAAexL,EAAM3b;QAGpC,IAAI0R,IAAwB;QACxBiK,EAAMjK,YACRA,IAAU7X,KAAKutB,GAAWzL,EAAMjK;QAGlC,IAAIC,IAAsB;QAK1B,OAJIgK,EAAMhK,UACRA,IAAQ9X,KAAKutB,GAAWzL,EAAMhK,SAGzB,IAAII,GACT3Q,GACAmQ,GACAC,GACAwV,GACAhnB,qBAEA0R,GACAC,GACAuB;;IAGJvZ,GACEimB;QAEA,MAAMlkB,IAAQ7B,KAAKwtB,GAAQzH,EAAWjK;QACtC,OAAa,QAATja,IACK,OAEA;YACL4rB,oBAAoB5rB;;;IAKlB/B,GAAQgc;QACd,QAAQA;UACN;YACE,OAAO;;UACT;YACE,OAAO;;UACT;YACE,OAAO;;UACT;YACE,OAp0BQhW;;;IAw0BdhG,GAASimB;QACP,IAAI5W;QACJ,MAAMyM,IAASmK,EAAWnK;QAc1B,OAXEzM,IADEyM,EAAO1B,OACA;YAAEoI,WAAWtiB,KAAK0tB,GAAkB9R;YAEpC;YAAEkG,OAAO9hB,KAAK2tB,GAAc/R;WAGvCzM,EAAO0M,WAAWkK,EAAWlK,UAEzBkK,EAAW7J,YAAYoI,OAAwB,MACjDnV,EAAO+M,cAAclc,KAAK4tB,GAAQ7H,EAAW7J;QAGxC/M;;IAGDrP,GAAS8X;QACf,IAAuB,MAAnBA,EAAQhS,QACV;QAEF,MAAMgmB,IAAShU,EAAQ9P,IAAIJ,KACrBA,aAAkBoS,KACb9Z,KAAK6tB,GAAqBnmB,KAj2BzB5B;QAs2BZ,OAAsB,MAAlB8lB,EAAOhmB,SACFgmB,EAAO,KAET;YAAEkC,iBAAiB;gBAAE7T,IAAI;gBAAOrC,SAASgU;;;;IAG1C9rB,GAAW4H;QACjB,OAAKA,SAE6B7B,MAAvB6B,EAAOqmB,cACT,EAAC/tB,KAAKguB,GAAgBtmB,YACG7B,MAAvB6B,EAAOumB,cACT,EAACjuB,KAAKkuB,GAAgBxmB,YACO7B,MAA3B6B,EAAOomB,kBACTpmB,EAAOomB,gBACXlW,QAAS9P,IAAIkQ,KAAKhY,KAAKotB,GAAWpV,IAClCmW,OAAO,CAACC,GAAOlmB,MAAYkmB,EAAMlV,OAAOhR,MAt3BjCpC,OA82BH;;IAcHhG,GAAQsa;QACd,IAAwB,MAApBA,EAASxU,QAGb,OAAOwU,EAAStS,IAAIumB,KAASruB,KAAKsuB,GAAgBD;;IAG5CvuB,GAAUsa;QAChB,OAAOA,EAAStS,IAAIumB,KAASruB,KAAKuuB,GAAkBF;;IAG9CvuB,GAAS0uB;QACf,OAAO;YACLjU,QAAQiU,EAAOjU;YACfhO,QAAQiiB,EAAOlU;;;IAIXxa,GAAW0uB;QACjB,MAAMjU,MAAWiU,EAAOjU,QAClBD,IAAWkU,EAAOjiB,UAAU;QAClC,OAAO,IAAI8N,GAAMC,GAAUC;;;IAI7Bza,GAAYkZ;QACV,OAAOuO,GAAWvO;;;IAIpBlZ,GAAckZ;QACZ,QAAQA;UACN,KAAK;YACH;;UACF,KAAK;YACH;;UACF;YACE;;;;IAKNlZ,GAAema;QACb,OAAO0N,GAAU1N;;IAGnBna,GAAiBma;QACf,QAAQA;UACN,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;YACH;;UACF,KAAK;UAEL;YACE,OA/7BQnU;;;IAm8BdhG,GAAqByH;QACnB,OAAO;YAAE6K,WAAW7K,EAAKD;;;IAG3BxH,GAAuB2uB;QACrB,OAAO7mB,EAAUukB,EAAiBsC,EAAyB;;;IAI7D3uB,GAAgB6X;QACd,OAAO;YACLjF,OAAO1S,KAAK0uB,GAAqB/W,EAAQjF;YACzCic,WAAW3uB,KAAK4uB,GAAYjX,EAAQqB;;;IAIxClZ,GAAkB6X;QAChB,OAAO,IAAIiB,GACT5Y,KAAK6uB,GAAuBlX,EAAc,QAC1C3X,KAAK8uB,GAAcnX,EAAQgX;;IAI/B7uB,GAAgB4H;QACd,OAAOoS,GAAYiV,OACjB/uB,KAAK6uB,GAAuBnnB,EAAOumB,YAAmB,QACtDjuB,KAAKgvB,GAAiBtnB,EAAOumB,YAAgB,KAC7CvmB,EAAOumB,YAAmB;;;IAK9BnuB,GAAqB4H;QACnB,yBAAIA,EAAOuS,IAAuB;YAChC,IAAIzJ,GAAW9I,EAAO7F,QACpB,OAAO;gBACLksB,aAAa;oBACXrb,OAAO1S,KAAK0uB,GAAqBhnB,EAAOgL;oBACxCuH,IAAI;;;YAGH,IAAI1J,EAAY7I,EAAO7F,QAC5B,OAAO;gBACLksB,aAAa;oBACXrb,OAAO1S,KAAK0uB,GAAqBhnB,EAAOgL;oBACxCuH,IAAI;;;;QAKZ,OAAO;YACLgU,aAAa;gBACXvb,OAAO1S,KAAK0uB,GAAqBhnB,EAAOgL;gBACxCuH,IAAIja,KAAKivB,GAAevnB,EAAOuS;gBAC/BpY,OAAO6F,EAAO7F;;;;IAKpB/B,GAAgB4H;QACd,QAAQA,EAAOqmB,YAAgB;UAC7B,KAAK;YACH,MAAMmB,IAAWlvB,KAAK6uB,GACpBnnB,EAAOqmB,YAAmB;YAE5B,OAAOjU,GAAYiV,OAAOG,sBAA0B;gBAClD/hB,aAAagiB;;;UAEjB,KAAK;YACH,MAAMC,IAAYpvB,KAAK6uB,GACrBnnB,EAAOqmB,YAAmB;YAE5B,OAAOjU,GAAYiV,OAAOK,sBAA2B;gBACnDC,WAAW;;;UAEf,KAAK;UAEL;YACE,OAjhCQvpB;;;IAqhCdhG,GAAemU;QACb,MAAMqb,IAA4B;QAIlC,OAHArb,EAAUzJ,OAAOpE,QAAQsM,KACvB4c,EAAgBhpB,KAAKoM,EAAMpL,OAEtB;YACLioB,YAAYD;;;IAIhBxvB,GAAiBgW;QACf,MAAM0Z,IAAQ1Z,EAAMyZ,cAAc;QAClC,OAAO,IAAIpd,GAAUqd,EAAM1nB,IAAIP,KAAQK,EAAUukB,EAAiB5kB;;;;SAItD4hB,GAAoB5hB;;IAElC,OACEA,EAAK3B,UAAU,KACC,eAAhB2B,EAAKxE,IAAI,MACO,gBAAhBwE,EAAKxE,IAAI;;;;;;;;;;;;;;;;;;;;;;;UCtgCA0G;IAEX3J,UAAmB2vB;QACbhmB,GAAgBgmB,YAClB3pB,MAEF2D,GAAgBgmB,WAAWA;;IAG7B3vB;QAIE,OAHK2J,GAAgBgmB,YACnB3pB,MAEK2D,GAAgBgmB;;;;;;;;;;;;;;;;;;;GCzD3B,OAAMC,KAAY,IAAIC,EAAO;;;SAGbC;IACd,OAAOF,GAAUG;;;SAGHC,GAAYC;IAC1BL,GAAUG,WAAWE;;;SAGP1I,GAAS2I,MAAgBlnB;IACvC,IAAI4mB,GAAUG,YAAYI,EAASC,OAAO;QACxC,MAAMC,IAAOrnB,EAAIhB,IAAIsoB;QACrBV,GAAUW,MAAM,cAAc1wB,OAAiBqwB,QAAUG;;;;SAI7C3T,GAASwT,MAAgBlnB;IACvC,IAAI4mB,GAAUG,YAAYI,EAASK,OAAO;QACxC,MAAMH,IAAOrnB,EAAIhB,IAAIsoB;QACrBV,GAAUa,MAAM,cAAc5wB,OAAiBqwB,QAAUG;;;;;;GAO7D,UAASC,GAAYtnB;IACnB,IAAmB,mBAARA,GACT,OAAOA;IACF;QACL,MAAM2mB,IAAWhmB,GAAgBC;QACjC;YACE,OAAO+lB,EAASe,GAAW1nB;UAC3B,OAAO2nB;;YAEP,OAAO3nB;;;;;;;;;;;;;;;;;;;;;;;;;;;;aChCGhD,GAAK4qB,IAAkB;;;IAGrC,MAAMlvB,IACJ,cAAc7B,mCAA6C+wB;;;;IAM7D,MALAlU,GAAShb,IAKH,IAAIF,MAAME;;;;;;;;aASF6B,GACdstB,GACAnvB;IAEKmvB,KACH7qB;;;;;;aAyBY8qB,GACd9nB;;AAEA+nB;IAMA,OAAO/nB;;;;;;;;;;;;;;;;;;UC7DIgoB;IACXhxB;;QAEE,MAAMixB,IACJ,kEAEIC,IAAcnsB,KAAKC,MAAM,MAAMisB,EAAMnrB,UAAUmrB,EAAMnrB;;gBAM3D,IAAIqrB,IAAS;QAEb,MAAOA,EAAOrrB,SADO,MACgB;YACnC,MAAMijB,IAAQpf,GAAgBC,KAAcwnB,GAAY;YACxD,KAAK,IAAIxqB,IAAI,GAAGA,IAAImiB,EAAMjjB,UAAUc;;;YAG9BuqB,EAAOrrB,SANM,MAMmBijB,EAAMniB,KAAKsqB,MAC7CC,KAAUF,EAAMI,OAAOtI,EAAMniB,KAAKqqB,EAAMnrB;;QAM9C,OAAOqrB;;;;SAIKhsB,GAAuBgC,GAASC;IAC9C,OAAID,IAAOC,KACD,IAEND,IAAOC,IACF,IAEF;;;0DAQOmF,GACdpF,GACAC,GACAlB;IAEA,OAAIiB,EAAKrB,WAAWsB,EAAMtB,UAGnBqB,EAAKmqB,MAAM,CAACvvB,GAAO4E,MAAUT,EAAWnE,GAAOqF,EAAMT;;;;;;;;;;;;;;;;;;UC3DjD4qB;;;;;;;;;;;;;IAaXvxB,YACWoQ,GACAohB,GACAC,GACAC,GACAC;kBAJAvhB,GACAlQ,sBAAAsxB,GACAtxB,YAAAuxB,GACAvxB,WAAAwxB,GACAxxB,wBAAAyxB;;;;;;MAQAC;IAEX5xB,YAAqBqQ,GAAmBC;QAAnBpQ,iBAAAmQ,GACnBnQ,KAAKoQ,WAAWA,KANU;;IAS5BuhB;QACE,OAV0B,gBAUnB3xB,KAAKoQ;;IAGdtQ,QAAQkF;QACN,OACEA,aAAiB0sB,MACjB1sB,EAAMmL,cAAcnQ,KAAKmQ,aACzBnL,EAAMoL,aAAapQ,KAAKoQ;;IAI5BtQ,EAAUkF;QACR,OACEC,GAAoBjF,KAAKmQ,WAAWnL,EAAMmL,cAC1ClL,GAAoBjF,KAAKoQ,UAAUpL,EAAMoL;;;;;;;;;;;;;;;;;;;;;;;;;UCtClCwhB;IAWX9xB,YAAoB+xB;kBAAAA;;;;;;;QAJpB7xB,UAEI;;2EAKJF,IAAIkJ;QACF,MAAM8oB,IAAK9xB,KAAK6xB,GAAS7oB,IACnB2R,IAAU3a,KAAK+xB,GAAMD;QAC3B,SAAgBjsB,MAAZ8U,GAGJ,KAAK,OAAOqX,GAAUnwB,MAAU8Y,GAC9B,IAAIqX,EAASxsB,QAAQwD,IACnB,OAAOnH;;IAMb/B,IAAIkJ;QACF,YAAyBnD,MAAlB7F,KAAK+C,IAAIiG;;iDAIlBlJ,IAAIkJ,GAAcnH;QAChB,MAAMiwB,IAAK9xB,KAAK6xB,GAAS7oB,IACnB2R,IAAU3a,KAAK+xB,GAAMD;QAC3B,SAAgBjsB,MAAZ8U,GAAJ;YAIA,KAAK,IAAIjU,IAAI,GAAGA,IAAIiU,EAAQ/U,QAAQc,KAClC,IAAIiU,EAAQjU,GAAG,GAAGlB,QAAQwD,IAExB,aADA2R,EAAQjU,KAAK,EAACsC,GAAKnH;YAIvB8Y,EAAQrU,KAAK,EAAC0C,GAAKnH;eATjB7B,KAAK+xB,GAAMD,KAAM,EAAC,EAAC9oB,GAAKnH;;;;WAe5B/B,OAAOkJ;QACL,MAAM8oB,IAAK9xB,KAAK6xB,GAAS7oB,IACnB2R,IAAU3a,KAAK+xB,GAAMD;QAC3B,SAAgBjsB,MAAZ8U,GACF,QAAO;QAET,KAAK,IAAIjU,IAAI,GAAGA,IAAIiU,EAAQ/U,QAAQc,KAClC,IAAIiU,EAAQjU,GAAG,GAAGlB,QAAQwD,IAMxB,OALuB,MAAnB2R,EAAQ/U,gBACH5F,KAAK+xB,GAAMD,KAElBnX,EAAQsX,OAAOvrB,GAAG;SAEb;QAGX,QAAO;;IAGT5G,QAAQ8G;QACNR,EAAQpG,KAAK+xB,IAAO,CAACnM,GAAGtP;YACtB,KAAK,OAAOmH,GAAG1Q,MAAMuJ,GACnB1P,EAAG6W,GAAG1Q;;;IAKZjN;QACE,OAAOuJ,EAAQrJ,KAAK+xB;;;;;;;;;;;;;;;;;;;;;;;MCrEXG;;;;;;;;;;;;IAYXpyB,YACSqyB,GACAxnB,GACAynB,GACAC;QAHAryB,eAAAmyB,aACAxnB,GACA3K,qBAAAoyB,GACApyB,iBAAAqyB;;;;;;;;;;WAcTvyB,GACEwyB,GACAnf,GACAof;QAUA,MAAMC,IAAkBD,EAAYC;QAQpC,KAAK,IAAI9rB,IAAI,GAAGA,IAAI1G,KAAKqyB,UAAUzsB,QAAQc,KAAK;YAC9C,MAAMkkB,IAAW5qB,KAAKqyB,UAAU3rB;YAChC,IAAIkkB,EAAS5hB,IAAIxD,QAAQ8sB,IAAS;gBAChC,MAAM9e,IAAiBgf,EAAgB9rB;gBACvCyM,IAAWyX,EAASnV,GAAsBtC,GAAUK;;;QAGxD,OAAOL;;;;;;;;WAUTrT,GACEwyB,GACAnf;;;QAYA,KAAK,MAAMyX,KAAY5qB,KAAKoyB,eACtBxH,EAAS5hB,IAAIxD,QAAQ8sB,OACvBnf,IAAWyX,EAASlV,GAClBvC,GACAA,GACAnT,KAAK2K;QAKX,MAAMgJ,IAAUR;;gBAGhB,KAAK,MAAMyX,KAAY5qB,KAAKqyB,WACtBzH,EAAS5hB,IAAIxD,QAAQ8sB,OACvBnf,IAAWyX,EAASlV,GAClBvC,GACAQ,GACA3T,KAAK2K;QAIX,OAAOwI;;;;;WAOTrT,GAAwB2yB;;;;QAItB,IAAIC,IAAmBD;QAUvB,OATAzyB,KAAKqyB,UAAUjsB,QAAQusB;YACrB,MAAMC,IAAkB5yB,KAAK0V,GAC3Bid,EAAE3pB,KACFypB,EAAU1vB,IAAI4vB,EAAE3pB;YAEd4pB,MACFF,IAAmBA,EAAiB5V,GAAO6V,EAAE3pB,KAAK4pB;YAG/CF;;IAGT5yB;QACE,OAAOE,KAAKqyB,UAAUlE,OACpB,CAAC5f,GAAMokB,MAAMpkB,EAAKuR,IAAI6S,EAAE3pB,MACxB4X;;IAIJ9gB,QAAQkF;QACN,OACEhF,KAAKmyB,YAAYntB,EAAMmtB,WACvB9lB,GAAYrM,KAAKqyB,WAAWrtB,EAAMqtB,WAAW,CAAC9f,GAAGC,MAAMD,EAAE/M,QAAQgN,OACjEnG,GAAYrM,KAAKoyB,eAAeptB,EAAMotB,eAAe,CAAC7f,GAAGC,MACvDD,EAAE/M,QAAQgN;;;;qEAOLqgB;IACX/yB,YACWgzB,GACAC,GACAP,GACAQ;;;;;IAKAC;QARAjzB,aAAA8yB,aACAC,aACAP,GACAxyB,mBAAAgzB,aAKAC;;;;;;WAQXnzB,YACEgzB,GACAC,GACAG,GACAF;QAnKiC3vB,GAsK/ByvB,EAAMT,UAAUzsB,WAAWstB,EAAQttB;QAOrC,IAAIutB,IAAazS;QACjB,MAAM2R,IAAYS,EAAMT;QACxB,KAAK,IAAI3rB,IAAI,GAAGA,IAAI2rB,EAAUzsB,QAAQc,KACpCysB,IAAaA,EAAWrW,GAAOuV,EAAU3rB,GAAGsC,KAAKkqB,EAAQxsB,GAAGmM;QAG9D,OAAO,IAAIggB,GACTC,GACAC,GACAG,GACAF,GACAG;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;UC5LOC;IAeXtzB,YAAYuzB;;;QAZZrzB,UAAqD,MACrDA,UAAkD;;QAG1CA,mBAAwB6F,GACxB7F,kBAA2B6F,GACnC7F,WAAiB;;;QAIjBA,WAA2B,GAGzBqzB,EACExxB;YACE7B,KAAKszB,MAAS,GACdtzB,KAAKmP,SAAStN,GACV7B,KAAKuzB;;;YAGPvzB,KAAKuzB;WAGThD;YACEvwB,KAAKszB,MAAS,GACdtzB,KAAKuwB,QAAQA,GACTvwB,KAAKwzB,MACPxzB,KAAKwzB,GAAcjD;;;IAM3BzwB,MACE8G;QAEA,OAAO5G,KAAKsI,UAAKzC,GAAWe;;IAG9B9G,KACE2zB,GACAC;QAMA,OAJI1zB,KAAK2zB,MACP7tB,MAEF9F,KAAK2zB,MAAmB,GACpB3zB,KAAKszB,KACFtzB,KAAKuwB,QAGDvwB,KAAK4zB,GAAYF,GAAS1zB,KAAKuwB,SAF/BvwB,KAAK6zB,GAAYJ,GAAQzzB,KAAY,UAKvC,IAAIozB,GAAsB,CAAClxB,GAAS4xB;YACzC9zB,KAAKuzB,KAAgB1xB;gBACnB7B,KAAK6zB,GAAYJ,GAAQ5xB,GAAOyG,KAAKpG,GAAS4xB;eAEhD9zB,KAAKwzB,KAAiBjD;gBACpBvwB,KAAK4zB,GAAYF,GAASnD,GAAOjoB,KAAKpG,GAAS4xB;;;;IAMvDh0B;QACE,OAAO,IAAImC,QAAQ,CAACC,GAAS4xB;YAC3B9zB,KAAKsI,KAAKpG,GAAS4xB;;;IAIfh0B,GACN8G;QAEA;YACE,MAAMuI,IAASvI;YACf,OAAIuI,aAAkBikB,KACbjkB,IAEAikB,GAAmBlxB,QAAQiN;UAEpC,OAAOshB;YACP,OAAO2C,GAAmBU,OAAUrD;;;IAIhC3wB,GACN2zB,GACA5xB;QAEA,OAAI4xB,IACKzzB,KAAK+zB,GAAiB,MAAMN,EAAO5xB,MAGnCuxB,GAAmBlxB,QAAYL;;IAIlC/B,GACN4zB,GACAnD;QAEA,OAAImD,IACK1zB,KAAK+zB,GAAiB,MAAML,EAAQnD,MAEpC6C,GAAmBU,OAAUvD;;IAMxCzwB,eAAkBqP;QAChB,OAAO,IAAIikB,GAA6B,CAAClxB,GAAS4xB;YAChD5xB,EAAQiN;;;IAIZrP,cAAiBywB;QACf,OAAO,IAAI6C,GAAsB,CAAClxB,GAAS4xB;YACzCA,EAAOvD;;;IAIXzwB;;;IAGEk0B;QAEA,OAAO,IAAIZ,GAAyB,CAAClxB,GAAS4xB;YAC5C,IAAIhO,IAAgB,GAChBmO,IAAgB,GAChBC,KAAO;YAEXF,EAAI5tB,QAAQkL;kBACRwU,GACFxU,EAAQhJ,KACN;sBACI2rB,GACEC,KAAQD,MAAkBnO,KAC5B5jB;mBAGJiyB,KAAOL,EAAOK;gBAIlBD,KAAO,GACHD,MAAkBnO,KACpB5jB;;;;;;;;WAWNpC,UACEs0B;QAEA,IAAIhZ,IAAiCgY,GAAmBlxB,SACtD;QAEF,KAAK,MAAMmyB,KAAaD,GACtBhZ,IAAIA,EAAE9S,KAAKgsB,KACLA,IACKlB,GAAmBlxB,QAAiBoyB,KAEpCD;QAIb,OAAOjZ;;IAkBTtb,eACEy0B,GACAvc;QAEA,MAAMwc,IAA4C;QAIlD,OAHAD,EAAWnuB,QAAQ,CAACoM,GAAGiiB;YACrBD,EAASluB,KAAK0R,EAAE5O,KAAKpJ,MAAMwS,GAAGiiB;YAEzBz0B,KAAK00B,GAAQF;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCpMXG;IACX70B,YACW80B,GACAC,GACAC;kBAFAF,aACAC,aACAC;;;;;;;WASXh1B,GACEi1B,GACA/rB;QAEA,OAAOhJ,KAAK60B,GACTG,GAA0CD,GAAa/rB,GACvDV,KAAK2sB,KAAWj1B,KAAKk1B,GAAoBH,GAAa/rB,GAAKisB;;6EAIxDn1B,GACNi1B,GACA/rB,GACAmsB;QAEA,OAAOn1B,KAAK40B,GAAoBQ,GAASL,GAAa/rB,GAAKV,KAAK0M;YAC9D,KAAK,MAAM8d,KAASqC,GAClBngB,IAAM8d,EAAMpd,GAAiB1M,GAAKgM;YAEpC,OAAOA;;;;;IAMHlV,GACNi1B,GACAhT,GACAkT;QAEA,IAAI/B,IAAU5S;QAOd,OANAyB,EAAK3b,QAAQ,CAAC4C,GAAKqsB;YACjB,KAAK,MAAMvC,KAASmC,GAClBI,IAAYvC,EAAMpd,GAAiB1M,GAAKqsB;YAE1CnC,IAAUA,EAAQpW,GAAO9T,GAAKqsB;YAEzBnC;;;;;;;WASTpzB,GACEi1B,GACAxmB;QAEA,OAAOvO,KAAK40B,GACTU,WAAWP,GAAaxmB,GACxBjG,KAAKyZ,KAAQ/hB,KAAKu1B,GAAwBR,GAAahT;;;;;WAO5DjiB,GACEi1B,GACAS;QAEA,OAAOx1B,KAAK60B,GACTY,GAA2CV,GAAaS,GACxDltB,KAAK2sB;YACJ,MAAMlT,IAAO/hB,KAAK01B,GAChBX,GACAS,GACAP;YAEF,IAAI/B,IAAU7S;YASd,OARA0B,EAAK3b,QAAQ,CAAC4C,GAAKmK;;gBAEZA,MACHA,IAAW,IAAIyC,GAAW5M,GAAK3D,EAAgB2B,SAEjDksB,IAAUA,EAAQpW,GAAO9T,GAAKmK;gBAGzB+f;;;;;;;;;;WAYbpzB,GACEi1B,GACAjT,GACA6T;QAEA,OAAI7T,EAAM5H,OACDla,KAAK41B,GAAkCb,GAAajT,EAAMva,QACxDua,EAAM+T,OACR71B,KAAK81B,GACVf,GACAjT,GACA6T,KAGK31B,KAAK+1B,GACVhB,GACAjT,GACA6T;;IAKE71B,GACNi1B,GACAva;;QAGA,OAAOxa,KAAKg2B,GAAYjB,GAAa,IAAIxsB,EAAYiS,IAAUlS,KAC7D6K;YACE,IAAIhE,IAASqR;YAIb,OAHIrN,aAAoBC,OACtBjE,IAASA,EAAO2N,GAAO3J,EAASnK,KAAKmK,KAEhChE;;;IAKLrP,GACNi1B,GACAjT,GACA6T;QAMA,MAAMjtB,IAAeoZ,EAAMpK;QAC3B,IAAIwb,IAAU1S;QACd,OAAOxgB,KAAK80B,GACTmB,GAAqBlB,GAAarsB,GAClCJ,KAAK4tB,KAGG9C,GAAmBhtB,QAAQ8vB,GAAUzJ;YAC1C,MAAM0J,IAAkBrU,EAAMsU,GAC5B3J,EAAOzV,MAAMtO;YAEf,OAAO1I,KAAK+1B,GACVhB,GACAoB,GACAR,GACArtB,KAAKkK;gBACLA,EAAEpM,QAAQ,CAAC4C,GAAKgM;oBACdke,IAAUA,EAAQpW,GAAO9T,GAAKgM;;;WAGjC1M,KAAK,MAAM4qB;;IAIZpzB,GACNi1B,GACAjT,GACA6T;;QAGA,IAAIzC,GACAmD;QACJ,OAAOr2B,KAAK40B,GACT0B,GAA0BvB,GAAajT,GAAO6T,GAC9CrtB,KAAKiuB,MACJrD,IAAUqD,GACHv2B,KAAK60B,GAAc2B,GACxBzB,GACAjT,KAGHxZ,KAAKmuB,MACJJ,IAAkBI;QAOXz2B,KAAK02B,GACV3B,GACAsB,GACAnD,GACA5qB,KAAKquB;YACLzD,IAAUyD;YAEV,KAAK,MAAM7D,KAASuD,GAClB,KAAK,MAAMzL,KAAYkI,EAAMT,WAAW;gBACtC,MAAMrpB,IAAM4hB,EAAS5hB,KACf2K,IAAUuf,EAAQnwB,IAAIiG,IACtB4tB,IAAahM,EAASlV,GAC1B/B,GACAA,GACAmf,EAAMnoB;gBAGNuoB,IADE0D,aAAsBxjB,KACd8f,EAAQpW,GAAO9T,GAAK4tB,KAEpB1D,EAAQjW,OAAOjU;;aAMlCV,KAAK;;;QAGJ4qB,EAAQ9sB,QAAQ,CAAC4C,GAAKgM;YACf8M,EAAMnH,QAAQ3F,OACjBke,IAAUA,EAAQjW,OAAOjU;YAItBkqB;;IAILpzB,GACNi1B,GACA0B,GACAI;QAEA,IAAIC,IAAmClW;QACvC,KAAK,MAAMkS,KAAS2D,GAClB,KAAK,MAAM7L,KAAYkI,EAAMT,WAEzBzH,aAAoB7W,MACoB,SAAxC8iB,EAAkB9zB,IAAI6nB,EAAS5hB,SAE/B8tB,IAAmCA,EAAiChX,IAClE8K,EAAS5hB;QAMjB,IAAI2tB,IAAkBE;QACtB,OAAO72B,KAAK40B,GACTU,WAAWP,GAAa+B,GACxBxuB,KAAKyuB,MACJA,EAAgB3wB,QAAQ,CAAC4C,GAAKgM;YAChB,SAARA,KAAgBA,aAAe5B,OACjCujB,IAAkBA,EAAgB7Z,GAAO9T,GAAKgM;YAG3C2hB;;;;;;;;;;;;;;;;;;;;;;;;UCjSFK;IACXl3B,YACW+b,GACAsG,GACA8U,GACAC;QAHAl3B,gBAAA6b,GACA7b,iBAAAmiB,aACA8U,aACAC;;IAGXp3B,UACE+b,GACAsb;QAEA,IAAIF,IAAYrW,MACZsW,IAActW;QAElB,KAAK,MAAMgE,KAAauS,EAAalV,YACnC,QAAQ2C,EAAUjD;UAChB;YACEsV,IAAYA,EAAUnX,IAAI8E,EAAU5P,IAAIhM;YACxC;;UACF;YACEkuB,IAAcA,EAAYpX,IAAI8E,EAAU5P,IAAIhM;;;QAOlD,OAAO,IAAIguB,GACTnb,GACAsb,EAAahV,WACb8U,GACAC;;;;;;;;;;;;;;;;;;;;;;;;;UCnBOE;IAOXt3B,YACU6Q,GACR0mB;QADQr3B,qBAAA2Q,GAGJ0mB,MACFA,EAAqBC,KAAwBvb,KAC3C/b,KAAKu3B,GAAiBxb,IACxB/b,KAAKw3B,KAAyBzb,KAC5Bsb,EAAqBI,GAAoB1b;;IAIvCjc,GACN43B;QAGA,OADA13B,KAAK2Q,gBAAgB9L,KAAK8yB,IAAID,GAAuB13B,KAAK2Q,gBACnD3Q,KAAK2Q;;IAGd7Q;QACE,MAAM83B,MAAc53B,KAAK2Q;QAIzB,OAHI3Q,KAAKw3B,MACPx3B,KAAKw3B,GAAuBI,IAEvBA;;;;AA9BTR,SAAiD;;;;;;;;;;;;;;;;;;MCftCS;IAMX/3B;QACEE,KAAK83B,UAAU,IAAI71B,QAAQ,CAACC,GAAsB4xB;YAChD9zB,KAAKkC,UAAUA,GACflC,KAAK8zB,SAASA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCQPiE;IAMXj4B;;;;IAImBk4B;;;;IAIAC;;;;;;IAMAC,IApCoB;;;;UAyCpBC,IAvCU;;;;;UA6CVC,IA1CgB;kBAqBhBJ,aAIAC,aAMAC,aAKAC,aAMAC,GA9BnBp4B,UAAgC,GAChCA,UAAsD;;QAEtDA,UAA0BwE,KAAKC,OA6B7BzE,KAAKq4B;;;;;;;;WAUPv4B;QACEE,KAAKs4B,KAAgB;;;;;WAOvBx4B;QACEE,KAAKs4B,KAAgBt4B,KAAKo4B;;;;;;WAQ5Bt4B,GAAcma;;QAEZja,KAAKu4B;;;QAIL,MAAMC,IAA2B3zB,KAAKC,MACpC9E,KAAKs4B,KAAgBt4B,KAAKy4B,OAItBC,IAAe7zB,KAAK8yB,IAAI,GAAGnzB,KAAKC,QAAQzE,KAAK24B,KAG7CC,IAAmB/zB,KAAK8yB,IAC5B,GACAa,IAA2BE;;gBAGzBE,IAAmB,KACrBvR,GAtGU,sBAwGR,mBAAmBuR,UACjB,gBAAgB54B,KAAKs4B,YACrB,sBAAsBE,WACtB,iBAAiBE;QAIvB14B,KAAK64B,KAAe74B,KAAKg4B,GAAMc,GAC7B94B,KAAKi4B,IACLW,GACA,OACE54B,KAAK24B,KAAkBn0B,KAAKC,OACrBwV;;;QAMXja,KAAKs4B,MAAiBt4B,KAAKm4B,IACvBn4B,KAAKs4B,KAAgBt4B,KAAKk4B,OAC5Bl4B,KAAKs4B,KAAgBt4B,KAAKk4B,KAExBl4B,KAAKs4B,KAAgBt4B,KAAKo4B,OAC5Bp4B,KAAKs4B,KAAgBt4B,KAAKo4B;;IAI9Bt4B;QAC4B,SAAtBE,KAAK64B,OACP74B,KAAK64B,GAAaE,MAClB/4B,KAAK64B,KAAe;;IAIxB/4B;QAC4B,SAAtBE,KAAK64B,OACP74B,KAAK64B,GAAaN,UAClBv4B,KAAK64B,KAAe;;sFAKhB/4B;QACN,QAAQ+E,KAAKm0B,WAAW,MAAOh5B,KAAKs4B;;;;;;;;;;;;;;;;;;;;;;UC5I3BW;IAAbn5B;QACEE,UAAgC,IAAIk5B;;IAEpCp5B,GACEi1B,GACAoE;QAGA,OADAn5B,KAAKo5B,GAAsBtZ,IAAIqZ,IACxB/F,GAAmBlxB;;IAG5BpC,GACEi1B,GACArsB;QAEA,OAAO0qB,GAAmBlxB,QACxBlC,KAAKo5B,GAAsB9D,WAAW5sB;;;;;;;;UAU/BwwB;IAAbp5B;QACUE,aAAQ;;;IAKhBF,IAAIq5B;QAEF,MAAMzwB,IAAeywB,EAAepjB,KAC9BsjB,IAAaF,EAAevM,KAC5B0M,IACJt5B,KAAKyG,MAAMiC,MACX,IAAIwW,GAAwB/X,EAAanB,IACrCuzB,KAASD,EAAgBzZ,IAAIwZ;QAEnC,OADAr5B,KAAKyG,MAAMiC,KAAgB4wB,EAAgBxZ,IAAIuZ,IACxCE;;IAGTz5B,IAAIq5B;QACF,MAAMzwB,IAAeywB,EAAepjB,KAC9BsjB,IAAaF,EAAevM,KAC5B0M,IAAkBt5B,KAAKyG,MAAMiC;QACnC,OAAO4wB,KAAmBA,EAAgBzZ,IAAIwZ;;IAGhDv5B,WAAW4I;QAIT,QAFE1I,KAAKyG,MAAMiC,MACX,IAAIwW,GAAwB/X,EAAanB,IACxBoB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MC9CVoyB;IACX15B,YAAoB25B;kBAAAA;;IAEpB35B;QAEE,OADAE,KAAKy5B,MApBM,GAqBJz5B,KAAKy5B;;IAGd35B;;;;;QAKE,OAAO,IAAI05B,GAAkB;;IAG/B15B;;QAEE,OAAO,IAAI05B,IAAkB;;;;;;;;;;;;;;;;;;;;sEC+WjBE,GAA4BjJ;;;IAG1C,OAAkB,gCAAXA,EAAE9uB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MC3UEg4B;IAOX75B,YACmB85B,GACR3B,GACA4B,GACQ5f,GACA6f;kBAJAF,aACR3B,aACA4B,GACQ75B,UAAAia,aACA6f,GAPnB95B,UAA4B,IAAI63B;QAmFhC73B,YAAOA,KAAK+5B,GAASjC,QAAQ90B,KAAKg3B,KAAKh6B,KAAK+5B,GAASjC;;;;QAvEnD93B,KAAK+5B,GAASjC,QAAQmC,MAAM9F;;;;;;;;;;;;;;;WAiB9Br0B,UACE85B,GACA3B,GACAiC,GACAjgB,GACA6f;QAEA,MAAMK,IAAa31B,KAAKC,QAAQy1B,GAC1BE,IAAY,IAAIT,GACpBC,GACA3B,GACAkC,GACAlgB,GACA6f;QAGF,OADAM,EAAU1a,MAAMwa,IACTE;;;;;WAODt6B,MAAMo6B;QACZl6B,KAAKq6B,KAAcC,WAAW,MAAMt6B,KAAKu6B,MAAsBL;;;;;WAOjEp6B;QACE,OAAOE,KAAKu6B;;;;;;;;WAUdz6B,OAAO06B;QACoB,SAArBx6B,KAAKq6B,OACPr6B,KAAKy6B,gBACLz6B,KAAK+5B,GAASjG,OACZ,IAAIzyB,EACFlB,EAAKE,WACL,yBAAyBm6B,IAAS,OAAOA,IAAS;;IAQlD16B;QACNE,KAAK45B,GAAWc,GAAiB,MACN,SAArB16B,KAAKq6B,MACPr6B,KAAKy6B,gBACEz6B,KAAKia,KAAKjX,KAAKmM,KACbnP,KAAK+5B,GAAS73B,QAAQiN,OAGxBlN,QAAQC;;IAKbpC;QACmB,SAArBE,KAAKq6B,OACPr6B,KAAK85B,GAAgB95B,OACrBy6B,aAAaz6B,KAAKq6B,KAClBr6B,KAAKq6B,KAAc;;;;MAKZM;IAkCX76B;;QAhCAE,UAAiCiC,QAAQC;;;QAIzClC,UAAuCiC,QAAQC;;;QAI/ClC,WAAmC;;;QAInCA,UAA8D;;QAG9DA,UAAwB;;;QAIxBA,WAA8B;;QAG9BA,UAAoC;;QAGpCA,UAAkB,IAAI+3B,GAAmB/3B;;;;QAKzCA,UAA4B,MAAYA,KAAK46B,GAAQC;QAGnD,MAAMC,IAASrxB,GAAgBC,KAAcoxB;QACzCA,KAA6C,qBAA5BA,EAAOC,oBAC1BD,EAAOC,iBAAiB,oBAAoB/6B,KAAKg7B;;;;IAMrDC;QACE,OAAOj7B,KAAKk7B;;;;;WAOdp7B,GAAoCma;;QAElCja,KAAKm7B,QAAQlhB;;;;;WAOfna,GACEma;QAEAja,KAAKo7B;;QAELp7B,KAAKq7B,GAAgBphB;;;;;WAOfna,GACNma;QAGA,OADAja,KAAKo7B,MACEp7B,KAAKq7B,GAAgBphB;;;;;;;;WAU9Bna,SAAiCma;QAE/B,IADAja,KAAKo7B,OACAp7B,KAAKk7B,IAAiB;YACzBl7B,KAAKk7B,MAAkB;YACvB,MAAMJ,IAASrxB,GAAgBC,KAAcoxB;YACzCA,KACFA,EAAOQ,oBAAoB,oBAAoBt7B,KAAKg7B,WAEhDh7B,KAAKu7B,GAAyBthB;;;;;;WAQxCna,QAA2Bma;QAEzB,OADAja,KAAKo7B,MACDp7B,KAAKk7B,KAEA,IAAIj5B,QAAWC,WAEjBlC,KAAKq7B,GAAgBphB;;;;;;;;;WAW9Bna,GAAiBma;QACfja,KAAKo7B,MAEDp7B,KAAKk7B,OAITl7B,KAAKw7B,KAAgBx7B,KAAKw7B,GAAcx4B,KAAK;YAC3C,MAAM+2B,IAAW,IAAIlC,IACf4D,IAAaC;gBACjB;0BACQzhB,KACN8f,EAAS73B,WACTlC,KAAK46B,GAAQvC;kBACb,OAAO5H;oBACP,KAAIiJ,GAA4BjJ,IAK9B,MADAsJ,EAAS73B,WACHuuB;;wCAJNpJ,GA3TI,cA2Tc,4CAA4CoJ;oBAC9DzwB,KAAK46B,GAAQe,GAAcF;;;YAQjC,OADAz7B,KAAK06B,GAAiBe,IACf1B,EAASjC;;;IAIZh4B,GAAmCma;QACzC,MAAM2hB,IAAU57B,KAAK67B,GAAK74B,KAAK,OAC7BhD,KAAK87B,MAAsB,GACpB7hB,IACJggB,MAAO1J;;;;YASN,MARAvwB,KAAK0wB,KAAUH,GACfvwB,KAAK87B,MAAsB,GAE3Btf,GAAS,8BADO+T,EAAMwL,SAASxL,EAAM/uB,WAAW;YAM1C+uB;WAEPvtB,KAAKmM,MACJnP,KAAK87B,MAAsB,GACpB3sB;QAIb,OADAnP,KAAK67B,KAAOD,GACLA;;;;;;WAQT97B,GACEm4B,GACAiC,GACAjgB;QAEAja,KAAKo7B;;QAQDp7B,KAAKg8B,GAAex0B,QAAQywB,MAAY,MAC1CiC,IAAU;QAGZ,MAAME,IAAYT,GAAiBsC,GACjCj8B,MACAi4B,GACAiC,GACAjgB,GACAiiB,KACEl8B,KAAKm8B,GAAuBD;QAGhC,OADAl8B,KAAKo8B,GAAkB91B,KAAK8zB,IACrBA;;IAGDt6B;QACFE,KAAK0wB,MACP5qB;;;;;;;WAaJhG;;;;WAWAA;;;;;QAKE,IAAIu8B;QACJ;YACEA,IAAcr8B,KAAK67B,UACbQ;iBACCA,MAAgBr8B,KAAK67B;;;;;WAOhC/7B,GAAyBm4B;QACvB,KAAK,MAAMhe,KAAMja,KAAKo8B,IACpB,IAAIniB,EAAGge,OAAYA,GACjB,QAAO;QAGX,QAAO;;;;;;;;WAUTn4B,GAA6Bw8B;;QAE3B,OAAOt8B,KAAKu8B,KAAQv5B,KAAK;;YAEvBhD,KAAKo8B,GAAkB3tB,KAAK,CAAC+tB,GAAGC,MAAMD,EAAE3C,KAAe4C,EAAE5C;YAEzD,KAAK,MAAM5f,KAAMja,KAAKo8B,IAEpB,IADAniB,EAAG8e,0BACCuD,KAA+BriB,EAAGge,OAAYqE,GAChD;YAIJ,OAAOt8B,KAAKu8B;;;;;WAOhBz8B,GAAqBm4B;QACnBj4B,KAAKg8B,GAAe11B,KAAK2xB;;iEAInBn4B,GAAuBma;;QAE7B,MAAMxT,IAAQzG,KAAKo8B,GAAkB50B,QAAQyS;QAE7Cja,KAAKo8B,GAAkBnK,OAAOxrB,GAAO;;;;;;;aAQzBi2B,GACdjM,GACAT;IAGA,IADAxT,GArec,cAqeI,GAAGwT,MAAQS,MACzBiJ,GAA4BjJ,IAC9B,OAAO,IAAIpvB,EAAelB,EAAKgB,aAAa,GAAG6uB,MAAQS;IAEvD,MAAMA;;;;;;;;;;;;;;;;;;UCrVGkM;IA2BX78B;;;IAGW88B;;IAEAC;;;IAGAC;kBALAF,aAEAC,aAGAC;;IA5BXh9B,UAAqBi9B;QACnB,OAAO,IAAIJ,GACTI,GACAJ,GAAUK,IACVL,GAAUM;;;;AAVdN,SAAuC,GACvCA,QAA2C,SAC3CA,QAA2C,UAC3CA,QAAwD,IACxDA,QAAkE,KAUlEA,QAAqC,IAAIA,GACvCA,GAAUO,IACVP,GAAUK,IACVL,GAAUM;AAGIN,cAAsB,IAAIA,GACxCA,GAAUQ,IACV,GACA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCvDSC;IAmDXt9B;;IAEYu9B,GACFC,GACRC;QAFUv9B,mBAAAq9B,aACFC;;;;;;;QApBVt9B,UAA+B,IAAI0c,GACjCzX;;;QAKFjF,UAA2B,IAAI4xB,GAA4B4L,KACzDA,EAAE5uB;;;;;;QAQJ5O,UAAuCqF,EAAgB2B,OAYrDhH,KAAK60B,KAAgBwI,EAAYI,GAAiBF,IAClDv9B,KAAK09B,KAAkBL,EAAYM,MACnC39B,KAAK49B,KAAcP,EAAYQ,MAC/B79B,KAAK89B,KAAiB,IAAInJ,GACxB30B,KAAK09B,IACL19B,KAAK60B,IACL70B,KAAKq9B,YAAYU;QAEnB/9B,KAAKs9B,GAAYU,GAAsBh+B,KAAK89B;;qCAI9Ch+B;QACE,OAAOmC,QAAQC;;;;;;;;;;IAWjBpC,SAAuBgC;QACrB,IAAIm8B,IAAmBj+B,KAAK60B,IACxBqJ,IAAoBl+B,KAAK89B;QAE7B,MAAM3uB,UAAenP,KAAKq9B,YAAYc,eACpC,sBACA,YACAC;;;YAGE,IAAIC;YACJ,OAAOr+B,KAAK60B,GACTyJ,GAAsBF,GACtB91B,KAAKi2B,MACJF,IAAaE,GAEbN,IAAmBj+B,KAAKq9B,YAAYI,GAAiB37B;;;YAIrDo8B,IAAoB,IAAIvJ,GACtB30B,KAAK09B,IACLO,GACAj+B,KAAKq9B,YAAYU,OAEZE,EAAiBK,GAAsBF,KAE/C91B,KAAKk2B;gBACJ,MAAMC,IAA6B,IAC7BC,IAA2B;;gBAGjC,IAAIC,IAAc/d;gBAElB,KAAK,MAAMkS,KAASuL,GAAY;oBAC9BI,EAAgBn4B,KAAKwsB,EAAMX;oBAC3B,KAAK,MAAMvH,KAAYkI,EAAMT,WAC3BsM,IAAcA,EAAY7e,IAAI8K,EAAS5hB;;gBAI3C,KAAK,MAAM8pB,KAAS0L,GAAY;oBAC9BE,EAAcp4B,KAAKwsB,EAAMX;oBACzB,KAAK,MAAMvH,KAAYkI,EAAMT,WAC3BsM,IAAcA,EAAY7e,IAAI8K,EAAS5hB;;;;gCAM3C,OAAOk1B,EACJU,GAAaR,GAAKO,GAClBr2B,KAAKu2B,MACG;oBACLC,IAAAD;oBACAE,IAAAN;oBACAO,IAAAN;;;;QAWd,OAJA1+B,KAAK60B,KAAgBoJ,GACrBj+B,KAAK89B,KAAiBI,GACtBl+B,KAAKs9B,GAAYU,GAAsBh+B,KAAK89B,KAErC3uB;;4EAITrP,GAAWuyB;QACT,MAAM1nB,IAAiBvG,EAAUK,OAC3B8J,IAAO8jB,EAAUlE,OACrB,CAAC5f,GAAMokB,MAAMpkB,EAAKuR,IAAI6S,EAAE3pB,MACxB4X;QAGF,IAAIqe;QAEJ,OAAOj/B,KAAKq9B,YACTc,eAAe,2BAA2B,aAAaC,KAI/Cp+B,KAAK89B,GAAec,GAAaR,GAAK7vB,GAAMjG,KAAKyZ;YACtDkd,IAAeld;;;;;;YAOf,MAAMqQ,IAA4B;YAElC,KAAK,MAAMxH,KAAYyH,GAAW;gBAChC,MAAMzgB,IAAYgZ,EAASsU,GACzBD,EAAal8B,IAAI6nB,EAAS5hB;gBAEX,QAAb4I;;;;gBAIFwgB,EAAc9rB,KACZ,IAAIyN,GACF6W,EAAS5hB,KACT4I,GACAqF,GAAiBrF,EAAUkE,MAAe,WAC1C/C,GAAaE,QAAO;;YAM5B,OAAOjT,KAAK60B,GAAcsK,GACxBf,GACAzzB,GACAynB,GACAC;YAILrvB,KAAK8vB;YACJ,MAAMlR,IAAUkR,EAAMsM,GAAwBH;YAC9C,OAAO;gBAAE9M,SAASW,EAAMX;gBAASkN,IAAAzd;;;;;;;;;;;;;;;;;WAkBvC9hB,GACEyyB;QAEA,OAAOvyB,KAAKq9B,YAAYc,eACtB,qBACA,qBACAC;YACE,MAAMkB,IAAW/M,EAAYO,MAAMvkB,QAC7BgxB,IAAiBv/B,KAAK09B,GAAgB8B,GAAgB;gBAC1DC,KAAe;;YAEjB,OAAOz/B,KAAK60B,GACT6K,GAAiBtB,GAAK7L,EAAYO,OAAOP,EAAYS,aACrD1qB,KAAK,MACJtI,KAAK2/B,GAA4BvB,GAAK7L,GAAagN,IAEpDj3B,KAAK,MAAMi3B,EAAeruB,MAAMktB,IAChC91B,KAAK,MAAMtI,KAAK60B,GAAc+K,GAAwBxB,IACtD91B,KAAK,MAAMtI,KAAK89B,GAAec,GAAaR,GAAKkB;;;;;;;;WAW1Dx/B,GAAYqyB;QACV,OAAOnyB,KAAKq9B,YAAYc,eACtB,gBACA,qBACAC;YACE,IAAIyB;YACJ,OAAO7/B,KAAK60B,GACTiL,GAAoB1B,GAAKjM,GACzB7pB,KAAMwqB,MA1WUzvB,GA2WM,SAAVyvB,IACX+M,IAAe/M,EAAMvkB,QACdvO,KAAK60B,GAAckL,GAAoB3B,GAAKtL,KAEpDxqB,KAAK,MACGtI,KAAK60B,GAAc+K,GAAwBxB,IAEnD91B,KAAK,MACGtI,KAAK89B,GAAec,GAAaR,GAAKyB;;;;;;WAUvD//B;QACE,OAAOE,KAAKq9B,YAAYc,eACtB,uCACA,YACAC,KACSp+B,KAAK60B,GAAcmL,GAAgC5B;;2EAMhEt+B;QACE,OAAOE,KAAKq9B,YAAYc,eACtB,yBACA,YACAC,KACSp+B,KAAK60B,GAAcoL,GAAmB7B;;;;;;WAUnDt+B,GAAmBkzB;QACjB,OAAOhzB,KAAKq9B,YAAYc,eACtB,yBACA,qBACAC,KACSp+B,KAAK60B,GAAcqL,GAAmB9B,GAAKpL;;;;;WASxDlzB;QACE,OAAOE,KAAKq9B,YAAYc,eACtB,oCACA,YACAC,KAAOp+B,KAAK49B,GAAYuC,GAA6B/B;;;;;;;;;WAYzDt+B,GAAiB6mB;QACf,MAAMyZ,IAAgBzZ,EAAY3K;QAClC,IAAIqkB,IAA2BrgC,KAAKsgC;QAEpC,OAAOtgC,KAAKq9B,YACTc,eAAe,sBAAsB,qBAAqBC;YACzD,MAAMmB,IAAiBv/B,KAAK09B,GAAgB8B,GAAgB;gBAC1DC,KAAe;;;wBAIjBY,IAA2BrgC,KAAKsgC;YAEhC,MAAM9L,IAAW;YACjB7N,EAAYjE,GAActc,QAAQ,CAACob,GAAQ3F;gBACzC,MAAM0kB,IAAgBF,EAAyBt9B,IAAI8Y;gBACnD,KAAK0kB,GACH;;;;gCAMF/L,EAASluB,KACPtG,KAAK49B,GACF4C,GAAmBpC,GAAK5c,EAAO0B,IAAkBrH,GACjDvT,KAAK,MACGtI,KAAK49B,GAAY6C,GACtBrC,GACA5c,EAAOwB,IACPnH;gBAKR,MAAMK,IAAcsF,EAAOtF;;gCAE3B,IAAIA,EAAYoI,OAAwB,GAAG;oBACzC,MAAMoc,IAAgBH,EACnBI,GAAgBzkB,GAAakkB,GAC7BQ,GAAmBxC,EAAIyC;oBAC1BR,IAA2BA,EAAyBvjB,GAClDjB,GACA6kB;;;oBAMAtD,GAAW0D,GACTP,GACAG,GACAlf,MAGFgT,EAASluB,KACPtG,KAAK49B,GAAYmD,GAAiB3C,GAAKsC;;;YAM/C,IAAIM,IAAc3gB,MACd4gB,IAAcrgB;;;;;YAiElB,IAhEA+F,EAAY/D,GAAgBxc,QAAQ,CAAC4C,GAAKgM;gBACxCisB,IAAcA,EAAYnhB,IAAI9W;;;;YAKhCwrB,EAASluB,KACPi5B,EAAejK,WAAW8I,GAAK6C,GAAa34B,KAAK22B;gBAC/CtY,EAAY/D,GAAgBxc,QAAQ,CAAC4C,GAAKgM;oBACxC,MAAMksB,IAAcjC,EAAal8B,IAAIiG;;;;;wCAOnCgM,aAAeY,MACfZ,EAAInC,QAAQrN,QAAQH,EAAgB2B;;;;oBAKpCu4B,EAAe4B,GAAYn4B,GAAKo3B,IAChCY,IAAcA,EAAYlkB,GAAO9T,GAAKgM,MAEvB,QAAfksB,KACAlsB,EAAInC,QAAQpF,EAAUyzB,EAAYruB,WAAW,KACG,MAA/CmC,EAAInC,QAAQpF,EAAUyzB,EAAYruB,YACjCquB,EAAY1pB,oBAMd+nB,EAAe6B,GAASpsB,GAAKorB;oBAC7BY,IAAcA,EAAYlkB,GAAO9T,GAAKgM,MAEtCqS,GAhgBA,cAkgBE,uCACAre,GACA,sBACAk4B,EAAYruB,SACZ,mBACAmC,EAAInC;oBAIJ8T,EAAY9D,GAAuBhD,IAAI7W,MACzCwrB,EAASluB,KACPtG,KAAKq9B,YAAYgE,GAAkBC,GACjClD,GACAp1B;;kBAYPo3B,EAAc56B,QAAQH,EAAgB2B,QAAQ;gBACjD,MAAMu6B,IAAsBvhC,KAAK49B,GAC9BuC,GAA6B/B,GAC7B91B,KAAKk5B,KAQGxhC,KAAK49B,GAAY6D,GACtBrD,GACAA,EAAIyC,IACJT;gBAGN5L,EAASluB,KAAKi7B;;YAGhB,OAAOnO,GAAmBsB,GAAQF,GAC/BlsB,KAAK,MAAMi3B,EAAeruB,MAAMktB,IAChC91B,KAAK,MACGtI,KAAK89B,GAAevI,GACzB6I,GACA4C;WAIPh+B,KAAKg+B,MACJhhC,KAAKsgC,KAAqBD,GACnBW;;;;;;;;;;;;WAeLlhC,UACNygC,GACAG,GACAlf;;QAQA,IANAne,GACEq9B,EAAcxkB,YAAYoI,OAAwB,IAKI,MAApDic,EAAcrkB,YAAYoI,MAC5B,QAAO;;;;;;gBAWT,OAFEoc,EAAc1kB,GAAgB0lB,MAC9BnB,EAAcvkB,GAAgB0lB,OACf1hC,KAAK2hC,MAUpBngB,EAAOwB,GAAexc,OACtBgb,EAAOyB,GAAkBzc,OACzBgb,EAAO0B,GAAiB1c,OACT;;;;;;;;;WAMnB1G,SAA6B8hC;QAC3B;kBACQ5hC,KAAKq9B,YAAYc,eACrB,0BACA,aACAC,KACShL,GAAmBhtB,QACxBw7B,GACCC,KACQzO,GAAmBhtB,QACxBy7B,EAAW5K,IACVjuB,KACChJ,KAAKq9B,YAAYgE,GAAkBS,GACjC1D,GACAyD,EAAWhmB,UACX7S,IAEJV,KAAK,MACL8qB,GAAmBhtB,QACjBy7B,EAAW3K,IACVluB,KACChJ,KAAKq9B,YAAYgE,GAAkBU,GACjC3D,GACAyD,EAAWhmB,UACX7S;UAQhB,OAAOynB;YACP,KAAIiJ,GAA4BjJ,IAO9B,MAAMA;;;;;YAFNpJ,GAzpBQ,cAypBU,wCAAwCoJ;;QAM9D,KAAK,MAAMoR,KAAcD,GAAa;YACpC,MAAM/lB,IAAWgmB,EAAWhmB;YAE5B,KAAKgmB,EAAW1f,WAAW;gBACzB,MAAM4D,IAAa/lB,KAAKsgC,GAAmBv9B,IAAI8Y,IAOzCI,IAA+B8J,EAAW/J,IAC1CgmB,IAAoBjc,EAAWkc,GACnChmB;;gCAEFjc,KAAKsgC,KAAqBtgC,KAAKsgC,GAAmBxjB,GAChDjB,GACAmmB;;;;;;;;;WAYRliC,GAAkBoiC;QAChB,OAAOliC,KAAKq9B,YAAYc,eACtB,2BACA,YACAC,WACuBv4B,MAAjBq8B,MACFA,KZtuBqB;QYwuBhBliC,KAAK60B,GAAcsN,GACxB/D,GACA8D;;;;;WAURpiC,GAAakJ;QACX,OAAOhJ,KAAKq9B,YAAYc,eAAe,iBAAiB,YAAYC,KAC3Dp+B,KAAK89B,GAAe9H,GAAYoI,GAAKp1B;;;;;;;;;WAYhDlJ,GAAe8b;QACb,OAAO5b,KAAKq9B,YACTc,eAAe,mBAAmB,aAAaC;YAC9C,IAAIrY;YACJ,OAAO/lB,KAAK49B,GACTwE,GAAchE,GAAKxiB,GACnBtT,KAAM+5B,KACDA;;;;YAIFtc,IAAasc,GACNjP,GAAmBlxB,QAAQ6jB,MAE3B/lB,KAAK49B,GAAY0E,GAAiBlE,GAAK91B,KAAKuT,MACjDkK,IAAa,IAAIpK,GACfC,GACAC,oBAEAuiB,EAAIyC;YAEC7gC,KAAK49B,GACT2E,GAAcnE,GAAKrY,GACnBzd,KAAK,MAAMyd;WAKvB/iB,KAAK+iB,MACqD,SAArD/lB,KAAKsgC,GAAmBv9B,IAAIgjB,EAAWlK,cACzC7b,KAAKsgC,KAAqBtgC,KAAKsgC,GAAmBxjB,GAChDiJ,EAAWlK,UACXkK;QAEF/lB,KAAKwiC,GAAiB7tB,IAAIiH,GAAQmK,EAAWlK,YAExCkK;;;;;;;IASbjmB,GACEi1B,GACAnZ;QAEA,MAAMC,IAAW7b,KAAKwiC,GAAiBz/B,IAAI6Y;QAC3C,YAAiB/V,MAAbgW,IACKuX,GAAmBlxB,QACxBlC,KAAKsgC,GAAmBv9B,IAAI8Y,MAGvB7b,KAAK49B,GAAYwE,GAAcrN,GAAanZ;;;;;;;;;;IAYvD9b,GACE+b,GACA4mB;QAEA,MAAM1c,IAAa/lB,KAAKsgC,GAAmBv9B,IAAI8Y,IAMzC6mB,IAAOD,IAA0B,cAAc;QACrD,OAAOziC,KAAKq9B,YACTc,eAAe,kBAAkBuE,GAAMtE,KACjCqE,IAMIrP,GAAmBlxB,YALnBlC,KAAKq9B,YAAYgE,GAAkB7b,aACxC4Y,OAOLp7B,KAAK;YACJhD,KAAKsgC,KAAqBtgC,KAAKsgC,GAAmBrjB,OAAOpB,IACzD7b,KAAKwiC,GAAiB5tB,OAAOmR,EAAYnK;;;;;;;;;;WAY/C9b,GACEgiB,GACA6gB;QAEA,IAAI1mB,IAA+B5W,EAAgB2B,OAC/C47B,IAAahiB;QAEjB,OAAO5gB,KAAKq9B,YAAYc,eAAe,iBAAiB,YAAYC,KAC3Dp+B,KAAKoiC,GAAchE,GAAKtc,EAAMzI,MAClC/Q,KAAKyd;YACJ,IAAIA,GAGF,OAFA9J,IACE8J,EAAW9J,8BACNjc,KAAK49B,GACTiF,GAA2BzE,GAAKrY,EAAWlK,UAC3CvT,KAAK6G;gBACJyzB,IAAazzB;;WAIpB7G,KAAK,MACJtI,KAAKs9B,GAAYhH,GACf8H,GACAtc,GACA6gB,IACI1mB,IACA5W,EAAgB2B,OACpB27B,IAAqBC,IAAahiB,OAGrCtY,KAAKga,MACG;YAAEA,WAAAA;YAAWwgB,IAAAF;;;IAKpB9iC,GACNs+B,GACA7L,GACAgN;QAEA,MAAMzM,IAAQP,EAAYO,OACpBiQ,IAAUjQ,EAAMvkB;QACtB,IAAIy0B,IAAe5P,GAAmBlxB;QAiCtC,OAhCA6gC,EAAQ38B,QAAQksB;YACd0Q,IAAeA,EACZ16B,KAAK,MACGi3B,EAAenK,GAASgJ,GAAK9L,IAErChqB,KAAM26B;gBACL,IAAIjuB,IAAMiuB;gBACV,MAAMC,IAAa3Q,EAAYU,GAAYlwB,IAAIuvB;gBAl5B9BjvB,GAo5BA,SAAf6/B,MAGGluB,KAAOA,EAAInC,QAAQpF,OAAyB,OAC/CuH,IAAM8d,EAAMrd,GAAsB6c,GAAQtd,GAAKud,IAC1Cvd;;;;gBAaHuqB,EAAe6B,GAASpsB,GAAKud,EAAYQ;;YAK5CiQ,EAAa16B,KAAK,MACvBtI,KAAK60B,GAAckL,GAAoB3B,GAAKtL;;IAIhDhzB,GAAeqjC;QACb,OAAOnjC,KAAKq9B,YAAYc,eACtB,mBACA,qBACAC,KAAO+E,EAAiBC,GAAQhF,GAAKp+B,KAAKsgC;;;;;;;;;;;;;;;;;;;;;AAgIzC5E,eAAe2H,GACpBlP;IAEA,IACEA,EAAI5yB,SAASpB,EAAKW,uBCvkCpB,gIDwkCEqzB,EAAI3yB,SAIJ,MAAM2yB;IAFN9M,GAjiCY,cAiiCM;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAj9BpB+V,SAAsD;;MEhH3CkG;IAAbxjC;;QAEEE,UAAoB,IAAIkf,GAAUqkB,GAAaC;;QAG/CxjC,UAAuB,IAAIkf,GAAUqkB,GAAaE;;wEAGlD3jC;QACE,OAAOE,KAAK0jC,GAAUr6B;;2EAIxBvJ,GAAakJ,GAAkB8oB;QAC7B,MAAM6R,IAAM,IAAIJ,GAAav6B,GAAK8oB;QAClC9xB,KAAK0jC,KAAY1jC,KAAK0jC,GAAU5jB,IAAI6jB,IACpC3jC,KAAK4jC,KAAe5jC,KAAK4jC,GAAa9jB,IAAI6jB;;0EAI5C7jC,GAAcyO,GAAsBujB;QAClCvjB,EAAKnI,QAAQ4C,KAAOhJ,KAAK8hC,GAAa94B,GAAK8oB;;;;;WAO7ChyB,GAAgBkJ,GAAkB8oB;QAChC9xB,KAAK6jC,GAAU,IAAIN,GAAav6B,GAAK8oB;;IAGvChyB,GAAiByO,GAAsBujB;QACrCvjB,EAAKnI,QAAQ4C,KAAOhJ,KAAK+hC,GAAgB/4B,GAAK8oB;;;;;WAOhDhyB,GAAsBgyB;QACpB,MAAMgS,IAAWv7B,EAAYsU,OACvBknB,IAAW,IAAIR,GAAaO,GAAUhS,IACtCkS,IAAS,IAAIT,GAAaO,GAAUhS,IAAK,IACzCvjB,IAAsB;QAK5B,OAJAvO,KAAK4jC,GAAaK,GAAe,EAACF,GAAUC,KAASL;YACnD3jC,KAAK6jC,GAAUF,IACfp1B,EAAKjI,KAAKq9B,EAAI36B;YAETuF;;IAGTzO;QACEE,KAAK0jC,GAAUt9B,QAAQu9B,KAAO3jC,KAAK6jC,GAAUF;;IAGvC7jC,GAAU6jC;QAChB3jC,KAAK0jC,KAAY1jC,KAAK0jC,GAAU9uB,OAAO+uB,IACvC3jC,KAAK4jC,KAAe5jC,KAAK4jC,GAAahvB,OAAO+uB;;IAG/C7jC,GAAgBgyB;QACd,MAAMgS,IAAWv7B,EAAYsU,OACvBknB,IAAW,IAAIR,GAAaO,GAAUhS,IACtCkS,IAAS,IAAIT,GAAaO,GAAUhS,IAAK;QAC/C,IAAIvjB,IAAOqS;QAIX,OAHA5gB,KAAK4jC,GAAaK,GAAe,EAACF,GAAUC,KAASL;YACnDp1B,IAAOA,EAAKuR,IAAI6jB,EAAI36B;YAEfuF;;IAGTzO,GAAYkJ;QACV,MAAM26B,IAAM,IAAIJ,GAAav6B,GAAK,IAC5Bk7B,IAAWlkC,KAAK0jC,GAAUS,GAAkBR;QAClD,OAAoB,SAAbO,KAAqBl7B,EAAIxD,QAAQ0+B,EAASl7B;;;;MAIxCu6B;IACXzjC,YACSkJ,GACAo7B;QADApkC,WAAAgJ,aACAo7B;;wCAITtkC,UAAoBmH,GAAoBC;QACtC,OACEqB,EAAYvC,EAAWiB,EAAK+B,KAAK9B,EAAM8B,QACvC/D,GAAoBgC,EAAKm9B,IAAiBl9B,EAAMk9B;;wCAKpDtkC,UAAyBmH,GAAoBC;QAC3C,OACEjC,GAAoBgC,EAAKm9B,IAAiBl9B,EAAMk9B,OAChD77B,EAAYvC,EAAWiB,EAAK+B,KAAK9B,EAAM8B;;;;;;;;;;;;;;;;;;;;;;;;;;aClG7Bq7B,GAAeC,GAAsBnU;IACnD,IAAoB,MAAhBA,EAAKvqB,QACP,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,YAAY+jC,qCACV,yBACAC,GAAapU,EAAKvqB,QAAQ,cAC1B;;;;;;;;;aAYQ4+B,GACdF,GACAnU,GACAsU;IAEA,IAAItU,EAAKvqB,WAAW6+B,GAClB,MAAM,IAAIpjC,EACRlB,EAAKI,kBACL,YAAY+jC,kBACVC,GAAaE,GAAc,cAC3B,2BACAF,GAAapU,EAAKvqB,QAAQ,cAC1B;;;;;;;;;;aAaQ8+B,GACdJ,GACAnU,GACAwU;IAEA,IAAIxU,EAAKvqB,SAAS++B,GAChB,MAAM,IAAItjC,EACRlB,EAAKI,kBACL,YAAY+jC,2BACVC,GAAaI,GAAiB,cAC9B,2BACAJ,GAAapU,EAAKvqB,QAAQ,cAC1B;;;;;;;;;;aAaQg/B,GACdN,GACAnU,GACAwU,GACAE;IAEA,IAAI1U,EAAKvqB,SAAS++B,KAAmBxU,EAAKvqB,SAASi/B,GACjD,MAAM,IAAIxjC,EACRlB,EAAKI,kBACL,YAAY+jC,wBAAmCK,WAC7C,GAAGE,sCACHN,GAAapU,EAAKvqB,QAAQ,cAC1B;;;;;;;;;;;SA6BQk/B,GACdR,GACA3iB,GACArH,GACAyqB;IAEAC,GAAaV,GAAc3iB,GAAM,GAAGsjB,GAAQ3qB,eAAsByqB;;;;;;aAOpDG,GACdZ,GACA3iB,GACArH,GACAyqB;SAEiBl/B,MAAbk/B,KACFD,GAAgBR,GAAc3iB,GAAMrH,GAAUyqB;;;;;;aAQlCI,GACdb,GACA3iB,GACAyjB,GACAL;IAEAC,GAAaV,GAAc3iB,GAAM,GAAGyjB,YAAqBL;;;;;;aAO3CM,GACdf,GACA3iB,GACAyjB,GACAL;SAEiBl/B,MAAbk/B,KACFI,GAAkBb,GAAc3iB,GAAMyjB,GAAYL;;;SA+BtCO,GACdhB,GACAc,GACAG,GACAR,GACAS;SAEiB3/B,MAAbk/B,cAjCJT,GACAc,GACAG,GACAR,GACAS;QAEA,MAAMT,aAAoBU,QACxB,MAAM,IAAIpkC,EACRlB,EAAKI,kBACL,YAAY+jC,oBAA+Bc,OACzC,sCAAsCM,GAAiBX;QAI7D,KAAK,IAAIr+B,IAAI,GAAGA,IAAIq+B,EAASn/B,UAAUc,GACrC,KAAK8+B,EAAUT,EAASr+B,KACtB,MAAM,IAAIrF,EACRlB,EAAKI,kBACL,YAAY+jC,oBAA+Bc,OACzC,kBAAkBG,6BAA2C7+B,OAC7D,QAAQg/B,GAAiBX,EAASr+B;KAcxCi/B,CACErB,GACAc,GACAG,GACAR,GACAS;;;;;;;;;;SAoCUI,GACdtB,GACAuB,GACAT,GACAU,GACAC;SAEclgC,MAAVigC,cAlCJxB,GACAuB,GACAT,GACAU,GACAC;QAEA,MAAMC,IAAgC;QAEtC,KAAK,MAAM7qB,KAAO4qB,GAAU;YAC1B,IAAI5qB,MAAQ2qB,GACV;YAEFE,EAAoB1/B,KAAKo/B,GAAiBvqB;;QAG5C,MAAM8qB,IAAoBP,GAAiBI;QAC3C,MAAM,IAAIzkC,EACRlB,EAAKI,kBACL,iBAAiB0lC,0BAA0C3B,oBACzD,IAAIc,0BAAmCY,EAAoB3+B,KAAK;KAgBlE6+B,CACE5B,GACAuB,GACAT,GACAU,GACAC;;;;;;;;;;;;;AAgCN,SAASf,GACPV,GACA3iB,GACAkkB,GACAC;IAEA,IAAIK,KAAQ;IASZ,IAPEA,IADW,aAATxkB,IACMykB,GAAcN,KACJ,uBAATnkB,IACgB,mBAAVmkB,KAAgC,OAAVA,WAEtBA,MAAUnkB;KAGtBwkB,GAAO;QACV,MAAME,IAAcX,GAAiBI;QACrC,MAAM,IAAIzkC,EACRlB,EAAKI,kBACL,YAAY+jC,oBAA+BuB,OACzC,iBAAiBlkB,kBAAqB0kB;;;;;;;aAS9BD,GAAcN;IAC5B,OACmB,mBAAVA,KACG,SAAVA,MACC78B,OAAOq9B,eAAeR,OAAW78B,OAAOC,aACN,SAAjCD,OAAOq9B,eAAeR;;;oFAKZJ,GAAiBI;IAC/B,SAAcjgC,MAAVigC,GACF,OAAO;IACF,IAAc,SAAVA,GACT,OAAO;IACF,IAAqB,mBAAVA,GAIhB,OAHIA,EAAMlgC,SAAS,OACjBkgC,IAAQ,GAAGA,EAAMS,UAAU,GAAG;IAEzBC,KAAKC,UAAUX;IACjB,IAAqB,mBAAVA,KAAuC,oBAAVA,GAC7C,OAAO,KAAKA;IACP,IAAqB,mBAAVA,GAAoB;QACpC,IAAIA,aAAiBL,OACnB,OAAO;QACF;YACL,MAAMiB;;qBAe2BZ;gBACrC,IAAIA,EAAMjV,aAAa;oBACrB,MACMqC,IADgB,4BACQzjB,KAAKq2B,EAAMjV,YAAYnvB;oBACrD,IAAIwxB,KAAWA,EAAQttB,SAAS,GAC9B,OAAOstB,EAAQ;;gBAGnB,OAAO;;8DAvBsByT;YACzB,OAAID,IACK,YAAYA,aAEZ;;;IAGN,OAAqB,qBAAVZ,IACT,eA1VXhgC;;;SA6WgB8gC,GACdtC,GACAhqB,GACAyqB;IAEA,SAAiBl/B,MAAbk/B,GACF,MAAM,IAAI1jC,EACRlB,EAAKI,kBACL,YAAY+jC,wBAAmCW,GAAQ3qB,QACrD;;;;;;aASQusB,GACdvC,GACAhtB,GACAwvB;IAEA1gC,EAAQkR,GAA0B,CAACtO,GAAK4c;QACtC,IAAIkhB,EAAYt/B,QAAQwB,KAAO,GAC7B,MAAM,IAAI3H,EACRlB,EAAKI,kBACL,mBAAmByI,yBAA2Bs7B,UAC5C,wBACAwC,EAAYz/B,KAAK;;;;;;;aAUX0/B,GACdzC,GACA3iB,GACArH,GACAyqB;IAEA,MAAMsB,IAAcX,GAAiBX;IACrC,OAAO,IAAI1jC,EACTlB,EAAKI,kBACL,YAAY+jC,oBAA+BW,GAAQ3qB,QACjD,oBAAoBqH,kBAAqB0kB;;;SAI/BW,GACd1C,GACAhqB,GACA6D;IAEA,IAAIA,KAAK,GACP,MAAM,IAAI9c,EACRlB,EAAKI,kBACL,YAAY+jC,oBAA+BW,GACzC3qB,oDACiD6D;;;2DAMzD,UAAS8mB,GAAQgC;IACf,QAAQA;MACN,KAAK;QACH,OAAO;;MACT,KAAK;QACH,OAAO;;MACT,KAAK;QACH,OAAO;;MACT;QACE,OAAOA,IAAM;;;;;;GAOnB,UAAS1C,GAAa0C,GAAal/B;IACjC,OAAO,GAAGk/B,KAAOl/B,OAAiB,MAARk/B,IAAY,KAAK;;;;;;;;;;;;;;;;;;;oEC7c7C,UAASC;IACP,IAA0B,sBAAfj9B,YACT,MAAM,IAAI5I,EACRlB,EAAKc,eACL;;;2EAMN,UAASkmC;IACP,KAAK19B,GAAgBC,KAAc09B,IACjC,MAAM,IAAI/lC,EACRlB,EAAKc,eACL;;;;;;;;;UAYOomC;IAKXvnC,YAAYwnC;QACVH,MACAnnC,KAAKunC,KAAcD;;IAGrBxnC,wBAAwB0J;QACtBg7B,GAA0B,yBAAyBgD,WAAW,IAC9D1C,GAAgB,yBAAyB,UAAU,GAAGt7B;QACtD29B;QACA;YACE,OAAO,IAAIE,GAAK/9B,EAAWyG,iBAAiBvG;UAC5C,OAAOinB;YACP,MAAM,IAAIpvB,EACRlB,EAAKI,kBACL,kDAAkDkwB;;;IAKxD3wB,sBAAsB8J;QAGpB,IAFA46B,GAA0B,uBAAuBgD,WAAW,IAC5DN,QACMt9B,aAAiBK,aACrB,MAAM88B,GAAkB,uBAAuB,cAAc,GAAGn9B;QAElE,OAAO,IAAIy9B,GAAK/9B,EAAW0G,eAAepG;;IAG5C9J;QAGE,OAFA0kC,GAA0B,iBAAiBgD,WAAW,IACtDL,MACOnnC,KAAKunC,GAAYv4B;;IAG1BlP;QAGE,OAFA0kC,GAA0B,qBAAqBgD,WAAW,IAC1DN,MACOlnC,KAAKunC,GAAYze;;IAG1BhpB;QACE,OAAO,kBAAkBE,KAAKgP,aAAa;;IAG7ClP,QAAQkF;QACN,OAAOhF,KAAKunC,GAAY/hC,QAAQR,EAAMuiC;;;;;;;;;;;;;;;;;;;;;;;;;;;UCnE7B3/B;;;;;;;IAUX9H,eAAe2nC;kBFkFfnD,GACAziC,GACAF,GACA+lC;YAEA,MAAM7lC,aAAiB4jC,UAAU5jC,EAAM+D,SAAS8hC,GAC9C,MAAM,IAAIrmC,EACRlB,EAAKI,kBACL,YAAY+jC,oBAA+B3iC,yBACzC,yBACA,GAAG4iC,GAAamD,GAAqB;SE3FzCC,CACE,aACAF,GACA,cACA;QAGF,KAAK,IAAI/gC,IAAI,GAAGA,IAAI+gC,EAAW7hC,UAAUc,GAEvC,IADAo+B,GAAgB,aAAa,UAAUp+B,GAAG+gC,EAAW/gC,KACxB,MAAzB+gC,EAAW/gC,GAAGd,QAChB,MAAM,IAAIvE,EACRlB,EAAKI,kBACL;QAMNP,KAAK4nC,KAAgB,IAAIC,EAAkBJ;;IAa7C3nC;QACE,OAAO8H,GAAUkgC;;IAGnBhoC,QAAQkF;QACN,MAAMA,aAAiB4C,KACrB,MAAMm/B,GAAkB,WAAW,aAAa,GAAG/hC;QAErD,OAAOhF,KAAK4nC,GAAcpiC,QAAQR,EAAM4iC;;;;;;;;;GAZ1ChgC,SAAuC,IAAIA,GACzCigC,EAAkBhvB,IAAWvR;;;;;AAkBjC,MAAMygC,KAAW,IAAIh9B,OAAO;;;;;;;;;;;;;;;;;;;;;;;;;MCvDNi9B;IACpBloC,YAA+BmoC;kBAAAA;;;;MAOpBC,WAA6BF;IACxCloC;QACE2B,MAAM;;IAGR3B,GAAiBqoC;QACf,yBAAIA,EAAQC,IAIL,yBAAID,EAAQC,KAMXD,EAAQE,GACZ,8EAKIF,EAAQE,GACZ;;;QAIJ,OAlBEF,EAAQl0B,GAAU3N,KAAK6hC,EAAa,OAkB/B;;IAGTroC,QAAQkF;QACN,OAAOA,aAAiBkjC;;;;MAIfI,WAAsCN;IACjDloC;QACE2B,MAAM;;IAGR3B,GAAiBqoC;QACf,OAAO,IAAI11B,GAAe01B,EAAa,MAAEz3B,GAAyBwb;;IAGpEpsB,QAAQkF;QACN,OAAOA,aAAiBsjC;;;;MAIfC,WAAiCP;IAC5CloC,YAA6B0oC;QAC3B/mC,MAAM,oCADqB+mC;;IAI7B1oC,GAAiBqoC;;;;QAIf,MAAMM,IAAe,IAAIC,GACvB;YACEC;YACAC,YAAY5oC,KAAKioC;YACjBY,KAAc;WAEhBV,EAAQj4B,IACRi4B,EAAQz2B,YACRy2B,EAAQW,4BAEJC,IAAiB/oC,KAAKwoC,GAAU1gC,IACpCwJ,KAAW03B,GAAU13B,GAASm3B,KAE1BQ,IAAa,IAAIj4B,GAA6B+3B;QACpD,OAAO,IAAIt2B,GAAe01B,EAAQ5gC,MAAO0hC;;IAG3CnpC,QAAQkF;;QAEN,OAAOhF,SAASgF;;;;MAIPkkC,WAAkClB;IAC7CloC,YAAqB0oC;QACnB/mC,MAAM,qCADa+mC;;IAIrB1oC,GAAiBqoC;;;;QAIf,MAAMM,IAAe,IAAIC,GACvB;YACEC;YACAC,YAAY5oC,KAAKioC;YACjBY,KAAc;WAEhBV,EAAQj4B,IACRi4B,EAAQz2B,YACRy2B,EAAQW,4BAEJC,IAAiB/oC,KAAKwoC,GAAU1gC,IACpCwJ,KAAW03B,GAAU13B,GAASm3B,KAE1BQ,IAAa,IAAI13B,GAA8Bw3B;QACrD,OAAO,IAAIt2B,GAAe01B,EAAQ5gC,MAAO0hC;;IAG3CnpC,QAAQkF;;QAEN,OAAOhF,SAASgF;;;;MAIPmkC,WAAuCnB;IAClDloC,YAA6BspC;QAC3B3nC,MAAM,mCADqB2nC;;IAI7BtpC,GAAiBqoC;QACf,MAAMM,IAAe,IAAIC,GACvB;YACEC;YACAC,YAAY5oC,KAAKioC;WAEnBE,EAAQj4B,IACRi4B,EAAQz2B,YACRy2B,EAAQW,4BAEJn3B,IAAUq3B,GAAUhpC,KAAKopC,IAAUX,IACnCY,IAAmB,IAAI53B,GAC3B02B,EAAQz2B,YACRC;QAEF,OAAO,IAAIc,GAAe01B,EAAa,MAAEkB;;IAG3CvpC,QAAQkF;;QAEN,OAAOhF,SAASgF;;;;;;;;;;;;;;;;;;;;;;;;MC5JPskC;IAMXxpC,YAAY+L,GAAkBC;QAI5B,IAHA04B,GAA0B,YAAYgD,WAAW,IACjD1C,GAAgB,YAAY,UAAU,GAAGj5B,IACzCi5B,GAAgB,YAAY,UAAU,GAAGh5B;SACpCy9B,SAAS19B,MAAaA,KAAY,MAAMA,IAAW,IACtD,MAAM,IAAIxK,EACRlB,EAAKI,kBACL,4DAA4DsL;QAGhE,KAAK09B,SAASz9B,MAAcA,KAAa,OAAOA,IAAY,KAC1D,MAAM,IAAIzK,EACRlB,EAAKI,kBACL,+DAA+DuL;QAInE9L,KAAKwpC,KAAO39B,GACZ7L,KAAKypC,KAAQ39B;;;;WAMfD;QACE,OAAO7L,KAAKwpC;;;;WAMd19B;QACE,OAAO9L,KAAKypC;;IAGd3pC,QAAQkF;QACN,OAAOhF,KAAKwpC,OAASxkC,EAAMwkC,MAAQxpC,KAAKypC,OAAUzkC,EAAMykC;;;;;WAO1D3pC,EAAWkF;QACT,OACEC,GAAoBjF,KAAKwpC,IAAMxkC,EAAMwkC,OACrCvkC,GAAoBjF,KAAKypC,IAAOzkC,EAAMykC;;;;;;;;;;;;;;;;;;;GC9B5C,OAAMC,KAAuB;;4EAGhBC;IACX7pC,YACWkU,GACAC,GACAc;QAFA/U,YAAAgU,aACAC,GACAjU,uBAAA+U;;IAGXjV,GAAYkJ,GAAkBuK;QAC5B,MAAM8e,IAAY;QAWlB,OAVuB,SAAnBryB,KAAKiU,KACPoe,EAAU/rB,KACR,IAAIyN,GAAc/K,GAAKhJ,KAAKgU,MAAMhU,KAAKiU,IAAWV,MAGpD8e,EAAU/rB,KAAK,IAAIgN,GAAYtK,GAAKhJ,KAAKgU,MAAMT;QAE7CvT,KAAK+U,gBAAgBnP,SAAS,KAChCysB,EAAU/rB,KAAK,IAAIwO,GAAkB9L,GAAKhJ,KAAK+U,mBAE1Csd;;;;gFAKEuX;IACX9pC,YACWkU,GACAC,GACAc;QAFA/U,YAAAgU,aACAC,GACAjU,uBAAA+U;;IAGXjV,GAAYkJ,GAAkBuK;QAC5B,MAAM8e,IAAY,EAChB,IAAIte,GAAc/K,GAAKhJ,KAAKgU,MAAMhU,KAAKiU,IAAWV;QAKpD,OAHIvT,KAAK+U,gBAAgBnP,SAAS,KAChCysB,EAAU/rB,KAAK,IAAIwO,GAAkB9L,GAAKhJ,KAAK+U;QAE1Csd;;;;AAyBX,SAASwX,GAAQzB;IACf,QAAQA;MACN;;cACA;;cACA;QACE,QAAO;;MACT;MACA;QACE,QAAO;;MACT;QACE,MAjGCtiC;;;;uEA0HM4iC;;;;;;;;;;;;;;;;;;;IAqBX5oC,YACWgqC,GACA55B,GACAwB,GACAo3B,GACT/zB,GACAd;QALSjU,gBAAA8pC,aACA55B,GACAlQ,kBAAA0R,GACA1R,iCAAA8oC;;;aAMejjC,MAApBkP,KACF/U,KAAK+pC,MAEP/pC,KAAK+U,kBAAkBA,KAAmB,IAC1C/U,KAAKiU,KAAYA,KAAa;;IAGhC1M;QACE,OAAOvH,KAAK8pC,SAASviC;;IAGvBohC;QACE,OAAO3oC,KAAK8pC,SAAS1B;;6EAIvBtoC,GAAYkqC;QACV,OAAO,IAAItB,mCACJ1oC,KAAK8pC,WAAaE,IACvBhqC,KAAKkQ,IACLlQ,KAAK0R,YACL1R,KAAK8oC,2BACL9oC,KAAK+U,iBACL/U,KAAKiU;;IAITnU,GAAqB4S;;QACnB,MAAMu3B,kBAAYjqC,KAAKuH,mCAAMyP,MAAMtE,IAC7By1B,IAAUnoC,KAAKkqC,GAAY;YAAE3iC,MAAM0iC;YAAWpB,KAAc;;QAElE,OADAV,EAAQgC,GAAoBz3B,IACrBy1B;;IAGTroC,GAAyB4S;;QACvB,MAAMu3B,kBAAYjqC,KAAKuH,mCAAMyP,MAAMtE,IAC7By1B,IAAUnoC,KAAKkqC,GAAY;YAAE3iC,MAAM0iC;YAAWpB,KAAc;;QAElE,OADAV,EAAQ4B,MACD5B;;IAGTroC,GAAqB2G;;;QAGnB,OAAOzG,KAAKkqC,GAAY;YAAE3iC,WAAM1B;YAAWgjC,KAAc;;;IAG3D/oC,GAAY06B;QACV,MAAM4P,KACHpqC,KAAKuH,QAAQvH,KAAKuH,KAAK8B,MACpB,KACA,oBAAoBrJ,KAAKuH,KAAK7F;QACpC,OAAO,IAAIL,EACTlB,EAAKI,kBACL,YAAYP,KAAK8pC,SAASlB,4CACxBpO,IACA4P;;sFAKNtqC,SAASsS;QACP,YACgEvM,MAA9D7F,KAAKiU,GAAUnH,KAAK4F,KAASN,EAAUE,EAAWI,YAG5C7M,MAFN7F,KAAK+U,gBAAgBjI,KAAK6F,KACxBP,EAAUE,EAAWK,EAAUD;;IAK7B5S;;;QAGN,IAAKE,KAAKuH,MAGV,KAAK,IAAIb,IAAI,GAAGA,IAAI1G,KAAKuH,KAAK3B,QAAQc,KACpC1G,KAAKmqC,GAAoBnqC,KAAKuH,KAAKxE,IAAI2D;;IAInC5G,GAAoBuG;QAC1B,IAAuB,MAAnBA,EAAQT,QACV,MAAM5F,KAAKqoC,GAAY;QAEzB,IAAIwB,GAAQ7pC,KAAKooC,OAAesB,GAAqB7hC,KAAKxB,IACxD,MAAMrG,KAAKqoC,GAAY;;;;;;;UAShBgC;IAGXvqC,YACmBoQ,GACA44B,GACjBp3B;kBAFiBxB,GACAlQ,iCAAA8oC,GAGjB9oC,KAAK0R,aACHA,KAAcjI,GAAgBC,KAAc4gC,GAAcp6B;;+DAI9DpQ,GAAa8oC,GAAoB9C;QAC/B,MAAMqC,IAAUnoC,KAAKuqC,iBAAkC3B;QACvD4B,GAAoB,uCAAuCrC,GAASrC;QACpE,MAAM2E,IAAaC,GAAY5E,GAAOqC;QAEtC,OAAO,IAAIwB,GACT,IAAIt1B,GAAYo2B;yBACC,MACjBtC,EAAQpzB;;yEAKZjV,GACE8oC,GACA9C,GACAvW;QAEA,MAAM4Y,IAAUnoC,KAAKuqC,sBAAuC3B;QAC5D4B,GAAoB,uCAAuCrC,GAASrC;QACpE,MAAM2E,IAAaC,GAAY5E,GAAOqC;QAEtC,IAAIl0B,GACAc;QAEJ,IAAKwa,GAGE;YACL,MAAMob,IAAmC;YAEzC,KAAK,MAAMC,KAAqBrb,GAAY;gBAC1C,IAAInd;gBAEJ,IAAIw4B,aAA6BC,IAC/Bz4B,IAAYw4B,EAAkBhD,SACzB;oBAAA,IAAiC,mBAAtBgD,GAMhB,MA/SH9kC;oBA0SGsM,IAAY04B,GACVlC,GACAgC;;gBAQJ,KAAKzC,EAAQ4C,SAAS34B,IACpB,MAAM,IAAI/Q,EACRlB,EAAKI,kBACL,UAAU6R;gBAIT44B,GAAkBL,GAAqBv4B,MAC1Cu4B,EAAoBrkC,KAAK8L;;YAI7B6B,IAAY,IAAI9B,GAAUw4B,IAC1B51B,IAAkBozB,EAAQpzB,gBAAgBrN,OAAOiL,KAC/CsB,EAAUg3B,GAAOt4B,EAAUD;eAnC7BuB,IAAY,IAAI9B,GAAUg2B,EAAQl0B,KAClCc,IAAkBozB,EAAQpzB;QAqC5B,OAAO,IAAI40B,GACT,IAAIt1B,GAAYo2B,IAChBx2B,GACAc;;uDAKJjV,GAAgB8oC,GAAoB9C;QAClC,MAAMqC,IAAUnoC,KAAKuqC,oBAAqC3B;QAC1D4B,GAAoB,uCAAuCrC,GAASrC;QAEpE,MAAMoF,IAA8B,IAC9BT,IAAa,IAAIh2B;QACvBrO,EAAQ0/B,GAAwB,CAAC98B,GAAKnH;YACpC,MAAM0F,IAAOujC,GAAgClC,GAAY5/B,IAEnDmiC,IAAehD,EAAQiD,GAAyB7jC;YACtD,IAAI1F,aAAiBqmC;;YAEnBgD,EAAe5kC,KAAKiB,SACf;gBACL,MAAM8jC,IAAcrC,GAAUnnC,GAAOspC;gBAClB,QAAfE,MACFH,EAAe5kC,KAAKiB,IACpBkjC,EAAW91B,IAAIpN,GAAM8jC;;;QAK3B,MAAMC,IAAO,IAAIn5B,GAAU+4B;QAC3B,OAAO,IAAItB,GACTa,EAAW51B,MACXy2B,GACAnD,EAAQpzB;;sEAKZjV,GACE8oC,GACAl2B,GACA7Q,GACA0pC;QAEA,MAAMpD,IAAUnoC,KAAKuqC,oBAAqC3B,IACpDr6B,IAAO,EAACi9B,GAAsB5C,GAAYl2B,MAC1CnG,IAAS,EAAC1K;QAEhB,IAAI0pC,EAAoB3lC,SAAS,KAAM,GACrC,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,YAAYqoC,gDACV;QAIN,KAAK,IAAIliC,IAAI,GAAGA,IAAI6kC,EAAoB3lC,QAAQc,KAAK,GACnD6H,EAAKjI,KACHklC,GACE5C,GACA2C,EAAoB7kC,MAGxB6F,EAAOjG,KAAKilC,EAAoB7kC,IAAI;QAGtC,MAAMwkC,IAA8B,IAC9BT,IAAa,IAAIh2B;;;QAIvB,KAAK,IAAI/N,IAAI6H,EAAK3I,SAAS,GAAGc,KAAK,KAAKA,GACtC,KAAKskC,GAAkBE,GAAgB38B,EAAK7H,KAAK;YAC/C,MAAMa,IAAOgH,EAAK7H,IACZ7E,IAAQ0K,EAAO7F,IACfykC,IAAehD,EAAQiD,GAAyB7jC;YACtD,IAAI1F,aAAiBqmC;;YAEnBgD,EAAe5kC,KAAKiB,SACf;gBACL,MAAM8jC,IAAcrC,GAAUnnC,GAAOspC;gBAClB,QAAfE,MACFH,EAAe5kC,KAAKiB,IACpBkjC,EAAW91B,IAAIpN,GAAM8jC;;;QAM7B,MAAMC,IAAO,IAAIn5B,GAAU+4B;QAC3B,OAAO,IAAItB,GACTa,EAAW51B,MACXy2B,GACAnD,EAAQpzB;;qDAKJjV,GACNsoC,GACAQ;QAEA,OAAO,IAAIF,GACT;YACEC,IAAAP;YACAQ,YAAAA;YACArhC,MAAMK,EAAU6O;YAChBoyB,KAAc;WAEhB7oC,KAAKkQ,IACLlQ,KAAK0R,YACL1R,KAAK8oC;;;;;;;;WAWThpC,GACE8oC,GACA9C,GACA2F,KAAc;QAYd,OANezC,GAAUlD,GAJT9lC,KAAKuqC,GACnBkB,+CACA7C;;;;;;;;;;;;aAqBUI,GACdlD,GACAqC;IAEA,IAAIuD,GAAoB5F,IAEtB,OADA0E,GAAoB,4BAA4BrC,GAASrC,IAClD4E,GAAY5E,GAAOqC;IACrB,IAAIrC,aAAiBkC;;;;;;;;;;IAO1B,OA2EJ,SACEnmC,GACAsmC;;QAGA,KAAK0B,GAAQ1B,EAAQC,KACnB,MAAMD,EAAQE,GACZ,GAAGxmC,EAAMomC;QAGb,IAAqB,SAAjBE,EAAQ5gC,MACV,MAAM4gC,EAAQE,GACZ,GAAGxmC,EAAMomC;QAIb,MAAM3yB,IAAiBzT,EAAMopB,GAAiBkd;QAC1C7yB,KACF6yB,EAAQpzB,gBAAgBzO,KAAKgP;;;;;;GA9F7Bq2B,EAAwB7F,GAAOqC,IACxB;IAQP;;;IAJIA,EAAQ5gC,QACV4gC,EAAQl0B,GAAU3N,KAAK6hC,EAAQ5gC,OAG7Bu+B,aAAiBL,OAAO;;;;;;;QAO1B,IACE0C,EAAQ2B,SAAS8B,gCACjBzD,EAAQC,IAER,MAAMD,EAAQE,GAAY;QAE5B,OA+BN,SAAoBz+B,GAAkBu+B;YACpC,MAAM57B,IAAsB;YAC5B,IAAIs/B,IAAa;YACjB,KAAK,MAAMC,KAASliC,GAAO;gBACzB,IAAImiC,IAAc/C,GAChB8C,GACA3D,EAAQ6D,GAAqBH;gBAEZ,QAAfE;;;gBAGFA,IAAc;oBAAE1c,WAAW;oBAE7B9iB,EAAOjG,KAAKylC,IACZF;;YAEF,OAAO;gBAAEv/B,YAAY;oBAAEC,QAAAA;;;SA/CZ0/B,CAAWnG,GAAoBqC;;IAEtC,OA+EN,SACEtmC,GACAsmC;QAEA,IAAc,SAAVtmC,GACF,OAAO;YAAEwtB,WAAW;;QACf,IAAqB,mBAAVxtB,GAChB,OAAOsmC,EAAQz2B,WAAWw6B,GAASrqC;QAC9B,IAAqB,oBAAVA,GAChB,OAAO;YAAEsJ,cAActJ;;QAClB,IAAqB,mBAAVA,GAChB,OAAO;YAAE4I,aAAa5I;;QACjB,IAAIA,aAAiB2C,MAAM;YAChC,MAAMc,IAAYlB,EAAU+nC,SAAStqC;YACrC,OAAO;gBAAEuJ,gBAAgB+8B,EAAQz2B,WAAWqX,EAAYzjB;;;QACnD,IAAIzD,aAAiBuC,GAAW;;;;YAIrC,MAAMkB,IAAY,IAAIlB,EACpBvC,EAAMwC,SACiC,MAAvCQ,KAAKC,MAAMjD,EAAMyC,cAAc;YAEjC,OAAO;gBAAE8G,gBAAgB+8B,EAAQz2B,WAAWqX,EAAYzjB;;;QACnD,IAAIzD,aAAiBynC,IAC1B,OAAO;YACL19B,eAAe;gBACbC,UAAUhK,EAAMgK;gBAChBC,WAAWjK,EAAMiK;;;QAGhB,IAAIjK,aAAiBwlC,IAC1B,OAAO;YAAE+E,YAAYjE,EAAQz2B,WAAWkc,GAAQ/rB;;QAC3C,IAAIA,aAAiBwqC,IAAmB;YAC7C,MAAMC,IAASnE,EAAQj4B,IACjBq8B,IAAU1qC,EAAM2qC,UAAUC;YAChC,KAAKF,EAAQ/mC,QAAQ8mC,IACnB,MAAMnE,EAAQE,GACZ,wCACE,GAAGkE,EAAQp8B,aAAao8B,EAAQn8B,4BAChC,gBAAgBk8B,EAAOn8B,aAAam8B,EAAOl8B;YAGjD,OAAO;gBACL1E,gBAAgBy8B,EAAQz2B,WAAW0X,GACjCvnB,EAAM6qC,GAAKnlC,MACX1F,EAAM2qC,UAAUC;;;QAGf,SAAc5mC,MAAVhE,KAAuBsmC,EAAQW,2BACxC,OAAO;QAEP,MAAMX,EAAQE,GACZ,4BAA4B3C,GAAiB7jC;;;;;;;;GApItC8qC,EAAiB7G,GAAOqC;;;AAKrC,SAASuC,GACP5hC,GACAq/B;IAEA,MAAM39B,IAA0B;IAiBhC,OAfInB,EAAQP;;;IAGNq/B,EAAQ5gC,QAAQ4gC,EAAQ5gC,KAAK3B,SAAS,KACxCuiC,EAAQl0B,GAAU3N,KAAK6hC,EAAQ5gC,QAGjCnB,EAAQ0C,GAAK,CAACE,GAAamS;QACzB,MAAMkwB,IAAcrC,GAAU7tB,GAAKgtB,EAAQyE,GAAqB5jC;QAC7C,QAAfqiC,MACF7gC,EAAOxB,KAAOqiC;QAKb;QAAE9gC,UAAU;YAAEC,QAAAA;;;;;AAsHvB,SAASkhC,GAAoB5F;IAC3B,SACmB,mBAAVA,KACG,SAAVA,KACEA,aAAiBL,SACjBK,aAAiBthC,QACjBshC,aAAiB1hC,KACjB0hC,aAAiBwD,MACjBxD,aAAiBuB,MACjBvB,aAAiBuG,MACjBvG,aAAiBkC;;;AAIvB,SAASwC,GACPhpC,GACA2mC,GACArC;IAEA,KAAK4F,GAAoB5F,OAAWM,GAAcN,IAAQ;QACxD,MAAMO,IAAcX,GAAiBI;QACrC,MAAoB,gBAAhBO,IAEI8B,EAAQE,GAAY7mC,IAAU,sBAE9B2mC,EAAQE,GAAY7mC,IAAU,MAAM6kC;;;;;;aAQhCmF,GACd5C,GACArhC;IAEA,IAAIA,aAAgBsjC,IAClB,OAAOtjC,EAAKqgC;IACP,IAAoB,mBAATrgC,GAChB,OAAOujC,GAAgClC,GAAYrhC;IAGnD,MAAM,IAAIlG,EACRlB,EAAKI,kBACL,YAAYqoC,iCAHE;;;;;;;;;GAepB,UAASkC,GACPlC,GACArhC;IAEA;QACE,gBH1oBmCA;YAErC,IADcA,EAAKslC,OAAO9E,OACb,GACX,MAAM,IAAI1mC,EACRlB,EAAKI,kBACL,uBAAuBgH,gCACrB;YAGN;gBACE,OAAO,IAAIK,MAAaL,EAAKE,MAAM;cACnC,OAAOgpB;gBACP,MAAM,IAAIpvB,EACRlB,EAAKI,kBACL,uBAAuBgH,kCACrB;;SG2nBGulC,CAAuBvlC,GAAMqgC;MACpC,OAAOnX;QACP,MAAMjvB,KAYY+uB,IAZWE,cAaPnvB,QAAQivB,EAAM/uB,UAAU+uB,EAAM7uB;QAZpD,MAAM,IAAIL,EACRlB,EAAKI,kBACL,YAAYqoC,iCAA0CpnC;;;;;;IAS5D,IAAsB+uB;;;AAKtB,SAASya,GAAkBp+B,GAAuBC;IAChD,OAAOD,EAASyE,KAAKtE,KAAKA,EAAEvH,QAAQqH;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCxmBhBkgC;IAkBpBjtC,YACUk4B,GACRgV,GACQC,GACEC,GACFC,GACEC;kBALFpV,aAEAiV,aACEC,aACFC,GACEntC,gBAAAotC,GAnBJptC;;;;;;QAMRA,UAAqB,GAErBA,UAAmD,MAC3CA,cAA+C,MAYrDA,KAAK46B,KAAU,IAAI7C,GAAmBC,GAAOgV;;;;;;;;WAU/CltC;QACE,4BACEE,KAAK0jB,0BACL1jB,KAAK0jB,6BACL1jB,KAAK0jB;;;;;WAQT5jB;QACE,wBAAOE,KAAK0jB;;;;;;;;WAUd5jB;0BACME,KAAK0jB,QAST1jB,KAAK2C,SARH3C,KAAKqtC;;;;;;;WAiBTvtC;QACME,KAAKstC,cACDttC,KAAKutC;;;;;;;;;WAYfztC;QAMEE,KAAK0jB,0BACL1jB,KAAK46B,GAAQvC;;;;;;;;;;;WAafv4B;;;QAGME,KAAKwtC,QAA+B,SAAnBxtC,KAAKytC,OACxBztC,KAAKytC,KAAYztC,KAAKg4B,GAAMc,GAC1B94B,KAAKitC,IAvJW,KAyJhB,MAAMjtC,KAAK0tC;;wDAMP5tC,GAAYkwB;QACpBhwB,KAAK2tC,MACL3tC,KAAK4tC,OAAQC,KAAK7d;;uFAIZlwB;QACN,IAAIE,KAAKwtC;;;QAGP,OAAOxtC,KAAKutC;;gDAKRztC;QACFE,KAAKytC,OACPztC,KAAKytC,GAAUlV,UACfv4B,KAAKytC,KAAY;;;;;;;;;;;;;;WAiBb3tC,YACNguC,GACAvd;;QASAvwB,KAAK2tC,MACL3tC,KAAK46B,GAAQrC;;;QAIbv4B,KAAK+tC,wBAEDD;;QAEF9tC,KAAK46B,GAAQvC,UACJ9H,KAASA,EAAMhvB,SAASpB,EAAKU;;QAEtC2b,GAAS+T,EAAM7uB,aACf8a,GACE;QAEFxc,KAAK46B,GAAQoT,QACJzd,KAASA,EAAMhvB,SAASpB,EAAKS;;;QAGtCZ,KAAKmtC,GAAoBc;;QAIP,SAAhBjuC,KAAK4tC,WACP5tC,KAAKkuC,MACLluC,KAAK4tC,OAAOL,SACZvtC,KAAK4tC,SAAS;;;QAKhB5tC,KAAK0jB,QAAQoqB;;cAGP9tC,KAAKotC,SAASe,GAAQ5d;;;;;WAOpBzwB;IAiBFA;QAMNE,KAAK0jB;QAEL,MAAM0qB,IAAsBpuC,KAAKquC,GAA0BruC,KAAK+tC,KAG1DA,IAAa/tC,KAAK+tC;;gBAExB/tC,KAAKmtC,GAAoBhqC,WAAWH,KAClCsrC;;;;;YAKMtuC,KAAK+tC,OAAeA;;;;YAItB/tC,KAAKuuC,GAAYD;WAGpB/d;YACC6d,EAAoB;gBAClB,MAAMI,IAAW,IAAIntC,EACnBlB,EAAKG,SACL,iCAAiCiwB,EAAM/uB;gBAEzC,OAAOxB,KAAKyuC,GAAkBD;;;;IAM9B1uC,GAAYwuC;QAMlB,MAAMF,IAAsBpuC,KAAKquC,GAA0BruC,KAAK+tC;QAEhE/tC,KAAK4tC,SAAS5tC,KAAK0uC,GAASJ,IAC5BtuC,KAAK4tC,OAAOe,GAAO;YACjBP,EAAoB,OAKlBpuC,KAAK0jB,uBACE1jB,KAAKotC,SAAUuB;YAG1B3uC,KAAK4tC,OAAOO,GAAS5d;YACnB6d,EAAoB,MACXpuC,KAAKyuC,GAAkBle;YAGlCvwB,KAAK4tC,OAAOgB,UAAW5e;YACrBoe,EAAoB,MACXpuC,KAAK4uC,UAAU5e;;;IAKpBlwB;QAKNE,KAAK0jB,0BAEL1jB,KAAK46B,GAAQe,GAAcD;YAMzB17B,KAAK0jB,0BACL1jB,KAAK0f;;;;IAMT5f,GAAkBywB;;;;;QAahB,OARAlJ,GAzbY,oBAybM,qBAAqBkJ,MAEvCvwB,KAAK4tC,SAAS,MAMP5tC,KAAKutC,sBAAmChd;;;;;;;WASzCzwB,GACN+uC;QAEA,OAAQjoC;YACN5G,KAAKg4B,GAAM0C,GAAiB,MACtB16B,KAAK+tC,OAAec,IACfjoC,OAEPygB,GAldM,oBAodJ;YAEKplB,QAAQC;;;;;;;;;;;UA0BZ4sC,WAA+B/B;IAK1CjtC,YACEk4B,GACAkV,GACA6B,GACQr9B,GACR07B;QAEA3rC,MACEu2B,0HAGAkV,GACA6B,GACA3B;QATMptC,kBAAA0R;;IAaA5R,GACRwuC;QAEA,OAAOtuC,KAAKktC,GAAW8B,GACrB,UACAV;;IAIMxuC,UAAUmvC;;QAElBjvC,KAAK46B,GAAQvC;QAEb,MAAMxS,IAAc7lB,KAAK0R,WAAWw9B,GAAgBD,IAC9CE,IAAWnvC,KAAK0R,WAAW09B,GAC/BH;QAEF,OAAOjvC,KAAKotC,SAAUiC,GAAcxpB,GAAaspB;;;;;;;WASnDrvC,GAAMimB;QACJ,MAAMupB,IAAyB;QAC/BA,EAAQl/B,WAAWpQ,KAAK0R,WAAW69B,IACnCD,EAAQE,YAAYxvC,KAAK0R,WAAW2H,GAAS0M;QAE7C,MAAM0pB,IAASzvC,KAAK0R,WAAWg+B,GAAsB3pB;QACjD0pB,MACFH,EAAQG,SAASA,IAGnBzvC,KAAK2vC,GAAYL;;;;;WAOnBxvC,GAAQ+b;QACN,MAAMyzB,IAAyB;QAC/BA,EAAQl/B,WAAWpQ,KAAK0R,WAAW69B,IACnCD,EAAQ9pB,eAAe3J,GACvB7b,KAAK2vC,GAAYL;;;;;;;;;;;;;;;;;;;;UAuCRM,WAA8B7C;IAOzCjtC,YACEk4B,GACAkV,GACA6B,GACQr9B,GACR07B;QAEA3rC,MACEu2B,sHAGAkV,GACA6B,GACA3B;QATMptC,kBAAA0R,GANV1R,WAA6B;;;;;;;;;QA2B7BA,uBAA8BsJ,EAAW6S;;;;;WAMzC0zB;QACE,OAAO7vC,KAAK8vC;;;IAIdhwC;QACEE,KAAK8vC,MAAqB,GAC1BruC,MAAMie;;IAGE5f;QACJE,KAAK8vC,MACP9vC,KAAK+vC,GAAe;;IAIdjwC,GACRwuC;QAEA,OAAOtuC,KAAKktC,GAAW8B,GACrB,SACAV;;IAIMxuC,UAAUkwC;QAQlB;;QANA3sC,KACI2sC,EAAchd,cAGlBhzB,KAAKiwC,kBAAkBjwC,KAAK0R,WAAWyY,GAAU6lB,EAAchd,cAE1DhzB,KAAK8vC,IAQH;;;;YAIL9vC,KAAK46B,GAAQvC;YAEb,MAAMnF,IAAUlzB,KAAK0R,WAAWw+B,GAC9BF,EAAcG,cACdH,EAAcrkB,aAEVoH,IAAgB/yB,KAAK0R,WAAWiY,YACpCqmB,EAAyB;YAE3B,OAAOhwC,KAAKotC,SAAUgD,GAAiBrd,GAAeG;;;QAdtD,OAtqBoB7vB,IAkqBjB2sC,EAAcG,gBAAsD,MAAtCH,EAAcG,aAAavqC,SAG5D5F,KAAK8vC,MAAqB,GACnB9vC,KAAKotC,SAAUiD;;;;;;WAuB1BvwC;;;QAKE,MAAMwvC,IAAwB;QAC9BA,EAAQl/B,WAAWpQ,KAAK0R,WAAW69B,IACnCvvC,KAAK2vC,GAAYL;;4EAInBxvC,GAAeuyB;QAWb,MAAMid,IAAwB;YAC5Btc,aAAahzB,KAAK0R,WAAWkc,GAAQ5tB,KAAKiwC;YAC1CK,QAAQje,EAAUvqB,IAAI8iB,KAAY5qB,KAAK0R,WAAW6+B,GAAW3lB;;QAG/D5qB,KAAK2vC,GAAYL;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AC7sBrB,MAAMkB;IAVN1wC;;;QAGEE,eAAY6F;;;IAQZ/F,YACkBotC,GACA6B,GACAr9B;QAEhBjQ,mBAJgByrC,GACAltC,mBAAA+uC,GACA/uC,kBAAA0R;;+DAMlB5R,GAAqB2wC,GAAiBnB;QACpC,OAAOtvC,KAAK+uC,YACT5rC,WACAH,KAAKsrC,KACGtuC,KAAKktC,GAAWwD,GAAqBD,GAASnB,GAAShB,IAE/DrU,MAAO1J;YAIN,MAHIA,EAAMhvB,SAASpB,EAAKS,mBACtBZ,KAAK+uC,YAAYd,KAEb1d;;;qFAKZzwB,GACE2wC,GACAnB;QAEA,OAAOtvC,KAAK+uC,YACT5rC,WACAH,KAAKsrC,KACGtuC,KAAKktC,GAAWyD,GACrBF,GACAnB,GACAhB,IAGHrU,MAAO1J;YAIN,MAHIA,EAAMhvB,SAASpB,EAAKS,mBACtBZ,KAAK+uC,YAAYd,KAEb1d;;;;;;;;;;;;;;;;;;;;;;;;;MCnDDqgB;IAoBX9wC,YAAoB+wC;kBAAAA;;QAlBpB7wC,UAAuB0gB,MACf1gB,iBAAwB,IAChCA,WAAoB;;;;;QAMpBA,UAAgD;;;;;;;QAQhDA,UAAwC,IAAI8wC;;IAI5ChxC,SAAayO;QAGX,IAFAvO,KAAK+wC,MAED/wC,KAAKqyB,UAAUzsB,SAAS,GAC1B,MAAM,IAAIvE,EACRlB,EAAKI,kBACL;QAGJ,MAAMwhB,UDoDH2Z,eACLmV,GACAtiC;YAEA,MAAMyiC,IAAgBpgB,GAAUigB,IAC1BI,IAAS;gBACb7gC,UAAU4gC,EAAct/B,WAAW69B;gBACnCjtB,WAAW/T,EAAKzG,IAAI2V,KAAKuzB,EAAct/B,WAAW+X,GAAOhM;eAErDyzB,UAAiBF,EAAcL,GAGnC,qBAAqBM,IAEjBlvB,IAAO,IAAI/L;YACjBk7B,EAAS9qC,QAAQ0P;gBACf,MAAMd,IAAMg8B,EAAct/B,WAAWy/B,GAAkBr7B;gBACvDiM,EAAKpN,IAAIK,EAAIhM,IAAItH,YAAYsT;;YAE/B,MAAM7F,IAA0B;YAMhC,OALAZ,EAAKnI,QAAQ4C;gBACX,MAAMgM,IAAM+M,EAAKhf,IAAIiG,EAAItH;gBAvGxB2B,KAwGY2R,IACb7F,EAAO7I,KAAK0O;gBAEP7F;SC7EciiC,CAA2BpxC,KAAK6wC,IAAWtiC;QAQ9D,OAPAwT,EAAK3b,QAAQ4O;YACPA,aAAeY,MAAcZ,aAAe5B,KAC9CpT,KAAKqxC,GAAcr8B,KAEnBlP;YAGGic;;IAGTjiB,IAAIkJ,GAAkBgL;QACpBhU,KAAKsxC,MAAMt9B,EAAKu9B,GAAYvoC,GAAKhJ,KAAKuT,GAAavK,MACnDhJ,KAAKwxC,GAAY1xB,IAAI9W;;IAGvBlJ,OAAOkJ,GAAkBgL;QACvB;YACEhU,KAAKsxC,MAAMt9B,EAAKu9B,GAAYvoC,GAAKhJ,KAAKyxC,GAAsBzoC;UAC5D,OAAOynB;YACPzwB,KAAK0xC,KAAiBjhB;;QAExBzwB,KAAKwxC,GAAY1xB,IAAI9W;;IAGvBlJ,OAAOkJ;QACLhJ,KAAKsxC,MAAM,EAAC,IAAI37B,GAAe3M,GAAKhJ,KAAKuT,GAAavK,QACtDhJ,KAAKwxC,GAAY1xB,IAAI9W;;IAGvBlJ;QAGE,IAFAE,KAAK+wC,MAED/wC,KAAK0xC,IACP,MAAM1xC,KAAK0xC;QAEb,IAAIC,IAAY3xC,KAAK4xC;;gBAErB5xC,KAAKqyB,UAAUjsB,QAAQwkB;YACrB+mB,IAAYA,EAAU10B,OAAO2N,EAAS5hB;;;;QAIxC2oC,EAAUvrC,QAAQ,CAAC4C,GAAK6oC;YACtB7xC,KAAKqyB,UAAU/rB,KAAK,IAAIuP,GAAe7M,GAAKhJ,KAAKuT,GAAavK;kBDX7D0yB,eACLmV,GACAxe;YAEA,MAAM2e,IAAgBpgB,GAAUigB,IAC1BI,IAAS;gBACb7gC,UAAU4gC,EAAct/B,WAAW69B;gBACnCe,QAAQje,EAAUvqB,IAAI6qB,KAAKqe,EAAct/B,WAAW6+B,GAAW5d;eAE3Due,UAAiBF,EAAcN,GAGnC,UAAUO;YACZ,OAAOD,EAAct/B,WAAWw+B,GAC9BgB,EAASf,cACTe,EAASvlB;SCFHmmB,CAAgB9xC,KAAK6wC,IAAW7wC,KAAKqyB,YAC3CryB,KAAK+xC,MAAY;;IAGXjyC,GAAckV;QACpB,IAAIg9B;QAEJ,IAAIh9B,aAAe5B,IACjB4+B,IAAah9B,EAAInC,cACZ;YAAA,MAAImC,aAAeY,KAIxB,MAxGJ9P;;YAsGIksC,IAAa3sC,EAAgB2B;;QAK/B,MAAMirC,IAAkBjyC,KAAK4xC,GAAa7uC,IAAIiS,EAAIhM;QAClD,IAAwB,SAApBipC;YACF,KAAKD,EAAWxsC,QAAQysC;;YAEtB,MAAM,IAAI5wC,EACRlB,EAAKY,SACL;eAIJf,KAAK4xC,KAAe5xC,KAAK4xC,GAAa90B,GAAO9H,EAAIhM,KAAKgpC;;;;;WAQlDlyC,GAAakJ;QACnB,MAAM6J,IAAU7S,KAAK4xC,GAAa7uC,IAAIiG;QACtC,QAAKhJ,KAAKwxC,GAAY3xB,IAAI7W,MAAQ6J,IACzBE,GAAaC,WAAWH,KAExBE,GAAawY;;;;WAOhBzrB,GAAsBkJ;QAC5B,MAAM6J,IAAU7S,KAAK4xC,GAAa7uC,IAAIiG;;;gBAGtC,KAAKhJ,KAAKwxC,GAAY3xB,IAAI7W,MAAQ6J,GAAS;YACzC,IAAIA,EAAQrN,QAAQH,EAAgB2B;;;;;;;;;;YAYlC,MAAM,IAAI3F,EACRlB,EAAKI,kBACL;;wBAIJ,OAAOwS,GAAaC,WAAWH;;;;QAI/B,OAAOE,GAAaE,QAAO;;IAIvBnT,MAAMuyB;QACZryB,KAAK+wC,MACL/wC,KAAKqyB,YAAYryB,KAAKqyB,UAAUnZ,OAAOmZ;;IAGjCvyB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCvJGoyC;IAyBXpyC,YACU85B,GACAuY;kBADAvY,aACAuY;;QAzBFnyC;;;;;;QAORA,UAA8B;;;;;;QAO9BA,UAA0D;;;;;;QAO1DA,WAAoC;;;;;;;;WAcpCF;QACmC,MAA7BE,KAAKoyC,OACPpyC,KAAKqyC,6BAMLryC,KAAKsyC,KAAmBtyC,KAAK45B,GAAWd,qDA1Dd,KA6DxB,OACE94B,KAAKsyC,KAAmB;QAKxBtyC,KAAKuyC,GACH,8CAGFvyC,KAAKqyC;QAMEpwC,QAAQC;;;;;;;WAYvBpC,GAAyBywB;kCACnBvwB,KAAK0jB,QACP1jB,KAAKqyC,+BAaLryC,KAAKoyC;QACDpyC,KAAKoyC,MA/GmB,MAgH1BpyC,KAAKwyC,MAELxyC,KAAKuyC,GACH,yBACE,6BAA6BhiB,EAAM7uB;QAGvC1B,KAAKqyC;;;;;;;;WAYXvyC,IAAI2yC;QACFzyC,KAAKwyC,MACLxyC,KAAKoyC,KAAsB,6BAEvBK;;;QAGFzyC,KAAK0yC,MAA4B,IAGnC1yC,KAAKqyC,GAAgBI;;IAGf3yC,GAAgB2yC;QAClBA,MAAazyC,KAAK0jB,UACpB1jB,KAAK0jB,QAAQ+uB,GACbzyC,KAAKmyC,GAAmBM;;IAIpB3yC,GAAmC6yC;QACzC,MAAMnxC,IACJ,4CAA4CmxC,QAC5C;QAGE3yC,KAAK0yC,MACPl2B,GAAShb,IACTxB,KAAK0yC,MAA4B,KAEjCrrB,GAxKU,sBAwKQ7lB;;IAId1B;QACwB,SAA1BE,KAAKsyC,OACPtyC,KAAKsyC,GAAiB/Z,UACtBv4B,KAAKsyC,KAAmB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCpHjBM;IAqDX9yC;;;;IAIU+yC;;IAEAhC,GACAjX,GACRuY,GACAW;kBALQD,aAEAhC,aACAjX;;;;;;;;;;;;;;;;;;QA1CV55B,UAAyC;;;;;;;;;;QAWzCA,UAAwB,IAAIgW,KAK5BhW,UAA8D;;;;;QAMtDA,uBAAiB,GAEjBA,kBAAY;;;;;;QAOpBA,WAA0B,GAexBA,KAAK8yC,KAAsBA,GAC3B9yC,KAAK8yC,GAAoBC,GAAazqB;YACpCsR,EAAWc,GAAiBgB;gBACtB17B,KAAKgzC,SACP3rB,GA5FM,eA8FJ;sBAEIrnB,KAAKizC;;YAKjBjzC,KAAKkzC,KAAqB,IAAIhB,GAC5BtY,GACAuY;;QAIFnyC,KAAKmzC,cH8BPtC,GACA7Y,GACAoV;YAEA,MAAM4D,IAAgBpgB,GAAUigB;YAChC,OAAO,IAAI/B,GACT9W,GACAgZ,EAAc9D,IACd8D,EAAcjC,aACdiC,EAAct/B,YACd07B;SGxCmBgG,CAAyBpzC,KAAK6wC,IAAWjX,GAAY;YACtEyZ,IAAQrzC,KAAKszC,GAAkBtZ,KAAKh6B;YACpCuzC,IAASvzC,KAAKwzC,GAAmBxZ,KAAKh6B;YACtCyzC,IAAezzC,KAAK0zC,GAAoB1Z,KAAKh6B;YAG/CA,KAAK2zC,cHSP9C,GACA7Y,GACAoV;YAEA,MAAM4D,IAAgBpgB,GAAUigB;YAChC,OAAO,IAAIjB,GACT5X,GACAgZ,EAAc9D,IACd8D,EAAcjC,aACdiC,EAAct/B,YACd07B;SGnBmBwG,CAAyB5zC,KAAK6wC,IAAWjX,GAAY;YACtEyZ,IAAQrzC,KAAK6zC,GAAkB7Z,KAAKh6B;YACpCuzC,IAASvzC,KAAK8zC,GAAmB9Z,KAAKh6B;YACtC+zC,IAAqB/zC,KAAKg0C,GAAyBha,KAAKh6B;YACxDi0C,IAAkBj0C,KAAKowC,GAAiBpW,KAAKh6B;;;;;;WAcjDF;QACE,OAAOE,KAAKk0C;;kDAIdp0C;QAEE,OADAE,KAAKm0C,kBAAiB,GACfn0C,KAAKo0C;;IAGNt0C;QACFE,KAAKgzC,SACPhzC,KAAK2zC,GAAY1D,wBAAwBjwC,KAAK6yC,GAAW5S,MAErDjgC,KAAKq0C,OACPr0C,KAAKs0C,OAELt0C,KAAKkzC,GAAmBv+B;;cAIpB3U,KAAKu0C;;;;;WAQfz0C;QACEE,KAAKm0C,kBAAiB,SAChBn0C,KAAKw0C;;QAGXx0C,KAAKkzC,GAAmBv+B;;IAGlB7U;cACAE,KAAK2zC,GAAYc,cACjBz0C,KAAKmzC,GAAYsB,QAEnBz0C,KAAK00C,GAAc9uC,SAAS,MAC9ByhB,GA7KU,eA+KR,8BAA8BrnB,KAAK00C,GAAc9uC;QAEnD5F,KAAK00C,KAAgB,KAGvB10C,KAAK20C;;IAGP70C;QACEunB,GAxLY,eAwLM,+BAClBrnB,KAAKm0C,kBAAiB,SAChBn0C,KAAKw0C;QACXx0C,KAAK8yC,GAAoB8B;;;QAIzB50C,KAAKkzC,GAAmBv+B;;;;;WAO1B7U,OAAOimB;QACD/lB,KAAK60C,GAAch1B,IAAIkG,EAAWlK;;QAKtC7b,KAAK60C,GAAclgC,IAAIoR,EAAWlK,UAAUkK,IAExC/lB,KAAKq0C;;QAEPr0C,KAAKs0C,OACIt0C,KAAKmzC,GAAY3F,QAC1BxtC,KAAK80C,GAAiB/uB;;;;;WAQ1BjmB,GAAS+b;QAMP7b,KAAK60C,GAAcjgC,OAAOiH,IACtB7b,KAAKmzC,GAAY3F,QACnBxtC,KAAK+0C,GAAmBl5B,IAGM,MAA5B7b,KAAK60C,GAAcruC,SACjBxG,KAAKmzC,GAAY3F,OACnBxtC,KAAKmzC,GAAY6B,OACRh1C,KAAKgzC;;;;QAIdhzC,KAAKkzC,GAAmBv+B;;oEAM9B7U,GAAuB+b;QACrB,OAAO7b,KAAK60C,GAAc9xC,IAAI8Y,MAAa;;oEAI7C/b,GAAuB+b;QACrB,OAAO7b,KAAKi1C,GAAWhuB,GAAuBpL;;;;;WAOxC/b,GAAiBimB;QACvB/lB,KAAKk1C,GAAuBhuB,GAA2BnB,EAAWlK,WAClE7b,KAAKmzC,GAAYgC,GAAMpvB;;;;;;WAQjBjmB,GAAmB+b;QACzB7b,KAAKk1C,GAAuBhuB,GAA2BrL,IACvD7b,KAAKmzC,GAAYiC,GAAQv5B;;IAGnB/b;QAMNE,KAAKk1C,KAAwB,IAAIzwB,GAAsBzkB,OACvDA,KAAKmzC,GAAYzzB,SACjB1f,KAAKkzC,GAAmBmC;;;;;WAOlBv1C;QACN,OACEE,KAAKgzC,SACJhzC,KAAKmzC,GAAY7F,QAClBttC,KAAK60C,GAAcruC,OAAO;;IAI9B1G;QACE,QAAQE,KAAKs1C,MAAmBt1C,KAAKu1C,aAAav1C,KAAKm0C;;IAGjDr0C;QACNE,KAAKk1C,KAAwB;;IAGvBp1C;QACNE,KAAK60C,GAAczuC,QAAQ,CAAC2f,GAAYlK;YACtC7b,KAAK80C,GAAiB/uB;;;IAIlBjmB,SAAyBywB;QAU/BvwB,KAAK20C;;QAGD30C,KAAKq0C,QACPr0C,KAAKkzC,GAAmBsC,OAExBx1C,KAAKs0C;;;;QAKLt0C,KAAKkzC,GAAmBv+B;;IAIpB7U,SACN+lB,GACA7J;QAKA;;QAFAhc,KAAKkzC,GAAmBv+B,4BAGtBkR,aAAuBpC,0BACvBoC,EAAYnC,SACZmC,EAAYjC;;;QAIZ;kBACQ5jB,KAAKy1C,GAAkB5vB;UAC7B,OAAO4K;YACPpJ,GA7VQ,eA+VN,oCACAxB,EAAYlC,UAAUtc,KAAK,MAC3BopB;kBAEIzwB,KAAK01C,GAA4BjlB;eAiB3C,IAZI5K,aAAuB1C,KACzBnjB,KAAKk1C,GAAuBS,GAAqB9vB,KACxCA,aAAuBtC,KAChCvjB,KAAKk1C,GAAuBU,GAAsB/vB,KAMlD7lB,KAAKk1C,GAAuBW,GAAmBhwB;SAG5C7J,EAAgBxW,QAAQH,EAAgB2B,QAC3C;YACE,MAAMw6B,UAAkCxhC,KAAK6yC,GAAW1S;YACpDnkB,EAAgBvO,EAAU+zB,MAA8B;;;kBAGpDxhC,KAAK81C,GAAmB95B;UAEhC,OAAOyU;YACPpJ,GA7XQ,eA6XU,6BAA6BoJ,UACzCzwB,KAAK01C,GAA4BjlB;;;;;;;WAUrC3wB,SAAkC2wB;QACxC,KAAIiJ,GAA4BjJ,IAsB9B,MAAMA;QAjBNzwB,KAAKs1C,MAAkB;;cAGjBt1C,KAAKw0C,MACXx0C,KAAKkzC,GAAmBv+B;;QAGxB3U,KAAK45B,GAAWmc,GAAiBra;YAC/BrU,GAtZQ,eAsZU;;;;kBAIZrnB,KAAK6yC,GAAW1S,MACtBngC,KAAKs1C,MAAkB,SACjBt1C,KAAKo0C;;;;;;;WAYTt0C,GAAmBkc;QAKzB,MAAM2K,IAAc3mB,KAAKk1C,GAAuBc,GAC9Ch6B;;;;QAuDF,OAlDA2K,EAAYjE,GAActc,QAAQ,CAACob,GAAQ3F;YACzC,IAAI2F,EAAOtF,YAAYoI,OAAwB,GAAG;gBAChD,MAAMyB,IAAa/lB,KAAK60C,GAAc9xC,IAAI8Y;;gCAEtCkK,KACF/lB,KAAK60C,GAAclgC,IACjBkH,GACAkK,EAAW4a,GAAgBnf,EAAOtF,aAAaF;;;;;QAQvD2K,EAAYhE,GAAiBvc,QAAQyV;YACnC,MAAMkK,IAAa/lB,KAAK60C,GAAc9xC,IAAI8Y;YAC1C,KAAKkK;;YAEH;;;wBAKF/lB,KAAK60C,GAAclgC,IACjBkH,GACAkK,EAAW4a,GACTr3B,EAAW6S,IACX4J,EAAW/J;;;YAMfhc,KAAK+0C,GAAmBl5B;;;;;YAMxB,MAAMo6B,IAAoB,IAAIt6B,GAC5BoK,EAAWnK,QACXC,qCAEAkK,EAAWhK;YAEb/b,KAAK80C,GAAiBmB;YAIjBj2C,KAAKi1C,GAAWiB,GAAiBvvB;;2CAIlC7mB,SACN+lB;QAGA,MAAM0K,IAAQ1K,EAAkB;QAChC,KAAK,MAAMhK,KAAYgK,EAAYlC;;QAE7B3jB,KAAK60C,GAAch1B,IAAIhE,aACnB7b,KAAKi1C,GAAWkB,GAAat6B,GAAU0U,IAC7CvwB,KAAK60C,GAAcjgC,OAAOiH,IAC1B7b,KAAKk1C,GAAuB1vB,aAAa3J;;;;;;;;;WAa/C/b;QACE,IAAIE,KAAKo2C,MAAyB;YAChC,MAAMC,IACJr2C,KAAK00C,GAAc9uC,SAAS,IACxB5F,KAAK00C,GAAc10C,KAAK00C,GAAc9uC,SAAS,GAAGusB,WzB5hB/B,GyB8hBnBW,UAAc9yB,KAAK6yC,GAAWyD,GAClCD;YAGY,SAAVvjB,IACgC,MAA9B9yB,KAAK00C,GAAc9uC,UACrB5F,KAAK2zC,GAAYqB,QAGnBh1C,KAAKu2C,GAAmBzjB,UAClB9yB,KAAKu0C;;QAIXv0C,KAAKw2C,QACPx2C,KAAKy2C;;;;;WAQD32C;QACN,OACEE,KAAKgzC,QAAmBhzC,KAAK00C,GAAc9uC,SA5hBtB;;;IAiiBzB9F;QACE,OAAOE,KAAK00C,GAAc9uC;;;;;WAOpB9F,GAAmBgzB;QAKzB9yB,KAAK00C,GAAcpuC,KAAKwsB,IAEpB9yB,KAAK2zC,GAAYnG,QAAYxtC,KAAK2zC,GAAY+C,MAChD12C,KAAK2zC,GAAY5D,GAAejd,EAAMT;;IAIlCvyB;QACN,OACEE,KAAKgzC,SACJhzC,KAAK2zC,GAAYrG,QAClBttC,KAAK00C,GAAc9uC,SAAS;;IAIxB9F;QAKNE,KAAK2zC,GAAYj0B;;IAGX5f;QACNE,KAAK2zC,GAAYgD;;IAGX72C;;QAEN,OAAOE,KAAK6yC,GACT3S,GAAmBlgC,KAAK2zC,GAAY1D,iBACpCjtC,KAAK;;YAEJ,KAAK,MAAM8vB,KAAS9yB,KAAK00C,IACvB10C,KAAK2zC,GAAY5D,GAAejd,EAAMT;WAGzC4H,MAAMoJ;;IAGHvjC,GACNizB,GACAG;QAQA,MAAMJ,IAAQ9yB,KAAK00C,GAAckC,SAC3BC,IAAUhkB,GAAoBnG,KAClCoG,GACAC,GACAG,GACAlzB,KAAK2zC,GAAY1D;QAEnB,OAAOjwC,KAAKi1C,GAAW6B,GAAqBD,GAAS7zC,KAAK,MAGjDhD,KAAKu0C;;IAIRz0C,SAAyBywB;;;QAY3BA,KAASvwB,KAAK00C,GAAc9uC,SAAS,MACnC5F,KAAK2zC,GAAY+C;;cAEb12C,KAAK+2C;;;;cAKL/2C,KAAKg3C;;;QAKTh3C,KAAKw2C,QACPx2C,KAAKy2C;;;IAMH32C,SAA2BywB;;;;QAIjC,IAAIjU,GAAiBiU,EAAMhvB,OAQzB,OAPA8lB,GAtpBU,eAwpBR,0EACArnB,KAAK2zC,GAAY1D;QAEnBjwC,KAAK2zC,GAAY1D,kBAAkB3mC,EAAW6S,IAEvCnc,KAAK6yC,GACT3S,GAAmB52B,EAAW6S,IAC9B8d,MAAMoJ;;IAOLvjC,SAAuBywB;;;QAG7B,IxC7nBKjU,GAD6B/a,IwC8nBRgvB,EAAMhvB,SxC7nBDA,MAASpB,EAAKY,SwC6nBN;;;YAGrC,MAAM+xB,IAAQ9yB,KAAK00C,GAAckC;;;;wBAOjC,OAFA52C,KAAK2zC,GAAYsD,MAEVj3C,KAAKi1C,GACTiC,GAAkBpkB,EAAMX,SAAS5B,GACjCvtB,KAAK,MAGGhD,KAAKu0C;;YxC7oBgBhzC;;;;;;;IwCopBpCzB;QACE,OAAO,IAAI8wC,GAAY5wC,KAAK6wC;;IAGtB/wC;QACNE,KAAKm0C,kBAAiB,SAChBn0C,KAAKw0C,MACXx0C,KAAKkzC,GAAmBv+B;cAClB3U,KAAKk0C;;IAGbp0C;QACME,KAAKgzC;;;;QAIP3rB,GA/sBU,eA+sBQ,4DACZrnB,KAAKizC;;;;WAOfnzC,SAAwBy1C;QACtBv1C,KAAKu1C,YAAYA,GAEbA,KAAav1C,KAAKm0C,uBACdn0C,KAAKk0C,kBACDqB,YACJv1C,KAAKw0C;QACXx0C,KAAKkzC,GAAmBv+B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;MCjWjBwiC;IAAbr3C;QACEE,uBAAkB8gB;;IAElBhhB,GAAe+b;QACb7b,KAAKo3C,kBAAkBp3C,KAAKo3C,gBAAgBt3B,IAAIjE;;IAGlD/b,GAAkB+b;QAChB7b,KAAKo3C,kBAAkBp3C,KAAKo3C,gBAAgBxiC,OAAOiH;;;;;WAOrD/b;QACE,MAAMkU,IAA0B;YAC9BojC,iBAAiBp3C,KAAKo3C,gBAAgBhwC;YACtCiwC,cAAc7yC,KAAKC;;QAErB,OAAO+hC,KAAKC,UAAUzyB;;;;;;;;UAwlBbsjC;IAAbx3C;QACEE,UAAqB,IAAIm3C,IACzBn3C,UAA+D,IAE/DA,UAA6C,MAC7CA,UAAkE,MAClEA,UAEW;;IAEXF,GAAmBqyB;;;IAInBryB,GACEqyB,GACAzO,GACA6M;;;IAKFzwB,GAAoB+b;QAElB,OADA7b,KAAKu3C,GAAWC,GAAe37B,IACxB7b,KAAKy3C,GAAW57B,MAAa;;IAGtC/b,GACE+b,GACA6H,GACA6M;QAEAvwB,KAAKy3C,GAAW57B,KAAY6H;;IAG9B5jB,GAAuB+b;QACrB7b,KAAKu3C,GAAWG,GAAkB77B;;IAGpC/b,GAAmB+b;QACjB,OAAO7b,KAAKu3C,GAAWH,gBAAgBv3B,IAAIhE;;IAG7C/b,GAAgB+b;eACP7b,KAAKy3C,GAAW57B;;IAGzB/b;QACE,OAAOE,KAAKu3C,GAAWH;;IAGzBt3C,GAAoB+b;QAClB,OAAO7b,KAAKu3C,GAAWH,gBAAgBv3B,IAAIhE;;IAG7C/b;QAEE,OADAE,KAAKu3C,KAAa,IAAIJ,IACfl1C,QAAQC;;IAGjBpC,GACEgC,GACA28B,GACAC;;;IAKF5+B,GAAe63C;;;IAIf73C;IAEAA,GAAoBic;;;;;;;;;;;;;;;;;;UCtkCT67B;IACX93C,YAAmBkJ;QAAAhJ,WAAAgJ;;;;MAER6uC;IACX/3C,YAAmBkJ;QAAAhJ,WAAAgJ;;;;;;;;UA6BR8uC;IAeXh4C,YACUgiB;;IAEAi2B;QAFA/3C,aAAA8hB,aAEAi2B,GAjBV/3C,UAAsC;;;;;;;QAOtCA,WAAkB;;QAGlBA,UAAyB4gB;;QAEzB5gB,UAAsB4gB,MAOpB5gB,KAAKg4C,KAAc,IAAIj3B,GAAYe,EAAMm2B,GAAcje,KAAKlY;;;;;WAO9Do2B;QACE,OAAOl4C,KAAK+3C;;;;;;;;;;;WAadj4C,GACEmiB,GACAk2B;QAEA,MAAMC,IAAYD,IACdA,EAAgBC,KAChB,IAAI72B,IACF82B,IAAiBF,IACnBA,EAAgBH,KAChBh4C,KAAKg4C;QACT,IAAIM,IAAiBH,IACjBA,EAAgBj2B,KAChBliB,KAAKkiB,IACLq2B,IAAiBF,GACjBG,KAAc;;;;;;;;;QAWlB,MAAMC,IACJz4C,KAAK8hB,MAAM42B,QAAqBL,EAAe7xC,SAASxG,KAAK8hB,MAAM3b,QAC/DkyC,EAAeM,SACf,MACAC,IACJ54C,KAAK8hB,MAAM+2B,QAAoBR,EAAe7xC,SAASxG,KAAK8hB,MAAM3b,QAC9DkyC,EAAejpC,UACf;;QAwFN,IAtFA6S,EAAWzE,GACT,CAACxU,GAAkB8vC;YACjB,MAAMC,IAASV,EAAet1C,IAAIiG;YAClC,IAAIsa,IAASw1B,aAAuB1lC,KAAW0lC,IAAc;YACzDx1B,MAQFA,IAAStjB,KAAK8hB,MAAMnH,QAAQ2I,KAAUA,IAAS;YAGjD,MAAM01B,MAA4BD,KAC9B/4C,KAAKkiB,GAAYrC,IAAIk5B,EAAO/vC,MAE1BiwC,MAA4B31B,MAC9BA,EAAO/L;;;YAGNvX,KAAKkiB,GAAYrC,IAAIyD,EAAOta,QAAQsa,EAAO5P;YAGhD,IAAIwlC,KAAgB;;wBAGpB,IAAIH,KAAUz1B,GAAQ;gBACFy1B,EAAO/kC,OAAOxO,QAAQ8d,EAAOtP,UAqBpCglC,MAA8BC,MACvCb,EAAUe,MAAM;oBAAEx3B;oBAA2B3M,KAAKsO;oBAClD41B,KAAgB,KArBXl5C,KAAKo5C,GAA4BL,GAAQz1B,OAC5C80B,EAAUe,MAAM;oBACdx3B;oBACA3M,KAAKsO;oBAEP41B,KAAgB,IAGbT,KACCz4C,KAAK8hB,MAAMm2B,GAAc30B,GAAQm1B,KAAkB,KACpDG,KACC54C,KAAK8hB,MAAMm2B,GAAc30B,GAAQs1B,KAAmB;;;;gBAKtDJ,KAAc;oBAOVO,KAAUz1B,KACpB80B,EAAUe,MAAM;gBAAEx3B;gBAAwB3M,KAAKsO;gBAC/C41B,KAAgB,KACPH,MAAWz1B,MACpB80B,EAAUe,MAAM;gBAAEx3B;gBAA0B3M,KAAK+jC;gBACjDG,KAAgB,IAEZT,KAAkBG;;;;YAIpBJ,KAAc;YAIdU,MACE51B,KACFi1B,IAAiBA,EAAez4B,IAAIwD,IAElCg1B,IADEW,IACeX,EAAex4B,IAAI9W,KAEnBsvC,EAAe1jC,OAAO5L,OAGzCuvC,IAAiBA,EAAe3jC,OAAO5L,IACvCsvC,IAAiBA,EAAe1jC,OAAO5L;YAO3ChJ,KAAK8hB,MAAM42B,QAAqB14C,KAAK8hB,MAAM+2B,MAC7C,MAAON,EAAe/xC,OAAOxG,KAAK8hB,MAAY,SAAE;YAC9C,MAAMi3B,IAAS/4C,KAAK8hB,MAAM42B,OACtBH,EAAeI,SACfJ,EAAenpC;YACnBmpC,IAAiBA,EAAe3jC,OAAOmkC,EAAQ/vC,MAC/CsvC,IAAiBA,EAAe1jC,OAAOmkC,EAAQ/vC,MAC/CovC,EAAUe,MAAM;gBAAEx3B;gBAA0B3M;;;QAQhD,OAAO;YACLqkC,IAAad;YACbe,IAAAlB;YACAmB,IAAAf;YACAgB,IAAalB;;;IAITx4C,GACNi5C,GACAz1B;;;;;;;;QASA,OACEy1B,EAAOxhC,MACP+L,EAAO5P,0BACN4P,EAAO/L;;;;;;;;;;;;;IAeZzX,GACEmiB,GACAw3B,GACA10B;QAMA,MAAM/C,IAAUhiB,KAAKg4C;QACrBh4C,KAAKg4C,KAAc/1B,EAAW+1B,IAC9Bh4C,KAAKkiB,KAAcD,EAAWC;;QAE9B,MAAMN,IAAUK,EAAWm2B,GAAUsB;QACrC93B,EAAQnT,KAAK,CAACkrC,GAAIC,MAsLtB,SAA2BD,GAAgBC;YACzC,MAAMvrB,IAAS7M;gBACb,QAAQA;kBACN;oBACE,OAAO;;kBACT;kBAEA;;;;oBAIE,OAAO;;kBACT;oBACE,OAAO;;kBACT;oBACE,OAtdY1b;;;YA0dlB,OAAOuoB,EAAMsrB,KAAMtrB,EAAMurB;;;;;;;;;;;;;;;;;GAvMnBC,EAAkBF,EAAGh4B,MAAMi4B,EAAGj4B,SAC9B3hB,KAAK8hB,MAAMm2B,GAAc0B,EAAG3kC,KAAK4kC,EAAG5kC,OAIxChV,KAAK85C,GAAkB/0B;QACvB,MAAMg1B,IAAeN,IACjBz5C,KAAKy5C,OACL,IAEEO,IADsC,MAA7Bh6C,KAAKi6C,GAAezzC,QAAcxG,KAAKkI,sCAEhDka,IAAmB43B,MAAiBh6C,KAAKk6C;QAG/C,IAFAl6C,KAAKk6C,KAAYF,GAEM,MAAnBp4B,EAAQhc,UAAiBwc,GAGtB;YAWL,OAAO;gBACL+sB,UAXyB,IAAIttB,GAC7B7hB,KAAK8hB,OACLG,EAAW+1B,IACXh2B,GACAJ,GACAK,EAAWC,sBACX83B,GACA53B;gDAC+B;gBAI/B+3B,IAAAJ;;;;QAdF,OAAO;YAAEI,IAAAJ;;;;;;WAuBbj6C,GAAuB63C;QACrB,OAAI33C,KAAKkI,kCAAWyvC;;;;;QAKlB33C,KAAKkI,MAAU,GACRlI,KAAKo6C,GACV;YACEf,IAAar5C,KAAKg4C;YAClBsB,IAAW,IAAI/3B;YACfi4B,IAAax5C,KAAKkiB;YAClBq3B,KAAa;;qCAEa,MAIvB;YAAEY,IAAc;;;;;WAOnBr6C,GAAgBkJ;;QAEtB,QAAIhJ,KAAK+3C,GAAiBl4B,IAAI7W;;UAIzBhJ,KAAKg4C,GAAYn4B,IAAI7W,OAOtBhJ,KAAKg4C,GAAYj1C,IAAIiG,GAAMuO;;;;;WAWzBzX,GAAkBilB;QACpBA,MACFA,EAAa/B,GAAe5c,QAC1B4C,KAAQhJ,KAAK+3C,KAAmB/3C,KAAK+3C,GAAiBj4B,IAAI9W,KAE5D+b,EAAa9B,GAAkB7c,QAAQ4C,UAMvC+b,EAAa7B,GAAiB9c,QAC5B4C,KAAQhJ,KAAK+3C,KAAmB/3C,KAAK+3C,GAAiBnjC,OAAO5L;QAE/DhJ,KAAKkI,KAAU6c,EAAa7c;;IAIxBpI;;QAEN,KAAKE,KAAKkI,IACR,OAAO;;;gBAKT,MAAMmyC,IAAoBr6C,KAAKi6C;QAC/Bj6C,KAAKi6C,KAAiBr5B,MACtB5gB,KAAKg4C,GAAY5xC,QAAQ4O;YACnBhV,KAAKs6C,GAAgBtlC,EAAIhM,SAC3BhJ,KAAKi6C,KAAiBj6C,KAAKi6C,GAAen6B,IAAI9K,EAAIhM;;;QAKtD,MAAM4Y,IAAiC;QAWvC,OAVAy4B,EAAkBj0C,QAAQ4C;YACnBhJ,KAAKi6C,GAAep6B,IAAI7W,MAC3B4Y,EAAQtb,KAAK,IAAIuxC,GAAqB7uC;YAG1ChJ,KAAKi6C,GAAe7zC,QAAQ4C;YACrBqxC,EAAkBx6B,IAAI7W,MACzB4Y,EAAQtb,KAAK,IAAIsxC,GAAmB5uC;YAGjC4Y;;;;;;;;;;;;;;;;;;;;;;IAuBT9hB,GAA8By6C;QAC5Bv6C,KAAK+3C,KAAmBwC,EAAY3X,IACpC5iC,KAAKi6C,KAAiBr5B;QACtB,MAAMqB,IAAajiB,KAAKw6C,GAAkBD,EAAYj4B;QACtD,OAAOtiB,KAAKo6C,GAAan4B,8BAAsC;;;;;;;;IASjEniB;QACE,OAAO+hB,GAAa44B,GAClBz6C,KAAK8hB,OACL9hB,KAAKg4C,IACLh4C,KAAKkiB,sBACLliB,KAAKk6C;;;;;;;;MC5bEQ;IAIX56C,YACmB85B,GACA+gB,GACAC,GACA7gB;kBAHAH,aACA+gB,GACA36C,sBAAA46C,aACA7gB,GAPnB/5B,UAPkB,GAgBhBA,KAAK46B,KAAU,IAAI7C,GACjB/3B,KAAK45B;;oEAMT95B;QACEE,KAAK66C;;IAGC/6C;QACNE,KAAK46B,GAAQe,GAAcD;YACzB,MAAM3G,IAAc/0B,KAAK26C,GAAYG,MAC/BC,IAAc/6C,KAAKg7C,GAAqBjmB;YAC1CgmB,KACFA,EACG/3C,KAAKmM;gBACJnP,KAAK45B,GAAWc,GAAiB,MACxB3F,EACJkmB,SACAj4C,KAAK;oBACJhD,KAAK+5B,GAAS73B,QAAQiN;mBAEvB8qB,MAAMihB;oBACLl7C,KAAKm7C,GAAuBD;;eAInCjhB,MAAMmhB;gBACLp7C,KAAKm7C,GAAuBC;;;;IAM9Bt7C,GAAqBi1B;QAC3B;YACE,MAAMgmB,IAAc/6C,KAAK46C,eAAe7lB;YACxC,QACE3qB,EAAkB2wC,MACjBA,EAAY9gB,SACZ8gB,EAAY/3C,OAOR+3C,KALL/6C,KAAK+5B,GAASjG,OACZxyB,MAAM;YAED;UAGT,OAAOivB;;YAGP,OADAvwB,KAAK+5B,GAASjG,OAAOvD,IACd;;;IAIHzwB,GAAuBywB;QACzBvwB,KAAKq7C,KAAU,KAAKr7C,KAAKs7C,GAA4B/qB,MACvDvwB,KAAKq7C,MAAW,GAChBr7C,KAAK45B,GAAWc,GAAiB,OAC/B16B,KAAK66C,MACE54C,QAAQC,eAGjBlC,KAAK+5B,GAASjG,OAAOvD;;IAIjBzwB,GAA4BywB;QAClC,IAAmB,oBAAfA,EAAM5uB,MAA0B;;;YAGlC,MAAMJ,IAAQgvB,EAAyBhvB;YACvC,OACW,cAATA,KACS,0BAATA,MACC+a,GAAiB/a;;QAGtB,QAAO;;;;;;;;;;;;;;;;;;;;;;;;ACvCX,MAAMg6C;IACJz7C;;;;IAISgiB;;;;;IAKAjG;;;;;;;IAOA2/B;QAZAx7C,aAAA8hB,GAKA9hB,gBAAA6b,GAOA7b,YAAAw7C;;;;iCAKX,OAAMC;IACJ37C,YAAmBkJ;QAAAhJ,WAAAgJ;;;;;;;QAQnBhJ,WAA4B;;;;;;;;;;;;;;;;;UAgCjB07C;IAsCX57C,YACY+yC,GACA8H;;IAEAgB,GACFn5C,GACAo5C;kBALE/I,aACA8H,aAEAgB,GACF37C,mBAAAwC,aACAo5C,GA3CV57C,UAA0D;QAE1DA,UAA8B,IAAI4xB,GAA4BiqB,KAC5DA,EAAEjtC,gBAEJ5O,UAA4B,IAAIgW;;;;;QAKhChW,UAAkD;;;;;QAKlDA,UAAoC,IAAI0c,GACtCnU,EAAYvC;;;;;QAMdhG,UAA2C,IAAIgW,KAI/ChW,UAA8B,IAAIsjC;;QAElCtjC,UAAgC;;QAIhCA,UAAiC,IAAIgW,KACrChW,UAAiCw5B,GAAkBsiB,MAE3C97C;;IAWR+7C;QACE,QAAO;;qFAITj8C,UAAUk8C;QAURh8C,KAAKg8C,KAAqBA;;;;;;WAQ5Bl8C,aAAagiB;QAGX,IAAIjG,GACAsb;QAHJn3B,KAAKi8C,GAAiB;QAKtB,MAAMC,IAAYl8C,KAAKm8C,GAAkBp5C,IAAI+e;QAC7C,IAAIo6B;;;;;;;QAOFrgC,IAAWqgC,EAAUrgC,UACrB7b,KAAK27C,GAAkBS,GAAoBvgC,IAC3Csb,IAAe+kB,EAAUV,KAAKa,WACzB;YACL,MAAMt2B,UAAmB/lB,KAAK6yC,GAAWyJ,GAAex6B,EAAMzI,OAExDiP,IAAStoB,KAAK27C,GAAkBS,GACpCr2B,EAAWlK;YAEbA,IAAWkK,EAAWlK,UACtBsb,UAAqBn3B,KAAKu8C,GACxBz6B,GACAjG,GACW,cAAXyM,IAEEtoB,KAAKw8C,MACPx8C,KAAK26C,GAAY8B,OAAO12B;;QAI5B,OAAOoR;;;;;WAOCr3B,SACRgiB,GACAjG,GACA3T;QAEA,MAAMqyC,UAAoBv6C,KAAK6yC,GAAW6J,GACxC56B;mCAC0B,IAEtB05B,IAAO,IAAI1D,GAAKh2B,GAAOy4B,EAAY3X,KACnC+Z,IAAiBnB,EAAKhB,GAAkBD,EAAYj4B,YACpDs6B,IAA0B95B,GAAaC,GAC3ClH,GACA3T,iCAAWlI,KAAK23C,cAEZ9V,IAAa2Z,EAAKpB,GACtBuC;oCAC4B38C,KAAKw8C,IACjCI;QAEF58C,KAAK68C,GAAoBhhC,GAAUgmB,EAAWkY;QAO9C,MAAM/lC,IAAO,IAAIunC,GAAUz5B,GAAOjG,GAAU2/B;QAO5C,OANAx7C,KAAKm8C,GAAkBxnC,IAAImN,GAAO9N,IAC9BhU,KAAK88C,GAAgBj9B,IAAIhE,KAC3B7b,KAAK88C,GAAgB/5C,IAAI8Y,GAAWvV,KAAKwb,KAEzC9hB,KAAK88C,GAAgBnoC,IAAIkH,GAAU,EAACiG;QAE/B+f,EAAWsN;;4CAIpBrvC,SAAegiB;QACb9hB,KAAKi8C,GAAiB;QAEtB,MAAMC,IAAYl8C,KAAKm8C,GAAkBp5C,IAAI+e,IAKvCi7B,IAAU/8C,KAAK88C,GAAgB/5C,IAAIm5C,EAAUrgC;;;gBACnD,IAAIkhC,EAAQn3C,SAAS,GAMnB,OALA5F,KAAK88C,GAAgBnoC,IACnBunC,EAAUrgC,UACVkhC,EAAQr1C,OAAOm0C,MAAMA,EAAEr2C,QAAQsc;aAEjC9hB,KAAKm8C,GAAkBvnC,OAAOkN;;gBAKhC,IAAI9hB,KAAKw8C,IAAiB;;;YAGxBx8C,KAAK27C,GAAkBqB,GAAuBd,EAAUrgC,WAC5B7b,KAAK27C,GAAkBsB,GACjDf,EAAUrgC,mBAIJ7b,KAAK6yC,GACRqK,GAAchB,EAAUrgC,wCAAuC,GAC/D7Y,KAAK;gBACJhD,KAAK27C,GAAkBwB,GAAgBjB,EAAUrgC,WACjD7b,KAAK26C,GAAYyC,GAASlB,EAAUrgC,WACpC7b,KAAKq9C,GAAuBnB,EAAUrgC;eAEvCoe,MAAMoJ;eAGXrjC,KAAKq9C,GAAuBnB,EAAUrgC,iBAChC7b,KAAK6yC,GAAWqK,GACpBhB,EAAUrgC;sCACmB;;;;;;;;;;;WAenC/b,YAAYgzB,GAAmBwqB;QAC7Bt9C,KAAKi8C,GAAiB;QAEtB;YACE,MAAM9sC,UAAenP,KAAK6yC,GAAW0K,GAAWzqB;YAChD9yB,KAAK27C,GAAkB6B,GAAmBruC,EAAOgjB,UACjDnyB,KAAKy9C,GAAoBtuC,EAAOgjB,SAASmrB,UACnCt9C,KAAK09C,GAAgCvuC,EAAOyS,WAC5C5hB,KAAK26C,GAAYpG;UACvB,OAAO9jB;;;YAGP,MAAMF,IAAQmM,GAA6BjM,GAAG;YAC9C6sB,EAAaxpB,OAAOvD;;;;;;;;;;;;;;;;;;;WAqBxBzwB,eACE85B,GACAghB,GACA7gB;QAEA,IAAI2gB,GACF9gB,GACA55B,KAAK26C,IACLC,GACA7gB,GACA4jB;;IAGJ79C,SAAuB6mB;QACrB3mB,KAAKi8C,GAAiB;QACtB;YACE,MAAMr6B,UAAgB5hB,KAAK6yC,GAAWqD,GAAiBvvB;;wBAEvDA,EAAYjE,GAActc,QAAQ,CAAC2e,GAAclJ;gBAC/C,MAAM+hC,IAAkB59C,KAAK69C,GAA+B96C,IAC1D8Y;gBAEE+hC;;;gBA9W8Bv6C,GAkX9B0hB,EAAa/B,GAAexc,OAC1Bue,EAAa9B,GAAkBzc,OAC/Bue,EAAa7B,GAAiB1c,QAC9B,IAGAue,EAAa/B,GAAexc,OAAO,IACrCo3C,EAAgBE,MAAmB,IAC1B/4B,EAAa9B,GAAkBzc,OAAO,IA1XjBnD,GA4X5Bu6C,EAAgBE,MAGT/4B,EAAa7B,GAAiB1c,OAAO,MA/XhBnD,GAiY5Bu6C,EAAgBE;gBAGlBF,EAAgBE,MAAmB;sBAMnC99C,KAAK09C,GAAgC97B,GAAS+E;UACpD,OAAO4J;kBACD8S,GAAyB9S;;;;;;WAQnCzwB,GACE63C,GACAoG;QAEA/9C,KAAKi8C,GAAiB;QACtB,MAAM+B,IAAmB;QACzBh+C,KAAKm8C,GAAkB/1C,QAAQ,CAAC0b,GAAOo6B;YACrC,MAAMra,IAAaqa,EAAUV,KAAKyC,GAAuBtG;YAKrD9V,EAAWsN,YACb6O,EAAiB13C,KAAKu7B,EAAWsN;YAGrCnvC,KAAKg8C,GAAoBkC,GAAoBvG,IAC7C33C,KAAKg8C,GAAoB3M,GAAc2O,IACvCh+C,KAAK23C,cAAcA;;IAGrB73C,SAAmB+b,GAAoBsY;QACrCn0B,KAAKi8C,GAAiB;;QAGtBj8C,KAAK27C,GAAkBwC,GAAiBtiC,GAAU,YAAYsY;QAE9D,MAAMypB,IAAkB59C,KAAK69C,GAA+B96C,IAAI8Y,IAC1DuiC,IAAWR,KAAmBA,EAAgB50C;QACpD,IAAIo1C,GAAU;;;;;;;YAQZ,IAAIx7B,IAAkB,IAAIlG,GACxBnU,EAAYvC;YAEd4c,IAAkBA,EAAgB9F,GAChCshC,GACA,IAAIxoC,GAAWwoC,GAAU/4C,EAAgB2B;YAE3C,MAAM6b,IAAyBjC,KAAiBd,IAAIs+B,IAC9CC,IAAQ,IAAI57B,GAChBpd,EAAgB2B;iCACK,IAAIgP;oCACD,IAAIkJ,GAAoBja,KAChD2d,GACAC;kBAGI7iB,KAAKk2C,GAAiBmI;;;;;;YAO5Br+C,KAAKs+C,KAA0Bt+C,KAAKs+C,GAAwBrhC,OAC1DmhC,IAEFp+C,KAAK69C,GAA+BjpC,OAAOiH,IAC3C7b,KAAKu+C;qBAECv+C,KAAK6yC,GACRqK,GAAcrhC,kCAAwC,GACtD7Y,KAAK,MAAMhD,KAAKq9C,GAAuBxhC,GAAUsY,IACjD8F,MAAMoJ;;IAIbvjC,SACE0+C;QAEAx+C,KAAKi8C,GAAiB;QAEtB,MAAM9pB,IAAUqsB,EAAoB1rB,MAAMX;;;;;gBAM1CnyB,KAAKy+C,GAAoBtsB,cAAoB,OAE7CnyB,KAAK0+C,GAA8BvsB;QAEnC;YACE,MAAMvQ,UAAgB5hB,KAAK6yC,GAAWnT,GACpC8e;YAEFx+C,KAAK27C,GAAkBgD,GAAoBxsB,GAAS,uBAC9CnyB,KAAK09C,GAAgC97B;UAC3C,OAAO2O;kBACD8S,GAAyB9S;;;IAInCzwB,SACEqyB,GACA5B;QAEAvwB,KAAKi8C,GAAiB;;;;;QAMtBj8C,KAAKy+C,GAAoBtsB,GAAS5B,IAElCvwB,KAAK0+C,GAA8BvsB;QAEnC;YACE,MAAMvQ,UAAgB5hB,KAAK6yC,GAAW+L,GAAYzsB;YAClDnyB,KAAK27C,GAAkBgD,GAAoBxsB,GAAS,YAAY5B,UAC1DvwB,KAAK09C,GAAgC97B;UAC3C,OAAO2O;kBACD8S,GAAyB9S;;;;;;WAQnCzwB,SAAoCuzB;QAC7BrzB,KAAK26C,GAAY3H,QACpB3rB,GAhfU,cAkfR;QAKJ;YACE,MAAMw3B,UAAuB7+C,KAAK6yC,GAAW7S;YAC7C,K7BpiByB,M6BoiBrB6e;;YAGF,YADAxrB,EAASnxB;YAIX,MAAM48C,IAAY9+C,KAAK++C,GAAuBh8C,IAAI87C,MAAmB;YACrEC,EAAUx4C,KAAK+sB,IACfrzB,KAAK++C,GAAuBpqC,IAAIkqC,GAAgBC;UAChD,OAAOruB;YACP,MAAMuuB,IAAiBtiB,GACrBjM,GACA;YAEF4C,EAASS,OAAOkrB;;;;;;WAQZl/C,GAA8BqyB;SACnCnyB,KAAK++C,GAAuBh8C,IAAIovB,MAAY,IAAI/rB,QAAQitB;YACvDA,EAASnxB;YAGXlC,KAAK++C,GAAuBnqC,OAAOud;;uFAI7BryB,GAAwCm/C;QAC9Cj/C,KAAK++C,GAAuB34C,QAAQ04C;YAClCA,EAAU14C,QAAQitB;gBAChBA,EAASS,OAAO,IAAIzyB,EAAelB,EAAKE,WAAW4+C;;YAIvDj/C,KAAK++C,GAAuBG;;IAGtBp/C,GACNqyB,GACAkB;QAEA,IAAI8rB,IAAen/C,KAAKo/C,GAAsBp/C,KAAKwC,YAAY68C;QAC1DF,MACHA,IAAe,IAAIziC,GACjBzX,MAGJk6C,IAAeA,EAAariC,GAAOqV,GAASkB,IAC5CrzB,KAAKo/C,GAAsBp/C,KAAKwC,YAAY68C,OAAWF;;;;;WAO/Cr/C,GAAoBqyB,GAAkB5B;QAC9C,IAAI4uB,IAAen/C,KAAKo/C,GAAsBp/C,KAAKwC,YAAY68C;;;gBAI/D,IAAIF,GAAc;YAChB,MAAM9rB,IAAW8rB,EAAap8C,IAAIovB;YAC9BkB,MAKE9C,IACF8C,EAASS,OAAOvD,KAEhB8C,EAASnxB,WAEXi9C,IAAeA,EAAaliC,OAAOkV,KAErCnyB,KAAKo/C,GAAsBp/C,KAAKwC,YAAY68C,OAAWF;;;IAIjDr/C,GACR+b,GACA0U,IAAsB;QAEtBvwB,KAAK27C,GAAkBqB,GAAuBnhC;QAQ9C,KAAK,MAAMiG,KAAS9hB,KAAK88C,GAAgB/5C,IAAI8Y,IAC3C7b,KAAKm8C,GAAkBvnC,OAAOkN,IAC1ByO,KACFvwB,KAAKg8C,GAAoBsD,GAAax9B,GAAOyO;QAMjD,IAFAvwB,KAAK88C,GAAgBloC,OAAOiH,IAExB7b,KAAKw8C,IAAiB;YACNx8C,KAAKu/C,GAAkBC,GAAsB3jC,GACrDzV,QAAQg4C;gBACKp+C,KAAKu/C,GAAkBE,GAAYrB;;gBAGtDp+C,KAAK0/C,GAAkBtB;;;;IAMvBt+C,GAAkBkJ;;;QAGxB,MAAM22C,IAAgB3/C,KAAKs+C,GAAwBv7C,IAAIiG;QACjC,SAAlB22C,MAKJ3/C,KAAK26C,GAAYyC,GAASuC,IAC1B3/C,KAAKs+C,KAA0Bt+C,KAAKs+C,GAAwBrhC,OAAOjU,IACnEhJ,KAAK69C,GAA+BjpC,OAAO+qC,IAC3C3/C,KAAKu+C;;IAGGz+C,GACR+b,GACAk+B;QAEA,KAAK,MAAM6F,KAAe7F,GACxB,IAAI6F,aAAuBhI,IACzB53C,KAAKu/C,GAAkBzd,GAAa8d,EAAY52C,KAAK6S,IACrD7b,KAAK6/C,GAAiBD,SACjB,IAAIA,aAAuB/H,IAAsB;YACtDxwB,GApoBQ,cAooBU,kCAAkCu4B,EAAY52C,MAChEhJ,KAAKu/C,GAAkBxd,GAAgB6d,EAAY52C,KAAK6S;YACnC7b,KAAKu/C,GAAkBE,GAC1CG,EAAY52C;;YAIZhJ,KAAK0/C,GAAkBE,EAAY52C;eAGrClD;;IAKEhG,GAAiB8/C;QACvB,MAAM52C,IAAM42C,EAAY52C;QACnBhJ,KAAKs+C,GAAwBv7C,IAAIiG,OACpCqe,GAtpBU,cAspBQ,4BAA4Bre,IAC9ChJ,KAAK8/C,GAAyBx5C,KAAK0C;QACnChJ,KAAKu+C;;;;;;;;;WAYDz+C;QACN,MACEE,KAAK8/C,GAAyBl6C,SAAS,KACvC5F,KAAKs+C,GAAwB93C,OAAOxG,KAAK47C,MACzC;YACA,MAAM5yC,IAAMhJ,KAAK8/C,GAAyBlJ,SACpC+I,IAAgB3/C,KAAK+/C,GAAuBz3C;YAClDtI,KAAK69C,GAA+BlpC,IAClCgrC,GACA,IAAIlE,GAAgBzyC,KAEtBhJ,KAAKs+C,KAA0Bt+C,KAAKs+C,GAAwBxhC,GAC1D9T,GACA22C,IAEF3/C,KAAK26C,GAAY8B,OACf,IAAI9gC,GACFzD,GAAMoU,GAAOtjB,EAAIzB,MAAM8R,MACvBsmC,6BAEAvoB,GAAe4oB;;;;IAOvBlgD;QACE,OAAOE,KAAKs+C;;;IAIdx+C;QACE,OAAOE,KAAK8/C;;IAGJhgD,SACR8hB,GACA+E;QAEA,MAAMs5B,IAA2B,IAC3BC,IAA2C,IAC3CC,IAAyC;QAE/CngD,KAAKm8C,GAAkB/1C,QAAQ,CAACwf,GAAGs2B;YACjCiE,EAAiB75C,KACfrE,QAAQC,UACLc,KAAK;gBACJ,MAAM25C,IAAiBT,EAAUV,KAAKhB,GAAkB54B;gBACxD,OAAK+6B,EAAenE,KAMbx4C,KAAK6yC,GACT6J,GAAaR,EAAUp6B,kCAAiC,GACxD9e,KAAK,EAAGsf,WAAAA,OACA45B,EAAUV,KAAKhB,GACpBl4B,GACAq6B,MAVGA;;;;2BAcV35C,KAAM25C;gBACL,MAAM53B,IACJ4B,KAAeA,EAAYjE,GAAc3f,IAAIm5C,EAAUrgC,WACnDgmB,IAAaqa,EAAUV,KAAKpB,GAChCuC;4CAC4B38C,KAAKw8C,IACjCz3B;gBAMF,IAJA/kB,KAAK68C,GACHX,EAAUrgC,UACVgmB,EAAWkY,KAETlY,EAAWsN,UAAU;oBACnBnvC,KAAKw8C,MACPx8C,KAAK27C,GAAkBwC,GACrBjC,EAAUrgC,UACVgmB,EAAWsN,SAAShtB,YAAY,gBAAgB;oBAIpD89B,EAAS35C,KAAKu7B,EAAWsN;oBACzB,MAAMltB,IAAa+U,GAAiBopB,GAClClE,EAAUrgC,UACVgmB,EAAWsN;oBAEb+Q,EAAqB55C,KAAK2b;;;kBAM9BhgB,QAAQ+xB,IAAImsB,IAClBngD,KAAKg8C,GAAoB3M,GAAc4Q,UACjCjgD,KAAK6yC,GAAWwN,GAAuBH;;IAGrCpgD,GAAiBwgD;IAO3BxgD,SAA6BgC;QAG3B,KAFqB9B,KAAKwC,YAAYgD,QAAQ1D,IAE7B;YACf,MAAMqN,UAAenP,KAAK6yC,GAAW0N,GAAiBz+C;YACtD9B,KAAKwC,cAAcV;;YAGnB9B,KAAKwgD,GACH;;YAGFxgD,KAAK27C,GAAkB4E,GACrBz+C,GACAqN,EAAOsvB,IACPtvB,EAAOuvB,WAEH1+B,KAAK09C,GAAgCvuC,EAAO0vB;;cAG9C7+B,KAAK26C,GAAY8F;;IAGzB3gD;QACE,OAAOE,KAAK26C,GAAYzG;;IAG1Bp0C;QACE,OAAOE,KAAK26C,GAAY+F;;IAG1B5gD,GAAuB+b;QACrB,MAAM+hC,IAAkB59C,KAAK69C,GAA+B96C,IAAI8Y;QAChE,IAAI+hC,KAAmBA,EAAgBE,IACrC,OAAOl9B,KAAiBd,IAAI89B,EAAgB50C;QACvC;YACL,IAAI23C,IAAS//B;YACb,MAAMm8B,IAAU/8C,KAAK88C,GAAgB/5C,IAAI8Y;YACzC,KAAKkhC,GACH,OAAO4D;YAET,KAAK,MAAM7+B,KAASi7B,GAAS;gBAC3B,MAAMb,IAAYl8C,KAAKm8C,GAAkBp5C,IAAI+e;gBAE7C6+B,IAASA,EAAOC,GAAU1E,EAAUV,KAAKqF;;YAE3C,OAAOF;;;;;;;;;;;;;;;;;;;;;;;;GC52Bb,OAAMG;IAANhhD;QACEE,eAAqC6F,GACrC7F,UAA6B;;;;;;;;UAgBlB+gD;IASXjhD,YAAoBm1C;kBAAAA,GARpBj1C,UAAkB,IAAI4xB,GAAqCiqB,KACzDA,EAAEjtC,gBAGI5O;QAERA,UAAwD,IAAI8wC,KAG1D9wC,KAAKi1C,GAAW+L,UAAUhhD;;IAG5BF,aAAastC;QACX,MAAMtrB,IAAQsrB,EAAStrB;QACvB,IAAIm/B,KAAc,GAEdC,IAAYlhD,KAAK+8C,GAAQh6C,IAAI+e;QAMjC,IALKo/B,MACHD,KAAc,GACdC,IAAY,IAAIJ,KAGdG,GACF;YACEC,EAAUC,WAAiBnhD,KAAKi1C,GAAWwH,OAAO36B;UAClD,OAAO2O;YACP,MAAMuuB,IAAiBtiB,GACrBjM,GACA,4BAA4B2c,EAAStrB;YAGvC,YADAsrB,EAASgU,QAAQpC;;QAKrBh/C,KAAK+8C,GAAQpoC,IAAImN,GAAOo/B,IACxBA,EAAUG,GAAU/6C,KAAK8mC;;QAGLA,EAAS6Q,GAAuBj+C,KAAK23C;QAMzD,IAAIuJ,EAAUC,IAAU;YACF/T,EAASkU,GAAeJ,EAAUC,OAEpDnhD,KAAKuhD;;;IAKXzhD,SAAestC;QACb,MAAMtrB,IAAQsrB,EAAStrB;QACvB,IAAI0/B,KAAa;QAEjB,MAAMN,IAAYlhD,KAAK+8C,GAAQh6C,IAAI+e;QACnC,IAAIo/B,GAAW;YACb,MAAMx6C,IAAIw6C,EAAUG,GAAU75C,QAAQ4lC;YAClC1mC,KAAK,MACPw6C,EAAUG,GAAUpvB,OAAOvrB,GAAG,IAC9B86C,IAA4C,MAA/BN,EAAUG,GAAUz7C;;QAIrC,IAAI47C,GAEF,OADAxhD,KAAK+8C,GAAQnoC,OAAOkN,IACb9hB,KAAKi1C,GAAWmI,GAASt7B;;IAIpChiB,GAAc2hD;QACZ,IAAIC,KAAc;QAClB,KAAK,MAAMP,KAAYM,GAAW;YAChC,MAAM3/B,IAAQq/B,EAASr/B,OACjBo/B,IAAYlhD,KAAK+8C,GAAQh6C,IAAI+e;YACnC,IAAIo/B,GAAW;gBACb,KAAK,MAAM9T,KAAY8T,EAAUG,IAC3BjU,EAASkU,GAAeH,OAC1BO,KAAc;gBAGlBR,EAAUC,KAAWA;;;QAGrBO,KACF1hD,KAAKuhD;;IAITzhD,GAAagiB,GAAcyO;QACzB,MAAM2wB,IAAYlhD,KAAK+8C,GAAQh6C,IAAI+e;QACnC,IAAIo/B,GACF,KAAK,MAAM9T,KAAY8T,EAAUG,IAC/BjU,EAASgU,QAAQ7wB;;;gBAMrBvwB,KAAK+8C,GAAQnoC,OAAOkN;;IAGtBhiB,GAAoB63C;QAClB33C,KAAK23C,cAAcA;QACnB,IAAI+J,KAAc;QAClB1hD,KAAK+8C,GAAQ32C,QAAQ,CAACwf,GAAGs7B;YACvB,KAAK,MAAM9T,KAAY8T,EAAUG;;YAE3BjU,EAAS6Q,GAAuBtG,OAClC+J,KAAc;YAIhBA,KACF1hD,KAAKuhD;;IAITzhD,GAA2B6hD;QACzB3hD,KAAK4hD,GAAyB9hC,IAAI6hC;;;QAGlCA,EAASr5C;;IAGXxI,GAA8B6hD;QAC5B3hD,KAAK4hD,GAAyBhtC,OAAO+sC;;;IAI/B7hD;QACNE,KAAK4hD,GAAyBx7C,QAAQu7C;YACpCA,EAASr5C;;;;;;;;;;UAsBFu5C;IAaX/hD,YACWgiB,GACDggC,GACRxqC;QAFStX,aAAA8hB,aACDggC;;;;;QAVV9hD,WAA6B,GAI7BA,UAAoC,MAE5BA,6CAONA,KAAKsX,UAAUA,KAAW;;;;;;;WAS5BxX,GAAeiiD;QAMb,KAAK/hD,KAAKsX,QAAQ0qC,wBAAwB;;YAExC,MAAM//B,IAAmC;YACzC,KAAK,MAAM2C,KAAam9B,EAAK9/B,iCACvB2C,EAAUjD,QACZM,EAAW3b,KAAKse;YAGpBm9B,IAAO,IAAIlgC,GACTkgC,EAAKjgC,OACLigC,EAAKhgC,MACLggC,EAAK//B,IACLC,GACA8/B,EAAK7/B,IACL6/B,EAAK5/B,WACL4/B,EAAK3/B;4CAC0B;;QAGnC,IAAIs/B,KAAc;QAYlB,OAXK1hD,KAAKiiD,KAKCjiD,KAAKkiD,GAAiBH,OAC/B/hD,KAAK8hD,GAAcx5C,KAAKy5C,IACxBL,KAAc,KANV1hD,KAAKmiD,GAAwBJ,GAAM/hD,KAAK23C,iBAC1C33C,KAAKoiD,GAAkBL;QACvBL,KAAc,IAOlB1hD,KAAK+hD,KAAOA,GACLL;;IAGT5hD,QAAQywB;QACNvwB,KAAK8hD,GAAcvxB,MAAMA;;qDAI3BzwB,GAAuB63C;QACrB33C,KAAK23C,cAAcA;QACnB,IAAI+J,KAAc;QASlB,OAPE1hD,KAAK+hD,OACJ/hD,KAAKiiD,MACNjiD,KAAKmiD,GAAwBniD,KAAK+hD,IAAMpK,OAExC33C,KAAKoiD,GAAkBpiD,KAAK+hD,KAC5BL,KAAc;QAETA;;IAGD5hD,GACNiiD,GACApK;;QAQA,KAAKoK,EAAK5/B,WACR,QAAO;;;gBAKT,MAAMkgC,gCAAc1K;;;gBAGpB,SAAI33C,KAAKsX,QAAQgrC,OAAyBD,QASlCN,EAAKhgC,KAAK1Y,mCAAasuC;;;IAGzB73C,GAAiBiiD;;;;;QAKvB,IAAIA,EAAK9/B,WAAWrc,SAAS,GAC3B,QAAO;QAGT,MAAM28C,IACJviD,KAAK+hD,MAAQ/hD,KAAK+hD,GAAKvqC,qBAAqBuqC,EAAKvqC;QACnD,UAAIuqC,EAAK3/B,OAAoBmgC,OACoB,MAAxCviD,KAAKsX,QAAQ0qC;;;;;IAShBliD,GAAkBiiD;QAKxBA,IAAOlgC,GAAa44B,GAClBsH,EAAKjgC,OACLigC,EAAKhgC,MACLggC,EAAK7/B,IACL6/B,EAAK5/B,YAEPniB,KAAKiiD,MAAqB,GAC1BjiD,KAAK8hD,GAAcx5C,KAAKy5C;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;UC9SfS;IAGX1iD,GAAsBg+B;QACpB99B,KAAKyiD,KAAqB3kB;;IAG5Bh+B,GACEi1B,GACAjT,GACA7F,GACA2mB;;;;QAUA,OAAI9gB,EAAM4gC,QAMNzmC,EAA6BzW,QAAQH,EAAgB2B,SALhDhH,KAAK2iD,GAA0B5tB,GAAajT,KAS9C9hB,KAAKyiD,GAAoB7jB,GAAa7J,GAAa6N,GAAYt6B,KACpEga;YACE,MAAMsgC,IAAkB5iD,KAAK6iD,GAAW/gC,GAAOQ;YAE/C,QACGR,EAAM42B,QAAqB52B,EAAM+2B,SAClC74C,KAAKw4C,GACH12B,EAAM1J,IACNwqC,GACAhgB,GACA3mB,KAGKjc,KAAK2iD,GAA0B5tB,GAAajT,MAGjD8N,QAAiBK,EAASC,SAC5B7I,GACE,wBACA,yDACApL,EAA6Bva,YAC7BogB,EAAMpgB;YAMH1B,KAAKyiD,GAAoBnsB,GAC9BvB,GACAjT,GACA7F,GACA3T,KAAKw6C;;;;YAILF,EAAgBx8C,QAAQ4O;gBACtB8tC,IAAiBA,EAAehmC,GAAO9H,EAAIhM,KAAKgM;gBAE3C8tC;;;;;+EAOPhjD,GACNgiB,GACAQ;;;QAIA,IAAIiU,IAAe,IAAIrX,GAAoB,CAAC5F,GAAIC,MAC9CuI,EAAMm2B,GAAc3+B,GAAIC;QAO1B,OALA+I,EAAUlc,QAAQ,CAACwf,GAAGzS;YAChBA,aAAoBC,MAAY0O,EAAMnH,QAAQxH,OAChDojB,IAAeA,EAAazW,IAAI3M;YAG7BojB;;;;;;;;;;;;WAcDz2B,GACNsY,GACA2qC,GACAngB,GACAogB;;;QAIA,IAAIpgB,EAAWp8B,SAASu8C,EAAsBv8C,MAC5C,QAAO;;;;;;;;;gBAWT,MAAMy8C,wBACJ7qC,IACI2qC,EAAsBpK,SACtBoK,EAAsB3zC;QAC5B,SAAK6zC,MAKHA,EAAezrC,oBACfyrC,EAAepwC,QAAQpF,EAAUu1C,KAA4B;;IAIzDljD,GACNi1B,GACAjT;QAUA,OARI8N,QAAiBK,EAASC,SAC5B7I,GACE,wBACA,gDACAvF,EAAMpgB;QAIH1B,KAAKyiD,GAAoBnsB,GAC9BvB,GACAjT,GACAzc,EAAgB2B;;;;;;;;;;;;;;;;;;;UC3KTk8C;IAmBXpjD,YACmBg1B,GACAuM;kBADAvM,aACAuM;;;;;QAhBnBrhC,UAAyC;;QAGzCA,UAA+B;;;;;QAMvBA,uBAA8BsJ,EAAW6S;;QAGjDnc,UAA+B,IAAIkf,GAAUqkB,GAAaC;;IAO1D1jC,GAAWi1B;QACT,OAAO3B,GAAmBlxB,QAAsC,MAA9BlC,KAAK60B,GAAcjvB;;IAGvD9F,GACEi1B,GACAjC,GACAE;QAEA,MAAMb,IAAUW,EAAMX,SAChBgxB,IAAanjD,KAAKojD,GAAuBjxB,GAAS;QA7B1B9uB,GA+Bb,MAAf8/C;;QAKYnjD,KAAK60B,GAAcsuB;QAUjC,OADAnjD,KAAKiwC,kBAAkBjd,GAChBI,GAAmBlxB;;IAG5BpC,GACEi1B;QAEA,OAAO3B,GAAmBlxB,QAAQlC,KAAKiwC;;IAGzCnwC,GACEi1B,GACA/B;QAGA,OADAhzB,KAAKiwC,kBAAkBjd,GAChBI,GAAmBlxB;;IAG5BpC,GACEi1B,GACApqB,GACAynB,GACAC;QAIA,MAAMF,IAAUnyB,KAAKqjD;QAGrB,IAFArjD,KAAKqjD,MAEDrjD,KAAK60B,GAAcjvB,SAAS,GAAG;YACnB5F,KAAK60B,GAAc70B,KAAK60B,GAAcjvB,SAAS;;QAO/D,MAAMktB,IAAQ,IAAIZ,GAChBC,GACAxnB,GACAynB,GACAC;QAEFryB,KAAK60B,GAAcvuB,KAAKwsB;;QAGxB,KAAK,MAAMlI,KAAYyH,GACrBryB,KAAKsjD,KAAuBtjD,KAAKsjD,GAAqBxjC,IACpD,IAAIyjB,GAAa3Y,EAAS5hB,KAAKmpB,KAGjCnyB,KAAK80B,GAAayuB,GAChBxuB,GACAnK,EAAS5hB,IAAIzB,KAAKqlB;QAItB,OAAOwG,GAAmBlxB,QAAQ4wB;;IAGpChzB,GACEi1B,GACA5C;QAEA,OAAOiB,GAAmBlxB,QAAQlC,KAAKwjD,GAAkBrxB;;IAG3DryB,GACEi1B,GACA5C;QAEA,MAAMkxB,IAAclxB,IAAU,GAIxBsxB,IAAWzjD,KAAK0jD,GAAeL,IAC/B58C,IAAQg9C,IAAW,IAAI,IAAIA;;;gBACjC,OAAOrwB,GAAmBlxB,QACxBlC,KAAK60B,GAAcjvB,SAASa,IAAQzG,KAAK60B,GAAcpuB,KAAS;;IAIpE3G;QACE,OAAOszB,GAAmBlxB,QACM,MAA9BlC,KAAK60B,GAAcjvB,UhCvIM,IgCuI2B5F,KAAKqjD,KAAc;;IAI3EvjD,GACEi1B;QAEA,OAAO3B,GAAmBlxB,QAAQlC,KAAK60B,GAAc3uB;;IAGvDpG,GACEi1B,GACA4uB;QAEA,MAAMjkC,IAAQ,IAAI6jB,GAAaogB,GAAa,IACtC98C,IAAM,IAAI08B,GAAaogB,GAAa/zC,OAAOg0C,oBAC3Cz0C,IAA0B;QAchC,OAbAnP,KAAKsjD,GAAqBrf,GAAe,EAACvkB,GAAO7Y,KAAM88B;YAKrD,MAAM7Q,IAAQ9yB,KAAKwjD,GAAkB7f,EAAIS;YAKzCj1B,EAAO7I,KAAKwsB;YAGPM,GAAmBlxB,QAAQiN;;IAGpCrP,GACEi1B,GACA8uB;QAEA,IAAIC,IAAiB,IAAI5kC,GAAkBja;QAe3C,OAbA4+C,EAAaz9C,QAAQu9C;YACnB,MAAMjkC,IAAQ,IAAI6jB,GAAaogB,GAAa,IACtC98C,IAAM,IAAI08B,GAAaogB,GAAa/zC,OAAOg0C;YACjD5jD,KAAKsjD,GAAqBrf,GAAe,EAACvkB,GAAO7Y,KAAM88B;gBAMrDmgB,IAAiBA,EAAehkC,IAAI6jB,EAAIS;;YAIrChR,GAAmBlxB,QAAQlC,KAAK+jD,GAAoBD;;IAG7DhkD,GACEi1B,GACAjT;;;QAQA,MAAMkiC,IAASliC,EAAMva,MACf08C,IAA8BD,EAAOp+C,SAAS;;;;;QAMpD,IAAIs+C,IAAYF;QACXz7C,EAAY0P,GAAcisC,OAC7BA,IAAYA,EAAUltC,MAAM;QAG9B,MAAM0I,IAAQ,IAAI6jB,GAAa,IAAIh7B,EAAY27C,IAAY;;;gBAI3D,IAAIJ,IAAiB,IAAI5kC,GAAkBja;QAmB3C,OAjBAjF,KAAKsjD,GAAqB58B,GAAaid;YACrC,MAAMwgB,IAAaxgB,EAAI36B,IAAIzB;YAC3B,SAAKy8C,EAAO1xC,EAAW6xC;;;;;;YAQjBA,EAAWv+C,WAAWq+C,MACxBH,IAAiBA,EAAehkC,IAAI6jB,EAAIS,OAEnC;WAER1kB,IAEI0T,GAAmBlxB,QAAQlC,KAAK+jD,GAAoBD;;IAGrDhkD,GAAoBskD;;;QAG1B,MAAMj1C,IAA0B;QAOhC,OANAi1C,EAASh+C,QAAQ+rB;YACf,MAAMW,IAAQ9yB,KAAKwjD,GAAkBrxB;YACvB,SAAVW,KACF3jB,EAAO7I,KAAKwsB;YAGT3jB;;IAGTrP,GACEi1B,GACAjC;QAvP8BzvB,GA4Pb,MAFErD,KAAKojD,GAAuBtwB,EAAMX,SAAS,aAK9DnyB,KAAK60B,GAAc+hB;QAEnB,IAAIyN,IAAarkD,KAAKsjD;QACtB,OAAOlwB,GAAmBhtB,QAAQ0sB,EAAMT,WAAYzH;YAClD,MAAM+Y,IAAM,IAAIJ,GAAa3Y,EAAS5hB,KAAK8pB,EAAMX;YAEjD,OADAkyB,IAAaA,EAAWzvC,OAAO+uB,IACxB3jC,KAAKqhC,GAAkBijB,GAC5BvvB,GACAnK,EAAS5hB;WAEVV,KAAK;YACNtI,KAAKsjD,KAAuBe;;;IAIhCvkD,GAAyBqyB;;;IAIzBryB,GACEs+B,GACAp1B;QAEA,MAAM26B,IAAM,IAAIJ,GAAav6B,GAAK,IAC5Bk7B,IAAWlkC,KAAKsjD,GAAqBnf,GAAkBR;QAC7D,OAAOvQ,GAAmBlxB,QAAQ8G,EAAIxD,QAAQ0+B,KAAYA,EAASl7B;;IAGrElJ,GACEs+B;QAQA,OANIp+B,KAAK60B,GAAcjvB,QAMhBwtB,GAAmBlxB;;;;;;;;;WAWpBpC,GAAuBqyB,GAAkB5U;QAM/C,OALcvd,KAAK0jD,GAAevxB;;;;;;;;;;WAiB5BryB,GAAeqyB;QACrB,IAAkC,MAA9BnyB,KAAK60B,GAAcjvB;;QAErB,OAAO;;;;;gBAQT,OAAOusB,IADcnyB,KAAK60B,GAAc,GAAG1C;;;;;WAQrCryB,GAAkBqyB;QACxB,MAAM1rB,IAAQzG,KAAK0jD,GAAevxB;QAClC,OAAI1rB,IAAQ,KAAKA,KAASzG,KAAK60B,GAAcjvB,SACpC,OAGK5F,KAAK60B,GAAcpuB;;;;;;;;;;;;;;;;;;;UC7UxB89C;;;;;IAWXzkD,YACmBg1B,GACA0vB;kBADA1vB,aACA0vB;;QAXXxkD,YAPD,IAAI0c,GACTnU,EAAYvC;;QASNhG,YAAO;;;;;;;WAiBPF,GACNi1B,GACA/f,GACA8U;QAOA,MAAM9gB,IAAMgM,EAAIhM,KACV8iC,IAAQ9rC,KAAK+hB,KAAKhf,IAAIiG,IACtBy7C,IAAe3Y,IAAQA,EAAMtlC,OAAO,GACpCk+C,IAAc1kD,KAAKwkD,GAAMxvC;QAU/B,OARAhV,KAAK+hB,OAAO/hB,KAAK+hB,KAAKjF,GAAO9T,GAAK;YAChC27C,IAAe3vC;YACfxO,MAAMk+C;YACN56B,UAAAA;YAGF9pB,KAAKwG,QAAQk+C,IAAcD,GAEpBzkD,KAAK80B,GAAayuB,GACvBxuB,GACA/rB,EAAIzB,KAAKqlB;;;;;;;WAUL9sB,GAAY6jD;QAClB,MAAM7X,IAAQ9rC,KAAK+hB,KAAKhf,IAAI4gD;QACxB7X,MACF9rC,KAAK+hB,OAAO/hB,KAAK+hB,KAAK9E,OAAO0mC,IAC7B3jD,KAAKwG,QAAQslC,EAAMtlC;;IAIvB1G,GACEi1B,GACA4uB;QAEA,MAAM7X,IAAQ9rC,KAAK+hB,KAAKhf,IAAI4gD;QAC5B,OAAOvwB,GAAmBlxB,QAAQ4pC,IAAQA,EAAM8Y,KAAgB;;IAGlE9kD,WACEi1B,GACA8uB;QAEA,IAAI3wB,IAAU5S;QAKd,OAJAujC,EAAaz9C,QAAQu9C;YACnB,MAAM7X,IAAQ9rC,KAAK+hB,KAAKhf,IAAI4gD;YAC5BzwB,IAAUA,EAAQpW,GAAO6mC,GAAa7X,IAAQA,EAAM8Y,KAAgB;YAE/DxxB,GAAmBlxB,QAAQgxB;;IAGpCpzB,GACEi1B,GACAjT,GACA6T;QAMA,IAAIzC,IAAU1S;;;gBAId,MAAMwjC,IAAS,IAAIz7C,EAAYuZ,EAAMva,KAAKyP,MAAM,MAC1C6tC,IAAW7kD,KAAK+hB,KAAKxC,GAAgBykC;QAC3C,MAAOa,EAASrlC,QAAW;YACzB,OAAMxW,KACJA,GACAnH,QAAO8iD,IAAEC,GAAa96B,UAAEA,MACtB+6B,EAASplC;YACb,KAAKqC,EAAMva,KAAK+K,EAAWtJ,EAAIzB,OAC7B;YAEEuiB,EAASrc,EAAUkoB,MAAkB,KAGrCivB,aAAyBxxC,MAAY0O,EAAMnH,QAAQiqC,OACrD1xB,IAAUA,EAAQpW,GAAO8nC,EAAc57C,KAAK47C;;QAGhD,OAAOxxB,GAAmBlxB,QAAQgxB;;IAGpCpzB,GACEi1B,GACA/c;QAEA,OAAOob,GAAmBhtB,QAAQpG,KAAK+hB,MAAO/Y,KAAqBgP,EAAEhP;;IAGvElJ,GAAgBwX;;;QAKd,OAAO,IAAIitC,GAA0BO,GAA2B9kD;;IAGlEF,GAAQs+B;QACN,OAAOhL,GAAmBlxB,QAAQlC,KAAKwG;;;;;;GAMzC+9C,SAA4C;ICvJ9CzkD;;;QAGEE,UAGI,IAAI4xB,GAAU5oB,KAAOA,EAAItH,aAK7B1B,WAAyB;;IAgBzB8pB,aAAuBjoB;QAQrB7B,KAAK+kD,KAAYljD;;IAGnBioB;QAKE,OAAO9pB,KAAK+kD;;;;;;;WASdjlD,GAAS8kD,GAA8B96B;QACrC9pB,KAAKglD,MACLhlD,KAAK8pB,WAAWA,GAChB9pB,KAAK4hB,GAAQjN,IAAIiwC,EAAc57C,KAAK47C;;;;;;;WAStC9kD,GAAYkJ,GAAkB8gB;QAC5B9pB,KAAKglD,MACDl7B,MACF9pB,KAAK8pB,WAAWA,IAElB9pB,KAAK4hB,GAAQjN,IAAI3L,GAAK;;;;;;;;;;;;WAcxBlJ,GACEi1B,GACA4uB;QAEA3jD,KAAKglD;QACL,MAAMC,IAAgBjlD,KAAK4hB,GAAQ7e,IAAI4gD;QACvC,YAAsB99C,MAAlBo/C,IACK7xB,GAAmBlxB,QAA8B+iD,KAEjDjlD,KAAKklD,GAAanwB,GAAa4uB;;;;;;;;;;;;WAe1C7jD,WACEi1B,GACA8uB;QAEA,OAAO7jD,KAAKmlD,GAAgBpwB,GAAa8uB;;;;;WAO3C/jD,MAAMi1B;QAGJ,OAFA/0B,KAAKglD,MACLhlD,KAAKolD,MAAiB,GACfplD,KAAKo6C,GAAarlB;;yDAIjBj1B;;;;;;;;;;;;;;;;;;ID2BRA,YAA6BulD;QAC3B5jD,mBAD2B4jD;;IAInBvlD,GACRi1B;QAEA,MAAMP,IAA4C;QAUlD,OATAx0B,KAAK4hB,GAAQxb,QAAQ,CAAC4C,GAAKgM;YACrBA,IACFwf,EAASluB,KACPtG,KAAKqlD,GAAcjkB,GAASrM,GAAa/f,GAAKhV,KAAK8pB,aAGrD9pB,KAAKqlD,GAAclkB,GAAYn4B;YAG5BoqB,GAAmBsB,GAAQF;;IAG1B10B,GACRi1B,GACA4uB;QAEA,OAAO3jD,KAAKqlD,GAAcjwB,GAASL,GAAa4uB;;IAGxC7jD,GACRi1B,GACA8uB;QAEA,OAAO7jD,KAAKqlD,GAAc/vB,WAAWP,GAAa8uB;;;;;;;;;;;;;;;;;;;;ME9L3CyB;IAsBXxlD,YAA6Bu9B;QAAAr9B,mBAAAq9B;;;;QAlB7Br9B,UAAkB,IAAI4xB,GAA8B4L,KAAKA,EAAE5uB;;QAGnD5O,iCAA4BqF,EAAgB2B;;QAE5ChH,uBAA4B;;QAEpCA,UAAsD;;;;;QAKtDA,UAAqB,IAAIsjC,IAEjBtjC,mBAAc,GAEtBA,UAA4Bw5B,GAAkB+rB;;IAI9CzlD,GACEs+B,GACApmB;QAGA,OADAhY,KAAKwmB,GAAQpgB,QAAQ,CAACwf,GAAGG,MAAe/N,EAAE+N,KACnCqN,GAAmBlxB;;IAG5BpC,GACEi1B;QAEA,OAAO3B,GAAmBlxB,QAAQlC,KAAKwhC;;IAGzC1hC,GACEi1B;QAEA,OAAO3B,GAAmBlxB,QAAQlC,KAAKwlD;;IAGzC1lD,GACEi1B;QAGA,OADA/0B,KAAKylD,kBAAkBzlD,KAAK0lD,GAAkBp9C,QACvC8qB,GAAmBlxB,QAAQlC,KAAKylD;;IAGzC3lD,GACEi1B,GACA4wB,GACAnkB;QAQA,OANIA,MACFxhC,KAAKwhC,4BAA4BA,IAE/BmkB,IAA8B3lD,KAAKwlD,OACrCxlD,KAAKwlD,KAAwBG;QAExBvyB,GAAmBlxB;;IAGpBpC,GAAeimB;QACrB/lB,KAAKwmB,GAAQ7R,IAAIoR,EAAWnK,QAAQmK;QACpC,MAAMlK,IAAWkK,EAAWlK;QACxBA,IAAW7b,KAAKylD,oBAClBzlD,KAAK0lD,KAAoB,IAAIlsB,GAAkB3d,IAC/C7b,KAAKylD,kBAAkB5pC,IAErBkK,EAAWhK,iBAAiB/b,KAAKwlD,OACnCxlD,KAAKwlD,KAAwBz/B,EAAWhK;;IAI5Cjc,GACEi1B,GACAhP;QAQA,OAFA/lB,KAAK4lD,GAAe7/B,IACpB/lB,KAAK6lD,eAAe,GACbzyB,GAAmBlxB;;IAG5BpC,GACEi1B,GACAhP;QAOA,OADA/lB,KAAK4lD,GAAe7/B,IACbqN,GAAmBlxB;;IAG5BpC,GACEi1B,GACAhP;QAUA,OAHA/lB,KAAKwmB,GAAQ5R,OAAOmR,EAAWnK,SAC/B5b,KAAKqkD,GAAW7E,GAAsBz5B,EAAWlK,WACjD7b,KAAK6lD,eAAe;QACbzyB,GAAmBlxB;;IAG5BpC,GACEi1B,GACA+wB,GACA1O;QAEA,IAAIruC,IAAQ;QACZ,MAAMg9C,IAA4C;QAalD,OAZA/lD,KAAKwmB,GAAQpgB,QAAQ,CAAC4C,GAAK+c;YAEvBA,EAAWhK,kBAAkB+pC,KACgB,SAA7C1O,EAAgBr0C,IAAIgjB,EAAWlK,cAE/B7b,KAAKwmB,GAAQ5R,OAAO5L,IACpB+8C,EAASz/C,KACPtG,KAAKgmD,GAA8BjxB,GAAahP,EAAWlK;YAE7D9S;YAGGqqB,GAAmBsB,GAAQqxB,GAAUz9C,KAAK,MAAMS;;IAGzDjJ,GACEi1B;QAEA,OAAO3B,GAAmBlxB,QAAQlC,KAAK6lD;;IAGzC/lD,GACEi1B,GACAnZ;QAEA,MAAMmK,IAAa/lB,KAAKwmB,GAAQzjB,IAAI6Y,MAAW;QAC/C,OAAOwX,GAAmBlxB,QAAQ6jB;;IAGpCjmB,GACEs+B,GACA7vB,GACAsN;QAGA,OADA7b,KAAKqkD,GAAW4B,GAAc13C,GAAMsN,IAC7BuX,GAAmBlxB;;IAG5BpC,GACEs+B,GACA7vB,GACAsN;QAEA7b,KAAKqkD,GAAW6B,GAAiB33C,GAAMsN;QACvC,MAAMwlB,IAAoBrhC,KAAKq9B,YAAYgE,IACrC7M,IAA4C;QAMlD,OALI6M,KACF9yB,EAAKnI,QAAQ4C;YACXwrB,EAASluB,KAAK+6B,EAAkBijB,GAAwBlmB,GAAKp1B;YAG1DoqB,GAAmBsB,GAAQF;;IAGpC10B,GACEs+B,GACAviB;QAGA,OADA7b,KAAKqkD,GAAW7E,GAAsB3jC,IAC/BuX,GAAmBlxB;;IAG5BpC,GACEs+B,GACAviB;QAEA,MAAMsqC,IAAenmD,KAAKqkD,GAAW+B,GAAgBvqC;QACrD,OAAOuX,GAAmBlxB,QAAQikD;;IAGpCrmD,GACEs+B,GACAp1B;QAEA,OAAOoqB,GAAmBlxB,QAAQlC,KAAKqkD,GAAW5E,GAAYz2C;;;;;;;;;;;;;;;;;;;;;;;;MCnLrDq9C;;;;;;;IAwBXvmD,YACEwmD;QAhBFtmD,UAAkE,IAGlEA,UAAkC,IAAIo3B,GAAe,IAErDp3B,WAAmB,GAajBA,KAAKumD,MAAW,GAChBvmD,KAAKqhC,KAAoBilB,EAAyBtmD;QAClDA,KAAK49B,KAAc,IAAI0nB,GAAkBtlD;QAGzCA,KAAK80B,KAAe,IAAImE,IACxBj5B,KAAK40B,KAAsB,IAAI2vB,GAC7BvkD,KAAK80B,IAJQ9f,KACbhV,KAAKqhC,GAAkBmlB,GAAaxxC;;IAQxClV;QACE,OAAOmC,QAAQC;;IAGjBpC;;QAGE,OADAE,KAAKumD,MAAW,GACTtkD,QAAQC;;IAGjBukD;QACE,OAAOzmD,KAAKumD;;IAGdzmD;;;IAIAA;QACE,OAAOE,KAAK80B;;IAGdh1B,GAAiBgC;QACf,IAAIk2B,IAAQh4B,KAAK0mD,GAAe5kD,EAAKu9C;QAQrC,OAPKrnB,MACHA,IAAQ,IAAIkrB,GACVljD,KAAK80B,IACL90B,KAAKqhC,KAEPrhC,KAAK0mD,GAAe5kD,EAAKu9C,OAAWrnB,IAE/BA;;IAGTl4B;QACE,OAAOE,KAAK49B;;IAGd99B;QACE,OAAOE,KAAK40B;;IAGd90B,eACEyd,GACAmlB,GACAikB;QAIAt/B,GA7FY,qBA6FM,yBAAyB9J;QAC3C,MAAM6gB,IAAM,IAAIwoB,GAAkB5mD,KAAK6mD,GAAev+C;QAEtD,OADAtI,KAAKqhC,GAAkBylB,MAChBH,EAAqBvoB,GACzB91B,KAAK6G,KACGnP,KAAKqhC,GACT0lB,GAAuB3oB,GACvB91B,KAAK,MAAM6G,IAEf63C,KACAhkD,KAAKmM,MACJivB,EAAI6oB;QACG93C;;IAIbrP,GACEi1B,GACA/rB;QAEA,OAAOoqB,GAAmB8zB,GACxBj+C,OAAOsD,OAAOvM,KAAK0mD,IAAgB5+C,IAAIkwB,KAAS,MAC9CA,EAAMynB,GAAY1qB,GAAa/rB;;;;;;;UAU1B49C;;;;;;;;;;IvBrIb9mD;QACEE,UAA2D;;IAI3DF,GAAuBstC;QACrBptC,KAAKmnD,GAAqB7gD,KAAK8mC;;IAGjCttC;QACEE,KAAKmnD,GAAqB/gD,QAAQgnC,KAAYA;;;IuB4HhDttC,YAAqB+gC;QACnBp/B,mBADmBo/B;;;;MAWVumB;IAMXtnD,YAAqCu9B;QAAAr9B,mBAAAq9B;;QAJrCr9B,UAA4C,IAAIsjC;;QAEhDtjC,UAAsD;;IAItDF,UAAeu9B;QACb,OAAO,IAAI+pB,GAAoB/pB;;IAGjCgqB;QACE,IAAKrnD,KAAKsnD,IAGR,OAAOtnD,KAAKsnD;QAFZ,MAhLqDxhD;;IAsLzDhG,GACEs+B,GACAviB,GACA7S;QAIA,OAFAhJ,KAAKunD,GAAoBzlB,GAAa94B,GAAK6S,IAC3C7b,KAAKwnD,GAAkB5yC,OAAO5L,IACvBoqB,GAAmBlxB;;IAG5BpC,GACEs+B,GACAviB,GACA7S;QAIA,OAFAhJ,KAAKunD,GAAoBxlB,GAAgB/4B,GAAK6S,IAC9C7b,KAAKwnD,GAAkB1nC,IAAI9W,IACpBoqB,GAAmBlxB;;IAG5BpC,GACEs+B,GACAp1B;QAGA,OADAhJ,KAAKwnD,GAAkB1nC,IAAI9W,IACpBoqB,GAAmBlxB;;IAG5BpC,aACEs+B,GACArY;QAEiB/lB,KAAKunD,GAAoB/H,GACxCz5B,EAAWlK,UAEJzV,QAAQ4C,KAAOhJ,KAAKwnD,GAAkB1nC,IAAI9W;QACnD,MAAMy+C,IAAQznD,KAAKq9B,YAAYQ;QAC/B,OAAO4pB,EACJ5kB,GAA2BzE,GAAKrY,EAAWlK,UAC3CvT,KAAKiG;YACJA,EAAKnI,QAAQ4C,KAAOhJ,KAAKwnD,GAAkB1nC,IAAI9W;WAEhDV,KAAK,MAAMm/C,EAAMC,GAAiBtpB,GAAKrY;;IAG5CjmB;QACEE,KAAKsnD,KAAqB,IAAIxW;;IAGhChxC,GACEs+B;;QAGA,MACMupB,IADQ3nD,KAAKq9B,YAAYM,KACJ6B;QAC3B,OAAOpM,GAAmBhtB,QACxBpG,KAAKwnD,IACJx+C,KACQhJ,KAAK4nD,GAAaxpB,GAAKp1B,GAAKV,KAAKs/C;YACjCA,KACHD,EAAaxmB,GAAYn4B;YAI/BV,KAAK,OACLtI,KAAKsnD,KAAqB,MACnBK,EAAaz2C,MAAMktB;;IAI9Bt+B,GACEs+B,GACAp1B;QAEA,OAAOhJ,KAAK4nD,GAAaxpB,GAAKp1B,GAAKV,KAAKs/C;YAClCA,IACF5nD,KAAKwnD,GAAkB5yC,OAAO5L,KAE9BhJ,KAAKwnD,GAAkB1nC,IAAI9W;;;IAKjClJ,GAAakV;;QAEX,OAAO;;IAGDlV,GACNs+B,GACAp1B;QAEA,OAAOoqB,GAAmB8zB,GAAG,EAC3B,MACE9zB,GAAmBlxB,QAAQlC,KAAKunD,GAAoB9H,GAAYz2C,KAClE,MAAMhJ,KAAKq9B,YAAYQ,KAAiB4hB,GAAYrhB,GAAKp1B,IACzD,MAAMhJ,KAAKq9B,YAAYwqB,GAAyBzpB,GAAKp1B;;;;;;;;;;;;;;;;;;;;;;;;MC1N9C8+C;IASXhoD,iBAAiBioD;QACf/nD,KAAK27C,KAAoB37C,KAAKgoD,GAAwBD,IACtD/nD,KAAKq9B,cAAcr9B,KAAKioD,GAAkBF,UACpC/nD,KAAKq9B,YAAY3d;QACvB1f,KAAKkoD,KAAcloD,KAAKmoD,GAAiCJ,IACzD/nD,KAAK6yC,KAAa7yC,KAAKooD,GAAiBL,IACxC/nD,KAAK26C,KAAc36C,KAAKqoD,GAAkBN,IAC1C/nD,KAAKi1C,KAAaj1C,KAAKsoD,GAAiBP;QACxC/nD,KAAKuoD,KAAevoD,KAAKwoD,GAAmBT,IAE5C/nD,KAAK27C,GAAkBxJ,KAAqBwF,KAC1C33C,KAAKi1C,GAAWgJ,GACdtG;QAGJ33C,KAAK26C,GAAY1F,KAAaj1C,KAAKi1C,UAE7Bj1C,KAAK6yC,GAAWnzB,eAChB1f,KAAK27C,GAAkBj8B,eACvB1f,KAAK26C,GAAYj7B;cAEjB1f,KAAK26C,GAAY8N,GAAkBzoD,KAAKi1C,GAAWuH;;IAG3D18C,GAAmBioD;QACjB,OAAO,IAAIhH,GAAa/gD,KAAKi1C;;IAG/Bn1C,GACEioD;QAEA,OAAO;;IAGTjoD,GAAiBioD;QACf,OAAO,IAAI3qB,GACTp9B,KAAKq9B,aACL,IAAImlB,IACJuF,EAAIxqB;;IAIRz9B,GAAkBioD;QAKhB,OAAO,IAAI1B,GAAkBe,GAAoBsB;;IAGnD5oD,GAAkBioD;QAChB,OAAO,IAAInV,GACT5yC,KAAK6yC,IACLkV,EAAIlX,IACJkX,EAAInuB,IACJ+d,KACE33C,KAAKi1C,GAAWgJ,GACdtG,yBAGJoQ,EAAIt4B,SAASk5B;;IAIjB7oD,GAAwBioD;QACtB,OAAO,IAAIzQ;;IAGbx3C,GAAiBioD;QACf,OAAO,IAAIrM,GACT17C,KAAK6yC,IACL7yC,KAAK26C,IACL36C,KAAK27C,IACLoM,EAAIxqB,IACJwqB,EAAInM;;IAIR97C,iBAAiB8oD;QACf,MAAM,IAAIvnD,EACRlB,EAAKW,qBA9HT;;;;;;;;;;;;;;;;;;;;;;;;;MC4BW+nD;IAmBX/oD,YACU2vB,GACAm5B,GACA7Z;;;;;;;;;IASAnV;QAXA55B,gBAAAyvB,aACAm5B,GACA5oD,mBAAA+uC,aASAnV,GAdO55B,gBAAW8wB,GAAOg4B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;WAqDnChpD,MACEipD,GACAC;QAEAhpD,KAAKipD;;;;;;;QAQL,MAAMC,IAAqB,IAAIrxB,IAQzBsxB,IAAoB,IAAItxB;;;;;;;gBAE9B,IAAIuxB,KAAc;;;;QA4BlB,OA3BAppD,KAAK+uC,YAAYsa,EAAkBvnD;YACjC,KAAKsnD,GAKH,OAJAA,KAAc,GAEd/hC,GAxHQ,mBAwHU,uBAAuBvlB,EAAK/B,MAEvCC,KAAKspD,GACVP,GACAC,GACAlnD,GACAqnD,GACAnmD,KAAKkmD,EAAmBhnD,SAASgnD,EAAmBp1B;YAEtD9zB,KAAK45B,GAAWmc,GAAiB,MACxB/1C,KAAKygD,GAAuB3+C;;;QAMzC9B,KAAK45B,GAAWc,GAAiB,MACxBwuB,EAAmBpxB,UAMrBqxB,EAAkBrxB;;kFAI3Bh4B;QAEE,OADAE,KAAKipD,MACEjpD,KAAK45B,GAAWuB,QAAQ,MACtBn7B,KAAKi1C,GAAWf;;;;;;;;;;;;;;;;;;;;;WAwBnBp0C,SACNipD,GACAC,GACAlnD,GACAqnD;QAEA;;;;YAKE,MAAMjc,UAAmBltC,KAAKyvB,SAAS85B,GAAevpD,KAAK4oD,KACrDl3C,IAAa1R,KAAKyvB,SAAS6a,GAC/BtqC,KAAK4oD,GAAa14C,KAEd2gC,ahB9IV3D,GACA6B,GACAr9B;gBAEA,OAAO,IAAI8+B,GAActD,GAAY6B,GAAar9B;agB0I5B83C,CAAatc,GAAYltC,KAAK+uC,aAAar9B;kBAEvDq3C,EAAkBU,WAAW;gBACjCC,IAAY1pD,KAAK45B;gBACjB+vB,IAAc3pD,KAAK4oD;gBACnBn5B,UAAUzvB,KAAKyvB;gBACfm6B,IAAA/Y;gBACAgZ,UAAU7pD,KAAK6pD;gBACfC,IAAahoD;gBACbioD,IArMiC;gBAsMjCC,IAAAhB;gBAGFhpD,KAAKq9B,cAAc0rB,EAAkB1rB,aACrCr9B,KAAK27C,KAAoBoN,EAAkBpN,IAC3C37C,KAAK6yC,KAAakW,EAAkBlW,IACpC7yC,KAAK26C,KAAcoO,EAAkBpO;YACrC36C,KAAKi1C,KAAa8T,EAAkB9T,IACpCj1C,KAAKkoD,KAAca,EAAkBb,IACrCloD,KAAKiqD,KAAWlB,EAAkBR;;;YAIlCvoD,KAAKq9B,YAAY6sB,GAA2BxuB;sBACpC17B,KAAKmqD;gBAGbhB,EAAkBjnD;UAClB,OAAOquB;;YAMP;;;YAHA44B,EAAkBr1B,OAAOvD,KAGpBvwB,KAAKoqD,GAAY75B,IACpB,MAAMA;YAOR,OALA85B,QAAQC,KACN,+EAEE/5B;YAEGvwB,KAAKspD,GACV,IAAIxB,IACJ;gBAAEyC,KAAS;eACXzoD,GACAqnD;;;;;;WASErpD,GAAYywB;QAClB,OAAmB,oBAAfA,EAAM5uB,OAEN4uB,EAAMhvB,SAASpB,EAAKW,uBACpByvB,EAAMhvB,SAASpB,EAAKc,kBAGE,sBAAjBupD,gBACPj6B,aAAiBi6B;;;;QAtPc,OAmQ7Bj6B,EAAMhvB,QApQgB,OAqQtBgvB,EAAMhvB;;;QAtQsB,OAyQ5BgvB,EAAMhvB;;;;;WAWJzB;QACN,IAAIE,KAAK45B,GAAW6wB,IAClB,MAAM,IAAIppD,EACRlB,EAAKW,qBACL;;IAKEhB,GAAuBgC;QAI7B,OAHA9B,KAAK45B,GAAW8wB,MAEhBrjC,GApSY,mBAoSM,uCAAuCvlB,EAAK/B;QACvDC,KAAKi1C,GAAWwL,GAAuB3+C;;qFAIhDhC;QAEE,OADAE,KAAKipD,MACEjpD,KAAK45B,GAAWuB,QAAQ,MACtBn7B,KAAKi1C,GAAWyL;;IAI3B5gD;QACE,OAAOE,KAAK45B,GAAW+wB,GAA2BjvB;;YAE5C17B,KAAKkoD,MACPloD,KAAKkoD,GAAYzT,cAGbz0C,KAAK26C,GAAY/F,YACjB50C,KAAK27C,GAAkB/G,YACvB50C,KAAKq9B,YAAYuX;;;;YAKvB50C,KAAK+uC,YAAY6b;;;;;;;WASrB9qD;QACEE,KAAKipD;QAEL,MAAMlvB,IAAW,IAAIlC;QAIrB,OAHA73B,KAAK45B,GAAWc,GAAiB,MACxB16B,KAAKi1C,GAAW4V,GAA8B9wB,KAEhDA,EAASjC;;IAGlBh4B,OACEgiB,GACA6/B,GACArqC;QAEAtX,KAAKipD;QACL,MAAM7b,IAAW,IAAIyU,GAAc//B,GAAO6/B,GAAUrqC;QAEpD,OADAtX,KAAK45B,GAAWc,GAAiB,MAAM16B,KAAKiqD,GAASxN,OAAOrP,KACrDA;;IAGTttC,GAASstC;;;QAGHptC,KAAK8qD,MAGT9qD,KAAK45B,GAAWc,GAAiB,MACxB16B,KAAKiqD,GAAS7M,GAAShQ;;IAIlCttC,SACEwyB;QAEAtyB,KAAKipD;QACL,MAAMlvB,IAAW,IAAIlC;QA4BrB,aA3BM73B,KAAK45B,GAAWuB,QAAQO;YAC5B;gBACE,MAAMvoB,UAAiBnT,KAAK6yC,GAAWkY,GAAaz4B;gBAChDnf,aAAoBC,KACtB2mB,EAAS73B,QAAQiR,KACRA,aAAoByC,KAC7BmkB,EAAS73B,QAAQ,QAEjB63B,EAASjG,OACP,IAAIzyB,EACFlB,EAAKgB,aACL;cAON,OAAOsvB;gBACP,MAAMuuB,IAAiBtiB,GACrBjM,GACA,2BAA2B6B;gBAE7ByH,EAASjG,OAAOkrB;;YAIbjlB,EAASjC;;IAGlBh4B,SAAiCgiB;QAC/B9hB,KAAKipD;QACL,MAAMlvB,IAAW,IAAIlC;QAsBrB,aArBM73B,KAAK45B,GAAWuB,QAAQO;YAC5B;gBACE,MAAM6e,UAAoBv6C,KAAK6yC,GAAW6J,GACxC56B;2CAC0B,IAEtB05B,IAAO,IAAI1D,GAAKh2B,GAAOy4B,EAAY3X,KACnC+Z,IAAiBnB,EAAKhB,GAAkBD,EAAYj4B,YACpDuf,IAAa2Z,EAAKpB,GACtBuC;6CAC4B;gBAE9B5iB,EAAS73B,QAAQ2/B,EAAoB;cACrC,OAAOpR;gBACP,MAAMuuB,IAAiBtiB,GACrBjM,GACA,4BAA4B3O;gBAE9BiY,EAASjG,OAAOkrB;;YAGbjlB,EAASjC;;IAGlBh4B,MAAMuyB;QACJryB,KAAKipD;QACL,MAAMlvB,IAAW,IAAIlC;QAIrB,OAHA73B,KAAK45B,GAAWc,GAAiB,MAC/B16B,KAAKi1C,GAAW3D,MAAMjf,GAAW0H,KAE5BA,EAASjC;;IAGlBh4B;QACE,OAAOE,KAAK4oD,GAAa14C;;IAG3BpQ,GAA2B6hD;QACzB3hD,KAAKipD,MACLjpD,KAAK45B,GAAWc,GAAiB,OAC/B16B,KAAKiqD,GAASe,GAA2BrJ,IAClC1/C,QAAQC;;IAInBpC,GAA8B6hD;;;QAGxB3hD,KAAK8qD,MAGT9qD,KAAK45B,GAAWc,GAAiB,OAC/B16B,KAAKiqD,GAASgB,GAA8BtJ,IACrC1/C,QAAQC;;IAInBgpD;;;;QAIE,OAAOlrD,KAAK45B,GAAW6wB;;IAGzB3qD,YACE86C;QAEA56C,KAAKipD;QACL,MAAMlvB,IAAW,IAAIlC;QAKrB,OAJA73B,KAAK45B,GAAWc,GAAiB,OAC/B16B,KAAKi1C,GAAW9W,eAAen+B,KAAK45B,IAAYghB,GAAgB7gB,IACzD93B,QAAQC;QAEV63B,EAASjC;;;;;;;;;;;;;;;;;;;;;;;;UChfPqzB;IAOXrrD,YAAoB6hD;QAAA3hD,gBAAA2hD;;;;;QAFZ3hD,cAAQ;;IAIhBF,KAAK+B;QACH7B,KAAKorD,GAAcprD,KAAK2hD,SAASr5C,MAAMzG;;IAGzC/B,MAAMywB;QACJvwB,KAAKorD,GAAcprD,KAAK2hD,SAASpxB,OAAOA;;IAG1CzwB;QACEE,KAAKqrD,SAAQ;;IAGPvrD,GAAiBwrD,GAA+BjN;QACjDr+C,KAAKqrD,SACR/wB,WAAW;YACJt6B,KAAKqrD,SACRC,EAAajN;WAEd;;;;;;;;;;;;;;;;;;;aCfOkN,GAAkBziD;;;;;IAChC,OAOF,SAA8BA,GAAc0iD;QAC1C,IAAmB,mBAAR1iD,KAA4B,SAARA,GAC7B,QAAO;QAGT,MAAM2iD,IAAS3iD;QACf,KAAK,MAAM4iD,KAAUF,GACnB,IAAIE,KAAUD,KAAoC,qBAAnBA,EAAOC,IACpC,QAAO;QAGX,QAAO;;;;;;;;;;;;;;;;;;;;;GAlBAC,EAAqB7iD,GAAK,EAAC,QAAQ,SAAS;;;MCYxC8iD;IACX9rD,YACmB0sC,GACAqf,GACAC,GACAC;QAHA/rD,iBAAAwsC,GACAxsC,6BAAA6rD,aACAC,GACA9rD,iBAAA+rD;;IAGnBjsD,GAAa+B;QACX,QAAQmJ,EAAUnJ;UAChB;YACE,OAAO;;UACT;YACE,OAAOA,EAAMsJ;;UACf;YACE,OAAOQ,EAAgB9J,EAAMmK,gBAAgBnK,EAAMsL;;UACrD;YACE,OAAOnN,KAAKgsD,GAAiBnqD,EAAqB;;UACpD;YACE,OAAO7B,KAAKisD,GAAuBpqD;;UACrC;YACE,OAAOA,EAAM4I;;UACf;YACE,OAAO,IAAI48B,GAAK77B,EAAoB3J,EAAiB;;UACvD;YACE,OAAO7B,KAAKksD,GAAiBrqD,EAAqB;;UACpD;YACE,OAAO7B,KAAKmsD,GAAgBtqD,EAAoB;;UAClD;YACE,OAAO7B,KAAKosD,GAAavqD,EAAiB;;UAC5C;YACE,OAAO7B,KAAKqsD,GAAcxqD,EAAe;;UAC3C;YACE,MAzDRiE;;;IA6DUhG,GAAcyK;QACpB,MAAM4E,IAAiC;QAIvC,OAHA/I,EAAQmE,EAASC,UAAU,IAAI,CAACxB,GAAKnH;YACnCsN,EAAOnG,KAAOhJ,KAAKssD,GAAazqD;YAE3BsN;;IAGDrP,GAAgB+B;QACtB,OAAO,IAAIynC,GACT39B,EAAgB9J,EAAMgK,WACtBF,EAAgB9J,EAAMiK;;IAIlBhM,GAAawM;QACnB,QAAQA,EAAWC,UAAU,IAAIzE,IAAIjG,KAAS7B,KAAKssD,GAAazqD;;IAG1D/B,GAAuB+B;QAC7B,QAAQ7B,KAAK8rD;UACX,KAAK;YACH,MAAMn7C,alExBE47C,EAAiB1qD;gBAC/B,MAAM8O,IAAgB9O,EAAM0I,SAAUC,OAA0B;gBAEhE,OAAIF,EAAkBqG,KACb47C,EAAiB57C,KAEnBA;akEkBqB47C,CAAiB1qD;YACvC,OAAqB,QAAjB8O,IACK,OAEF3Q,KAAKssD,GAAa37C;;UAC3B,KAAK;YACH,OAAO3Q,KAAKgsD,GAAiBthD,EAAkB7I;;UACjD;YACE,OAAO;;;IAIL/B,GAAiB+B;QACvB,MAAM2qD,IAAkB5hD,EAAmB/I,IACrCyD,IAAY,IAAIlB,EACpBooD,EAAgBnoD,SAChBmoD,EAAgB3hD;QAElB,OAAI7K,KAAK6rD,wBACAvmD,IAEAA,EAAUmnD;;IAIb3sD,GAAiB6B;QACvB,MAAM+qD,IAAevlD,EAAaqB,EAAW7G;QAvFrC0B,GAyFN8lB,GAAoBujC;QAGtB,MAAMx8C,IAAa,IAAIwhB,GAAWg7B,EAAa3pD,IAAI,IAAI2pD,EAAa3pD,IAAI,KAClEiG,IAAM,IAAIT,EAAYmkD,EAAajkD,EAAS;QAclD,OAZKyH,EAAW1K,QAAQxF,KAAKwsC,UAAUC;;QAErCjwB,GACE,YAAYxT,2BACV,4CACA,GAAGkH,EAAWC,aAAaD,EAAWE,4BACtC,iEACA,aAAapQ,KAAKwsC,UAAUC,GAAYt8B,aAAanQ,KAAKwsC,UAAUC,GAAYr8B,eAChF;QAIC,IAAIi8B,GAAkBrjC,GAAKhJ,KAAKwsC,WAAWxsC,KAAK+rD;;;;;;;;;;;;;;;;;;;;uBC3D3D;MAWaY,KAAuBhwB,GAAUQ;;;;;;;AAyB9C,MAAMyvB;IAmBJ9sD,YAAYgqC;;QACV,SAAsBjkC,MAAlBikC,EAASvY,MAAoB;YAC/B,SAAqB1rB,MAAjBikC,EAAStY,KACX,MAAM,IAAInwB,EACRlB,EAAKI,kBACL;YAGJP,KAAKuxB,OA/DU,4BAgEfvxB,KAAKwxB,OA/DS;eAiEd2T,GAAkB,YAAY,oBAAoB,QAAQ2E,EAASvY,OACnEvxB,KAAKuxB,OAAOuY,EAASvY,MAErB8T,GAA0B,YAAY,WAAW,OAAOyE,EAAStY;QACjExxB,KAAKwxB,oBAAMsY,EAAStY;QA0DtB,IAxDAqV,GAAoB,YAAYiD,GAAU,EACxC,QACA,OACA,eACA,yBACA,kBACA,gCACA;QAGFzE,GACE,YACA,UACA,eACAyE,EAASiF,cAEX/uC,KAAK+uC,cAAcjF,EAASiF;QAE5B1J,GACE,YACA,WACA,yBACAyE,EAAS+hB,wBAGXxmB,GACE,YACA,WACA,6BACAyE,EAAShB;;;SAK4B,MAAnCgB,EAAS+hB,wBACXrvC,GACE,6FAG0C,MAAnCstB,EAAS+hB,yBAClBrvC,GACE;QAIJxc,KAAK6rD,sCACH/hB,EAAS+hB;QACX7rD,KAAK8oC,0CACHgB,EAAShB;QAEXzD,GACE,YACA,UACA,kBACAyE,EAAS+iB,sBAEqBhnD,MAA5BikC,EAAS+iB,gBACX7sD,KAAK6sD,iBAAiBlwB,GAAUO,SAC3B;YACL,IACE4M,EAAS+iB,mBAAmBF,MAC5B7iB,EAAS+iB,iBAAiBlwB,GAAUmwB,IAEpC,MAAM,IAAIzrD,EACRlB,EAAKI,kBACL,mCAAmCo8B,GAAUmwB;YAG/C9sD,KAAK6sD,iBAAiB/iB,EAAS+iB;;QAInCxnB,GACE,YACA,WACA,gCACAyE,EAASijB;QAEX/sD,KAAKyxB,iCACHqY,EAASijB;;IAGbjtD,QAAQkF;QACN,OACEhF,KAAKuxB,SAASvsB,EAAMusB,QACpBvxB,KAAKwxB,QAAQxsB,EAAMwsB,OACnBxxB,KAAK6rD,0BAA0B7mD,EAAM6mD,yBACrC7rD,KAAK+uC,gBAAgB/pC,EAAM+pC,eAC3B/uC,KAAK6sD,mBAAmB7nD,EAAM6nD,kBAC9B7sD,KAAKyxB,qBAAqBzsB,EAAMysB,oBAChCzxB,KAAK8oC,8BAA8B9jC,EAAM8jC;;;;;;UAQlCkkB;;;;IA4BXltD,YACEmtD,GACA5qD,GACA0mD,IAAuC,IAAIjB;QAE3C,IAzBF9nD,UAAoD;;;QAapDA,UAAkB,IAAI26B,IAwQtB36B,gBAAW;YACT4U,QAAQ8mB;;;gBAGN17B,KAAKktD,YACCltD,KAAKmtD,GAAkBhD;;WAjQyB,mBAA5C8C,EAAgC31C,SAAsB;;;YAGhE,MAAM81C,IAAMH;YACZjtD,KAAKqtD,KAAeD,GACpBptD,KAAKysC,KAAcugB,GAAUM,GAAkBF,IAC/CptD,KAAKutD,KAAkBH,EAAIzrD,MAC3B3B,KAAKwtD,KAAe,IAAIprD,EAA4BC;eAC/C;YACL,MAAMorD,IAAWR;YACjB,KAAKQ,EAASt9C,WACZ,MAAM,IAAI9O,EACRlB,EAAKI,kBACL;YAIJP,KAAKysC,KAAc,IAAI/a,GAAW+7B,EAASt9C,WAAWs9C,EAASr9C;;YAE/DpQ,KAAKutD,KAAkB,aACvBvtD,KAAKwtD,KAAe,IAAIxrD;;QAG1BhC,KAAK0tD,KAAqB3E,GAC1B/oD,KAAK2tD,KAAY,IAAIf,GAAkB;;IAGzCgB;QAYE,OAPK5tD,KAAK6tD;;QAER7tD,KAAK6tD,KAAkB,IAAIxjB,GACzBrqC,KAAKysC,IACLzsC,KAAK2tD,GAAU7kB,6BAGZ9oC,KAAK6tD;;IAGd/tD,SAASguD;QACPtpB,GAA0B,sBAAsBgD,WAAW,IAC3D1C,GAAgB,sBAAsB,UAAU,GAAGgpB;QAEnD,MAAMC,IAAc,IAAInB,GAAkBkB;QAC1C,IAAI9tD,KAAKmtD,OAAqBntD,KAAK2tD,GAAUnoD,QAAQuoD,IACnD,MAAM,IAAI1sD,EACRlB,EAAKW,qBACL;QAMJd,KAAK2tD,KAAYI,QACeloD,MAA5BkoD,EAAYhf,gBACd/uC,KAAKwtD,c3EdTze;YAEA,KAAKA,GACH,OAAO,IAAI/sC;YAGb,QAAQ+sC,EAAYptB;cAClB,KAAK;gBACH,MAAMqsC,IAASjf,EAAYif;;gCAW3B,OATA3qD,KAEsB,mBAAX2qD,KACI,SAAXA,MACAA,EAAa,SACbA,EAAa,KAAmC;gBAI7C,IAAI7pD,EACT6pD,GACAjf,EAAYnrC,KAAgB;;cAGhC,KAAK;gBACH,OAAOmrC,EAAYif;;cAErB;gBACE,MAAM,IAAI3sD,EACRlB,EAAKI,kBACL;;;;;;;;;;;;;;;;;;;qF2EhBkB0tD;SAAwBF,EAAYhf;;IAI5DjvC;QAEE,OADAE,KAAKktD,MACEltD,KAAKmtD,GAAkBjZ;;IAGhCp0C;QAEE,OADAE,KAAKktD,MACEltD,KAAKmtD,GAAkBzM;;IAGhC5gD,kBAAkBgqC;;QAChB,IAAI9pC,KAAKmtD,IACP,MAAM,IAAI9rD,EACRlB,EAAKW,qBACL;QAMJ,IAAIotD,KAAkB;QActB,OAZIpkB,WAC8CjkC,MAA5CikC,EAASqkB,kCACX3xC,GACE;QAGJ0xC,gCACEpkB,EAASokB,uCACTpkB,EAASqkB;QAINnuD,KAAKouD,GAAgBpuD,KAAK0tD,IAAoB;YACnDnD,KAAS;YACTsC,gBAAgB7sD,KAAK2tD,GAAUd;YAC/BqB,iBAAAA;;;IAIJpuD;QACE,SAC4B+F,MAA1B7F,KAAKmtD,OACJntD,KAAKmtD,GAAiBrC,IAEvB,MAAM,IAAIzpD,EACRlB,EAAKW,qBACL;QAIJ,MAAMi5B,IAAW,IAAIlC;QAUrB,OATA73B,KAAKquD,GAAOC,GAAkC5yB;YAC5C;gBACE,MAAMktB,IAAe5oD,KAAKuuD;sBACpBvuD,KAAK0tD,GAAmBc,iBAAiB5F,IAC/C7uB,EAAS73B;cACT,OAAOuuB;gBACPsJ,EAASjG,OAAOrD;;YAGbsJ,EAASjC;;IAGlBh4B;QAEE,OADCE,KAAKotD,IAAqBqB,uBAAuB,cAC3CzuD,KAAKkB,SAAS0T;;IAGvB85C;QAEE,OADA1uD,KAAKktD,MACEltD,KAAKmtD,GAAkBrC;;IAGhChrD;QAEE,OADAE,KAAKktD,MACEltD,KAAKmtD,GAAkBwB;;IAKhC7uD,kBAAkB8uD;QAGhB,IAFA5uD,KAAKktD,MAED3B,GAAkBqD,IACpB,OAAO5uD,KAAK6uD,GAA0BD;QACjC;YACL9pB,GAAgB,+BAA+B,YAAY,GAAG8pB;YAC9D,MAAMjN,IAAkC;gBACtCr5C,MAAMsmD;;YAER,OAAO5uD,KAAK6uD,GAA0BlN;;;IAIlC7hD,GACN6hD;QAEA,MAGMmN,IAAgB,IAAI3D,GAAoB;YAC5C7iD,MAAM;gBACAq5C,EAASr5C,QACXq5C,EAASr5C;;YAGbioB,OATkB4D;gBAClB,MArbsBruB;;;QAgcxB,OADA9F,KAAKmtD,GAAkBnC,GAA2B8D,IAC3C;YACLA,EAAcC,MACd/uD,KAAKmtD,GAAkBlC,GAA8B6D;;;IAIzDhvD;QAQE,OAPKE,KAAKmtD;;;QAGRntD,KAAKouD,GAAgB,IAAItG,IAA2B;YAClDyC,KAAS;YAGNvqD,KAAKmtD;;IAGNrtD;QACN,OAAO,IAAIuxB,GACTrxB,KAAKysC,IACLzsC,KAAKutD,IACLvtD,KAAK2tD,GAAUp8B,MACfvxB,KAAK2tD,GAAUn8B,KACfxxB,KAAK2tD,GAAUl8B;;IAIX3xB,GACNipD,GACAC;QASA,MAAMJ,IAAe5oD,KAAKuuD;QAS1B,OAPAvuD,KAAKmtD,KAAmB,IAAItE,GAC1Bp/C,GAAgBC,MAChBk/C,GACA5oD,KAAKwtD,IACLxtD,KAAKquD,KAGAruD,KAAKmtD,GAAiBztC,MAAMqpC,GAAmBC;;IAGhDlpD,UAAyBstD;QAC/B,IA6+DctkD,IA7+DAskD,EAAI91C,SA6+DStO,IA7+DA,cA8+DtBC,OAAOC,UAAUC,eAAeC,KAAKN,GAAKE,IA7+D7C,MAAM,IAAI3H,EACRlB,EAAKI,kBACL;QA0+DR,IAAkBuI,GAAaE;;;;;;;;;;;;;;;;;;;;;;;;;;;WAt+D3B,MAAMmH,IAAYi9C,EAAI91C,QAAQnH;QAC9B,KAAKA,KAAkC,mBAAdA,GACvB,MAAM,IAAI9O,EACRlB,EAAKI,kBACL;QAGJ,OAAO,IAAImxB,GAAWvhB;;IAGxBi9C;QACE,KAAKptD,KAAKqtD,IACR,MAAM,IAAIhsD,EACRlB,EAAKW,qBACL;QAIJ,OAAOd,KAAKqtD;;IAYdvtD,WAAWkvD;QAIT,OAHAxqB,GAA0B,wBAAwBgD,WAAW,IAC7D1C,GAAgB,wBAAwB,oBAAoB,GAAGkqB;QAC/DhvD,KAAKktD,MACE,IAAI+B,GAAoB9nD,EAAaqB,EAAWwmD,IAAahvD;;IAGtEF,IAAIkvD;QAIF,OAHAxqB,GAA0B,iBAAiBgD,WAAW,IACtD1C,GAAgB,iBAAiB,oBAAoB,GAAGkqB;QACxDhvD,KAAKktD,MACE7gB,GAAkB6iB,GAAQ/nD,EAAaqB,EAAWwmD,IAAahvD;;IAGxEF,gBAAgB4I;QAQd,IAPA87B,GAA0B,6BAA6BgD,WAAW,IAClE1C,GACE,6BACA,oBACA,GACAp8B;QAEEA,EAAalB,QAAQ,QAAQ,GAC/B,MAAM,IAAInG,EACRlB,EAAKI,kBACL,0BAA0BmI,2BACxB;QAIN,OADA1I,KAAKktD,MACE,IAAIh1C,GACT,IAAIi3C,GAAchoD,EAAasP,GAAY/N,IAC3C1I;;IAIJF,eACE86C;QAIA,OAFApW,GAA0B,4BAA4BgD,WAAW,IACjE1C,GAAgB,4BAA4B,YAAY,GAAG8V;QACpD56C,KAAKktD,KAAyBn4B,YAClCA,KACQ6lB,EAAe,IAAIhK,GAAY5wC,MAAM+0B;;IAKlDj1B;QAGE,OAFAE,KAAKktD,MAEE,IAAIkC,GAAWpvD;;IAGxB6vB;QACE,QAAQD;UACN,KAAKK,EAASC;YACZ,OAAO;;UACT,KAAKD,EAASo/B;YACZ,OAAO;;UACT;;YAEE,OAAO;;;IAIbvvD,mBAAmBwvD;QAGjB,QAFA9qB,GAA0B,yBAAyBgD,WAAW,IAC9D1C,GAAgB,yBAAyB,oBAAoB,GAAGwqB;QACxDA;UACN,KAAK;YACHx/B,GAAYG,EAASC;YACrB;;UACF,KAAK;YACHJ,GAAYG,EAASK;YACrB;;UACF,KAAK;YACHR,GAAYG,EAASo/B;YACrB;;UACF;YACE,MAAM,IAAIhuD,EACRlB,EAAKI,kBACL,wBAAwB+uD;;;;;IAOhCxvD;QACE,OAAOE,KAAK2tD,GAAU9B;;;;;;UAObjb;IACX9wC,YACUyvD,GACAC;kBADAD,aACAC;;IAGV1vD,IACE2vD;QAEAjrB,GAA0B,mBAAmBgD,WAAW;QACxD,MAAM7D,IAAM+rB,GACV,mBACAD,GACAzvD,KAAKuvD;QAEP,OAAOvvD,KAAKwvD,GACTG,GAAO,EAAChsB,EAAI+I,MACZ1pC,KAAM+e;YACL,KAAKA,KAAwB,MAAhBA,EAAKnc,QAChB,OA5oBkBE;YA8oBpB,MAAMkP,IAAM+M,EAAK;YACjB,IAAI/M,aAAeY,IACjB,OAAO,IAAIg6C,GACT5vD,KAAKuvD,IACL5rB,EAAI+I,IACJ;8BACiB;qCACO,GACxB/I,EAAIksB;YAED,IAAI76C,aAAe5B,IACxB,OAAO,IAAIw8C,GACT5vD,KAAKuvD,IACL5rB,EAAI+I,IACJ13B;8BACiB;qCACO,GACxB2uB,EAAIksB;YAGN,MAlqBkB/pD;;;IAyqB1BhG,IACE2vD,GACA5tD,GACAyV;QAEAstB,GAA4B,mBAAmB4C,WAAW,GAAG;QAC7D,MAAM7D,IAAM+rB,GACV,mBACAD,GACAzvD,KAAKuvD;QAEPj4C,IAAUw4C,GAAmB,mBAAmBx4C;QAChD,OAAOy4C,GAAgBzrB,KAAgB0rB,GACrCrsB,EAAIksB,IACJhuD,GACA,oBAEIouD,IACJ34C,EAAQ44C,SAAS54C,EAAQ64C,cACrBnwD,KAAKuvD,GAAWa,GAAYC,GAC1B/rB,GACAyrB,GACAz4C,EAAQ64C,eAEVnwD,KAAKuvD,GAAWa,GAAYE,GAC1BhsB,GACAyrB;QAGR,OADA/vD,KAAKwvD,GAAa76C,IAAIgvB,EAAI+I,IAAMujB,IACzBjwD;;IAaTF,OACE2vD,GACAc,GACA1uD,MACG0pC;QAEH,IAAI5H,GACAssB;QAgCJ,OA7B+B,mBAAtBM,KACPA,aAA6B1lB,MAE7BnG,GAA4B,sBAAsB8C,WAAW;QAC7D7D,IAAM+rB,GACJ,sBACAD,GACAzvD,KAAKuvD,KAEPU,IAASjwD,KAAKuvD,GAAWa,GAAYI,GACnC,sBACAD,GACA1uD,GACA0pC,OAGF/G,GAA0B,sBAAsBgD,WAAW;QAC3D7D,IAAM+rB,GACJ,sBACAD,GACAzvD,KAAKuvD,KAEPU,IAASjwD,KAAKuvD,GAAWa,GAAYK,GACnC,sBACAF;QAIJvwD,KAAKwvD,GAAa3kC,OAAO8Y,EAAI+I,IAAMujB,IAC5BjwD;;IAGTF,OAAO2vD;QACLjrB,GAA0B,sBAAsBgD,WAAW;QAC3D,MAAM7D,IAAM+rB,GACV,sBACAD,GACAzvD,KAAKuvD;QAGP,OADAvvD,KAAKwvD,GAAa56C,OAAO+uB,EAAI+I,KACtB1sC;;;;MAIEovD;IAIXtvD,YAAoByvD;kBAAAA,GAHpBvvD,UAAqB,IACrBA,WAAqB;;IAIrBF,IACE2vD,GACA5tD,GACAyV;QAEAstB,GAA4B,kBAAkB4C,WAAW,GAAG,IAC5DxnC,KAAK0wD;QACL,MAAM/sB,IAAM+rB,GACV,kBACAD,GACAzvD,KAAKuvD;QAEPj4C,IAAUw4C,GAAmB,kBAAkBx4C;QAC/C,OAAOy4C,GAAgBzrB,KAAgB0rB,GACrCrsB,EAAIksB,IACJhuD,GACA,mBAEIouD,IACJ34C,EAAQ44C,SAAS54C,EAAQ64C,cACrBnwD,KAAKuvD,GAAWa,GAAYC,GAC1B/rB,GACAyrB,GACAz4C,EAAQ64C,eAEVnwD,KAAKuvD,GAAWa,GAAYE,GAC1BhsB,GACAyrB;QAKR,OAHA/vD,KAAK2wD,KAAa3wD,KAAK2wD,GAAWz3C,OAChC+2C,EAAO1e,GAAY5N,EAAI+I,IAAM35B,GAAawY,QAErCvrB;;IAaTF,OACE2vD,GACAc,GACA1uD,MACG0pC;QAIH,IAAI5H,GACAssB;QAkCJ,OArCAjwD,KAAK0wD,MAM0B,mBAAtBH,KACPA,aAA6B1lB,MAE7BnG,GAA4B,qBAAqB8C,WAAW;QAC5D7D,IAAM+rB,GACJ,qBACAD,GACAzvD,KAAKuvD,KAEPU,IAASjwD,KAAKuvD,GAAWa,GAAYI,GACnC,qBACAD,GACA1uD,GACA0pC,OAGF/G,GAA0B,qBAAqBgD,WAAW;QAC1D7D,IAAM+rB,GACJ,qBACAD,GACAzvD,KAAKuvD,KAEPU,IAASjwD,KAAKuvD,GAAWa,GAAYK,GACnC,qBACAF;QAIJvwD,KAAK2wD,KAAa3wD,KAAK2wD,GAAWz3C,OAChC+2C,EAAO1e,GAAY5N,EAAI+I,IAAM35B,GAAaE,QAAO,MAE5CjT;;IAGTF,OAAO2vD;QACLjrB,GAA0B,qBAAqBgD,WAAW,IAC1DxnC,KAAK0wD;QACL,MAAM/sB,IAAM+rB,GACV,qBACAD,GACAzvD,KAAKuvD;QAKP,OAHAvvD,KAAK2wD,KAAa3wD,KAAK2wD,GAAWz3C,OAChC,IAAIvD,GAAeguB,EAAI+I,IAAM35B,GAAawY,QAErCvrB;;IAGTF;QAGE,OAFAE,KAAK0wD,MACL1wD,KAAK4wD,MAAa,GACd5wD,KAAK2wD,GAAW/qD,SAAS,IACpB5F,KAAKuvD,GAAWrC,KAAyB5b,MAAMtxC,KAAK2wD,MAGtD1uD,QAAQC;;IAGTpC;QACN,IAAIE,KAAK4wD,IACP,MAAM,IAAIvvD,EACRlB,EAAKW,qBACL;;;;;;UAUKurC;IAIXvsC,YACS4sC,GACEF,GACAqjB;kBAFFnjB,GACE1sC,iBAAAwsC,aACAqjB,GAET7vD,KAAKmtD,KAAmBntD,KAAKwsC,UAAU0gB;;IAGzCptD,UACEyH,GACAilC,GACAuf;QAEA,IAAIxkD,EAAK3B,SAAS,KAAM,GACtB,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,+FAEE,GAAGgH,EAAKD,WAAyBC,EAAK3B;QAG5C,OAAO,IAAIymC,GAAkB,IAAI9jC,EAAYhB,IAAOilC,GAAWuf;;IAGjEj6B;QACE,OAAO9xB,KAAK0sC,GAAKnlC,KAAKwO;;IAGxB0W;QACE,OAAO,IAAIwiC,GACTjvD,KAAK0sC,GAAKnlC,KAAKqlB,KACf5sB,KAAKwsC,WACLxsC,KAAK6vD;;IAITtoD;QACE,OAAOvH,KAAK0sC,GAAKnlC,KAAKD;;IAGxBxH,WACEkvD;QASA,IAPAxqB,GAA0B,gCAAgCgD,WAAW,IACrE1C,GACE,gCACA,oBACA,GACAkqB;SAEGA,GACH,MAAM,IAAI3tD,EACRlB,EAAKI,kBACL;QAGJ,MAAMgH,IAAOJ,EAAaqB,EAAWwmD;QACrC,OAAO,IAAIC,GAAoBjvD,KAAK0sC,GAAKnlC,KAAKyP,MAAMzP,IAAOvH,KAAKwsC;;IAGlE1sC,QAAQkF;QACN,MAAMA,aAAiBqnC,KACrB,MAAMtF,GAAkB,WAAW,qBAAqB,GAAG/hC;QAE7D,OACEhF,KAAKwsC,cAAcxnC,EAAMwnC,aACzBxsC,KAAK0sC,GAAKlnC,QAAQR,EAAM0nC,OACxB1sC,KAAK6vD,OAAe7qD,EAAM6qD;;IAQ9B/vD,IAAI+B,GAAUyV;QACZstB,GAA4B,yBAAyB4C,WAAW,GAAG,IACnElwB,IAAUw4C,GAAmB,yBAAyBx4C;QACtD,OAAOy4C,GAAgBzrB,KAAgB0rB,GACrChwD,KAAK6vD,IACLhuD,GACA,0BAEIouD,IACJ34C,EAAQ44C,SAAS54C,EAAQ64C,cACrBnwD,KAAKwsC,UAAU4jB,GAAYC,GACzB/rB,GACAyrB,GACAz4C,EAAQ64C,eAEVnwD,KAAKwsC,UAAU4jB,GAAYE,GAAahsB,GAAcyrB;QAC5D,OAAO/vD,KAAKmtD,GAAiB7b,MAC3B2e,EAAO1e,GAAYvxC,KAAK0sC,IAAM35B,GAAawY;;IAU/CzrB,OACEywD,GACA1uD,MACG0pC;QAEH,IAAI0kB;QAqBJ,OAlB+B,mBAAtBM,KACPA,aAA6B1lB,MAE7BnG,GAA4B,4BAA4B8C,WAAW;QACnEyoB,IAASjwD,KAAKwsC,UAAU4jB,GAAYI,GAClC,4BACAD,GACA1uD,GACA0pC,OAGF/G,GAA0B,4BAA4BgD,WAAW;QACjEyoB,IAASjwD,KAAKwsC,UAAU4jB,GAAYK,GAClC,4BACAF,KAIGvwD,KAAKmtD,GAAiB7b,MAC3B2e,EAAO1e,GAAYvxC,KAAK0sC,IAAM35B,GAAaE,QAAO;;IAItDnT;QAEE,OADA0kC,GAA0B,4BAA4BgD,WAAW,IAC1DxnC,KAAKmtD,GAAiB7b,MAAM,EACjC,IAAI37B,GAAe3V,KAAK0sC,IAAM35B,GAAawY;;IAuB/CzrB,cAAcqwB;QACZyU,GACE,gCACA4C,WACA,GACA;QAEF,IAGIma,GAHArqC,IAA2C;YAC7C0qC,yBAAwB;WAGtB6O,IAAU;QAEa,mBAAlB1gC,EAAK0gC,MACXtF,GAAkBp7B,EAAK0gC,QAExBv5C,IAAU6Y,EAAK0gC,IACfhqB,GAAoB,gCAAgCvvB,GAAS,EAC3D;QAEF+tB,GACE,gCACA,WACA,0BACA/tB,EAAQ0qC;QAEV6O;QAGF,MAAMC,IAAkB;YACtB9O,wBAAwB1qC,EAAQ0qC;;QAgClC,OA7BIuJ,GAAkBp7B,EAAK0gC,MACzBlP,IAAWxxB,EAAK0gC,MAIhB/rB,GACE,gCACA,YACA+rB,GACA1gC,EAAK0gC;QAEP3rB,GACE,gCACA,YACA2rB,IAAU,GACV1gC,EAAK0gC,IAAU,KAEjB3rB,GACE,gCACA,YACA2rB,IAAU,GACV1gC,EAAK0gC,IAAU;QAEjBlP,IAAW;YACTr5C,MAAM6nB,EAAK0gC;YACXtgC,OAAOJ,EAAK0gC,IAAU;YACtBE,UAAU5gC,EAAK0gC,IAAU;YAGtB7wD,KAAKgxD,GAAmBF,GAAiBnP;;IAG1C7hD,GACNwX,GACAqqC;QAEA,IAAIsP,IAAc98B;YAChBk2B,QAAQ95B,MAAM,iCAAiC4D;;QAE7CwtB,EAASpxB,UACX0gC,IAAatP,EAASpxB,MAAMyJ,KAAK2nB;QAGnC,MAAMmN,IAAgB,IAAI3D,GAA4B;YACpD7iD,MAAM6mC;gBACJ,IAAIwS,EAASr5C,MAAM;oBAKjB,MAAM0M,IAAMm6B,EAASptB,KAAKhf,IAAI/C,KAAK0sC;oBAEnCiV,EAASr5C,KACP,IAAIsnD,GACF5vD,KAAKwsC,WACLxsC,KAAK0sC,IACL13B,GACAm6B,EAAShtB,WACTgtB,EAAS33B,kBACTxX,KAAK6vD;;;YAKbt/B,OAAO0gC;YAEHC,IAAmBlxD,KAAKmtD,GAAiB1Q,OAC7C0S,GAAc7iC,GAAOtsB,KAAK0sC,GAAKnlC,OAC/BunD,GACAx3C;QAGF,OAAO;YACLw3C,EAAcC,MACd/uD,KAAKmtD,GAAiB/P,GAAS8T;;;IAInCpxD,IAAIwX;QAGF,OAFAstB,GAA4B,yBAAyB4C,WAAW,GAAG,IACnE2pB,GAAmB,yBAAyB75C;QACrC,IAAIrV,QACT,CAACC,GAAkD4xB;YAC7Cxc,KAA8B,YAAnBA,EAAQymC,SACrB/9C,KAAKwsC,UACF0gB,KACAkE,GAA0BpxD,KAAK0sC,IAC/B1pC,KAAKgS;gBACJ9S,EACE,IAAI0tD,GACF5vD,KAAKwsC,WACLxsC,KAAK0sC,IACL13B;gCACe,GACfA,aAAe5B,MAAW4B,EAAIuC,IAC9BvX,KAAK6vD;eAGR/7B,KAEL9zB,KAAKqxD,GAAuBnvD,GAAS4xB,GAAQxc;;;IAM7CxX,GACNoC,GACA4xB,GACAxc;QAEA,MAAM8lC,IAAWp9C,KAAKgxD,GACpB;YACEhP,yBAAwB;YACxBsP,KAAuB;WAEzB;YACEhpD,MAAOy5C;;;gBAGL3E,MAEK2E,EAAK9uC,UAAU8uC,EAAKwP,SAASpvC;;;;;;;;gBAQhC2R,EACE,IAAIzyB,EACFlB,EAAKgB,aACL,4DAIJ4gD,EAAK9uC,UACL8uC,EAAKwP,SAASpvC,aACd7K,KACmB,aAAnBA,EAAQymC,SAERjqB,EACE,IAAIzyB,EACFlB,EAAKgB,aACL,gLAOJe,EAAQ6/C;;YAGZxxB,OAAOuD;;;IAKbh0B,cACEisD;QAEA,OAAO,IAAI1f,GAAqBrsC,KAAK0sC,IAAM1sC,KAAKwsC,WAAWuf;;;;AAI/D,MAAMyF;IACJ1xD,YACW0X,GACA2K;QADAniB,wBAAAwX,GACAxX,iBAAAmiB;;IAGXriB,QAAQkF;QACN,OACEhF,KAAKwX,qBAAqBxS,EAAMwS,oBAChCxX,KAAKmiB,cAAcnd,EAAMmd;;;;MAWlBytC;IAEX9vD,YACUyvD,GACA7iB,GACD+kB,GACCC,GACAC,GACS9B;kBALTN,aACA7iB,aACD+kB,aACCC,aACAC,aACS9B;;IAGnB/vD,KAAKwX;QAGH,IAFAstB,GAA4B,yBAAyB4C,WAAW,GAAG,IACnElwB,IAAUs6C,GAAwB,yBAAyBt6C;QACtDtX,KAAKyxD,IAEH;;;YAGL,IAAIzxD,KAAK6vD,IAAY;gBACnB,MAAM1gB,IAAW,IAAI0iB,GACnB7xD,KAAKuvD,IACLvvD,KAAK0sC,IACL1sC,KAAKyxD,IACLzxD,KAAK0xD,IACL1xD,KAAK2xD;gBAEP,OAAO3xD,KAAK6vD,GAAWiC,cAAc3iB,GAAU73B;;YAQ/C,OANuB,IAAIs0C,GACzB5rD,KAAKuvD,IACLvvD,KAAKuvD,GAAWwC,MAChBz6C,EAAQ06C;kCACSnsD,GAEGymD,GAAatsD,KAAKyxD,GAAU/nC;;;IAKxD5pB,IACEsS,GACAkF;QAIA,IAFAstB,GAA4B,wBAAwB4C,WAAW,GAAG,IAClElwB,IAAUs6C,GAAwB,wBAAwBt6C;QACtDtX,KAAKyxD,IAAW;YAClB,MAAM5vD,IAAQ7B,KAAKyxD,GAChBz9C,OACAtB,MAAM84B,GAAsB,wBAAwBp5B;YACvD,IAAc,SAAVvQ,GAAgB;gBAOlB,OANuB,IAAI+pD,GACzB5rD,KAAKuvD,IACLvvD,KAAKuvD,GAAWwC,MAChBz6C,EAAQ06C,kBACRhyD,KAAK6vD,IAEevD,GAAazqD;;;;IAMzCiwB;QACE,OAAO9xB,KAAK0sC,GAAKnlC,KAAKwO;;IAGxB4tB;QACE,OAAO,IAAI0I,GACTrsC,KAAK0sC,IACL1sC,KAAKuvD,IACLvvD,KAAK6vD;;IAIT58C;QACE,OAA0B,SAAnBjT,KAAKyxD;;IAGdF;QACE,OAAO,IAAIC,GAAiBxxD,KAAK2xD,IAAmB3xD,KAAK0xD;;IAG3D5xD,QAAQkF;QACN,MAAMA,aAAiB4qD,KACrB,MAAM7oB,GAAkB,WAAW,oBAAoB,GAAG/hC;QAE5D,OACEhF,KAAKuvD,OAAevqD,EAAMuqD,MAC1BvvD,KAAK0xD,OAAe1sD,EAAM0sD,MAC1B1xD,KAAK0sC,GAAKlnC,QAAQR,EAAM0nC,QACJ,SAAnB1sC,KAAKyxD,KACkB,SAApBzsD,EAAMysD,KACNzxD,KAAKyxD,GAAUjsD,QAAQR,EAAMysD,QACjCzxD,KAAK6vD,OAAe7qD,EAAM6qD;;;;MAKnBgC,WACHjC;IAER9vD,KAAKwX;QAMH,OALa7V,MAAMuS,KAAKsD;;;;MASfY;IACXpY,YACSmyD,GACEzlB,GACUqjB;kBAFZoC,GACEjyD,iBAAAwsC,aACUqjB;;IAGrB/vD,MACE4S,GACAw/C,GACArwD;QAEA2iC,GAA0B,eAAegD,WAAW,IACpDZ,GAAgB,eAAe,GAAG/kC;;QAGlC,MAUMoY,a3BloCRqqB,GACA6tB,GACA73C,GACAyqB;YAEA,KAAKotB,EAAM9gD,KAAKC,KAAWA,MAAYyzB,IACrC,MAAM,IAAI1jC,EACRlB,EAAKI,kBACL,iBAAiBmlC,GAAiBX,6BAChC,GAAGT,eAA0BW,GAAQ3qB,6BACrC,WAAW63C,EAAM9qD,KAAK;YAG5B,OAAO09B;S2BqnCMqtB,CAAmB,eAVH,gPAUsC,GAAGF;QAEpE,IAAIG;QACJ,MAAMjgD,IAAYo5B,GAAsB,eAAe94B;QACvD,IAAIN,EAAUuG,KAAc;YAC1B,8CACEsB,uDACAA,GAEA,MAAM,IAAI5Y,EACRlB,EAAKI,kBACL,qCAAqC0Z,QACnC;YAEC,sBAAIA,GAAoB;gBAC7Bja,KAAKsyD,GAAkCzwD,GAAOoY;gBAC9C,MAAMs4C,IAA6B;gBACnC,KAAK,MAAMjmD,KAAczK,GACvB0wD,EAAcjsD,KAAKtG,KAAKwyD,GAAqBlmD;gBAE/C+lD,IAAa;oBAAE/lD,YAAY;wBAAEC,QAAQgmD;;;mBAErCF,IAAaryD,KAAKwyD,GAAqB3wD;iCAGrCoY,uDAAsBA,KACxBja,KAAKsyD,GAAkCzwD,GAAOoY;QAEhDo4C,IAAaryD,KAAKwsC,UAAU4jB,GAAYqC,GACtC,eACA5wD;;+CAEqBoY;QAGzB,MAAMvS,IAASoS,GAAYiV,OAAO3c,GAAW6H,GAAIo4C;QAEjD,OADAryD,KAAK0yD,GAAkBhrD,IAChB,IAAIwQ,GACTlY,KAAKiyD,GAAOU,GAAUjrD,IACtB1H,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,QACE4S,GACAkgD;QASA,IAAIjkC;QACJ,IARAiW,GAA4B,iBAAiB4C,WAAW,GAAG,IAC3DtC,GACE,iBACA,oBACA,GACA0tB;aAGmB/sD,MAAjB+sD,KAA+C,UAAjBA,GAChCjkC,gCACK;YAAA,IAAqB,WAAjBikC,GAGT,MAAM,IAAIvxD,EACRlB,EAAKI,kBACL,mDAAmDqyD,SACjD;YALJjkC;;QAQF,IAA4B,SAAxB3uB,KAAKiyD,GAAOp6C,SACd,MAAM,IAAIxW,EACRlB,EAAKI,kBACL;QAIJ,IAA0B,SAAtBP,KAAKiyD,GAAOn6C,OACd,MAAM,IAAIzW,EACRlB,EAAKI,kBACL;QAIJ,MAAM6R,IAAYo5B,GAAsB,iBAAiB94B,IACnDiF,IAAU,IAAIiB,GAAQxG,GAAWuc;QAEvC,OADA3uB,KAAK6yD,GAAmBl7C,IACjB,IAAIO,GACTlY,KAAKiyD,GAAOa,GAAWn7C,IACvB3X,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,MAAMqe;QAIJ,OAHAqmB,GAA0B,eAAegD,WAAW,IACpD1C,GAAgB,eAAe,UAAU,GAAG3mB,IAC5C6oB,GAAuB,eAAe,GAAG7oB;QAClC,IAAIjG,GACTlY,KAAKiyD,GAAOc,GAAiB50C,IAC7Bne,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,YAAYqe;QAIV,OAHAqmB,GAA0B,qBAAqBgD,WAAW,IAC1D1C,GAAgB,qBAAqB,UAAU,GAAG3mB;QAClD6oB,GAAuB,qBAAqB,GAAG7oB,IACxC,IAAIjG,GACTlY,KAAKiyD,GAAOe,GAAgB70C,IAC5Bne,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,QACEmzD,MACGzoD;QAEHk6B,GAA4B,iBAAiB8C,WAAW;QACxD,MAAMpuB,IAAQpZ,KAAKkzD,GACjB,iBACAD,GACAzoD;qBACY;QAEd,OAAO,IAAI0N,GACTlY,KAAKiyD,GAAOkB,GAAY/5C,IACxBpZ,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,WACEmzD,MACGzoD;QAEHk6B,GAA4B,oBAAoB8C,WAAW;QAC3D,MAAMpuB,IAAQpZ,KAAKkzD,GACjB,oBACAD,GACAzoD;qBACY;QAEd,OAAO,IAAI0N,GACTlY,KAAKiyD,GAAOkB,GAAY/5C,IACxBpZ,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,UACEmzD,MACGzoD;QAEHk6B,GAA4B,mBAAmB8C,WAAW;QAC1D,MAAMpuB,IAAQpZ,KAAKkzD,GACjB,mBACAD,GACAzoD;qBACY;QAEd,OAAO,IAAI0N,GACTlY,KAAKiyD,GAAOmB,GAAUh6C,IACtBpZ,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,MACEmzD,MACGzoD;QAEHk6B,GAA4B,eAAe8C,WAAW;QACtD,MAAMpuB,IAAQpZ,KAAKkzD,GACjB,eACAD,GACAzoD;qBACY;QAEd,OAAO,IAAI0N,GACTlY,KAAKiyD,GAAOmB,GAAUh6C,IACtBpZ,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,QAAQkF;QACN,MAAMA,aAAiBkT,KACrB,MAAM6uB,GAAkB,WAAW,SAAS,GAAG/hC;QAEjD,OACEhF,KAAKwsC,cAAcxnC,EAAMwnC,aAAaxsC,KAAKiyD,GAAOzsD,QAAQR,EAAMitD;;IAIpEnyD,cACEisD;QAEA,OAAO,IAAI7zC,GAASlY,KAAKiyD,IAAQjyD,KAAKwsC,WAAWuf;;0EAI3CjsD,GACN8oC,GACAqqB,GACAzoD,GACA+P;QAGA,IADAqsB,GAAgBgC,GAAY,GAAGqqB,IAC3BA,aAAsBrD,IAAkB;YAC1C,IAAIplD,EAAO5E,SAAS,GAClB,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,kCAAkCqoC;YAGtC,MAAMmZ,IAAOkR;YACb,KAAKlR,EAAK9uC,QACR,MAAM,IAAI5R,EACRlB,EAAKM,WACL,yDACE,GAAGmoC;YAGT,OAAO5oC,KAAKqzD,GAAkBtR,EAAe,IAAExnC;;QAC1C;YACL,MAAM+4C,IAAY,EAACL,IAAY/5C,OAAO1O;YACtC,OAAOxK,KAAKuzD,GAAgB3qB,GAAY0qB,GAAW/4C;;;;;;;;;;;;;WAe/Cza,GAAkBkV,GAAeuF;QACvC,MAAMi5C,IAA0B;;;;;;;;gBAShC,KAAK,MAAM77C,KAAW3X,KAAKiyD,GAAOt6C,SAChC,IAAIA,EAAQjF,MAAMiG,KAChB66C,EAAWltD,KAAK2J,EAASjQ,KAAKwsC,UAAUC,IAAaz3B,EAAIhM,YACpD;YACL,MAAMnH,IAAQmT,EAAItC,MAAMiF,EAAQjF;YAChC,IAAIpI,EAAkBzI,IACpB,MAAM,IAAIR,EACRlB,EAAKI,kBACL,iGAEEoX,EAAQjF,QACR;YAGC,IAAc,SAAV7Q,GAEJ;gBACL,MAAM6Q,IAAQiF,EAAQjF,MAAMpL;gBAC5B,MAAM,IAAIjG,EACRlB,EAAKI,kBACL,mEACE,iCAAiCmS,qBACjC;;YAPJ8gD,EAAWltD,KAAKzE;;QAYtB,OAAO,IAAIwY,GAAMm5C,GAAYj5C;;;;WAMvBza,GACN8oC,GACAr8B,GACAgO;;QAGA,MAAM5C,IAAU3X,KAAKiyD,GAAO95C;QAC5B,IAAI5L,EAAO3G,SAAS+R,EAAQ/R,QAC1B,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,kCAAkCqoC,UAChC;QAKN,MAAM4qB,IAA0B;QAChC,KAAK,IAAI9sD,IAAI,GAAGA,IAAI6F,EAAO3G,QAAQc,KAAK;YACtC,MAAM+sD,IAAWlnD,EAAO7F;YAExB,IADyBiR,EAAQjR,GACZgM,MAAMiG,KAAc;gBACvC,IAAwB,mBAAb86C,GACT,MAAM,IAAIpyD,EACRlB,EAAKI,kBACL,yDACE,GAAGqoC,yBAAkC6qB;gBAG3C,KACGzzD,KAAKiyD,GAAOp8B,SACc,MAA3B49B,EAASjsD,QAAQ,MAEjB,MAAM,IAAInG,EACRlB,EAAKI,kBACL,uFACE,uBAAuBqoC,0CACvB,IAAI6qB;gBAGV,MAAMlsD,IAAOvH,KAAKiyD,GAAO1qD,KAAKyP,MAAM7P,EAAaqB,EAAWirD;gBAC5D,KAAKlrD,EAAY0P,GAAc1Q,IAC7B,MAAM,IAAIlG,EACRlB,EAAKI,kBACL,qEACE,+CAA+CqoC,0BAC/C,6BAA6BrhC,iDAC7B;gBAGN,MAAMyB,IAAM,IAAIT,EAAYhB;gBAC5BisD,EAAWltD,KAAK2J,EAASjQ,KAAKwsC,UAAUC,IAAazjC;mBAChD;gBACL,MAAM0qD,IAAU1zD,KAAKwsC,UAAU4jB,GAAYqC,GACzC7pB,GACA6qB;gBAEFD,EAAWltD,KAAKotD;;;QAIpB,OAAO,IAAIr5C,GAAMm5C,GAAYj5C;;IAsB/Bza,cAAcqwB;QACZyU,GAA4B,oBAAoB4C,WAAW,GAAG;QAC9D,IACIma,GADArqC,IAA2C,IAE3Cu5C,IAAU;QAyCd,OAvC2B,mBAAlB1gC,EAAK0gC,MACXtF,GAAkBp7B,EAAK0gC,QAExBv5C,IAAU6Y,EAAK0gC,IACfhqB,GAAoB,oBAAoBvvB,GAAS,EAC/C;QAEF+tB,GACE,oBACA,WACA,0BACA/tB,EAAQ0qC;QAEV6O,MAGEtF,GAAkBp7B,EAAK0gC,MACzBlP,IAAWxxB,EAAK0gC,MAEhB/rB,GAAgB,oBAAoB,YAAY+rB,GAAS1gC,EAAK0gC,KAC9D3rB,GACE,oBACA,YACA2rB,IAAU,GACV1gC,EAAK0gC,IAAU;QAEjB3rB,GACE,oBACA,YACA2rB,IAAU,GACV1gC,EAAK0gC,IAAU,KAEjBlP,IAAW;YACTr5C,MAAM6nB,EAAK0gC;YACXtgC,OAAOJ,EAAK0gC,IAAU;YACtBE,UAAU5gC,EAAK0gC,IAAU;YAG7B7wD,KAAK2zD,GAAyC3zD,KAAKiyD,KAC5CjyD,KAAKgxD,GAAmB15C,GAASqqC;;IAGlC7hD,GACNwX,GACAqqC;QAEA,IAAIsP,IAAc98B;YAChBk2B,QAAQ95B,MAAM,iCAAiC4D;;QAE7CwtB,EAASpxB,UACX0gC,IAAatP,EAASpxB,MAAMyJ,KAAK2nB;QAGnC,MAAMmN,IAAgB,IAAI3D,GAA4B;YACpD7iD,MAAO6G;gBACDwyC,EAASr5C,QACXq5C,EAASr5C,KACP,IAAIsrD,GACF5zD,KAAKwsC,WACLxsC,KAAKiyD,IACL9iD,GACAnP,KAAK6vD;;YAKbt/B,OAAO0gC;YAGH4C,IAAkB7zD,KAAKwsC,UAAU0gB,MACjCgE,IAAmB2C,EAAgBpX,OACvCz8C,KAAKiyD,IACLnD,GACAx3C;QAEF,OAAO;YACLw3C,EAAcC,MACd8E,EAAgBzW,GAAS8T;;;IAIrBpxD,GAAyCgiB;QAC/C,IAAIA,EAAM+2B,QAAqD,MAAjC/2B,EAAM3J,GAAgBvS,QAClD,MAAM,IAAIvE,EACRlB,EAAKc,eACL;;IAKNnB,IAAIwX;QAIF,OAHAstB,GAA4B,aAAa4C,WAAW,GAAG,IACvD2pB,GAAmB,aAAa75C,IAChCtX,KAAK2zD,GAAyC3zD,KAAKiyD,KAC5C,IAAIhwD,QACT,CAACC,GAA+C4xB;YAC1Cxc,KAA8B,YAAnBA,EAAQymC,SACrB/9C,KAAKwsC,UACF0gB,KACA4G,GAA2B9zD,KAAKiyD,IAChCjvD,KAAMm+C;gBACLj/C,EACE,IAAI0xD,GACF5zD,KAAKwsC,WACLxsC,KAAKiyD,IACL9Q,GACAnhD,KAAK6vD;eAGR/7B,KAEL9zB,KAAKqxD,GAAuBnvD,GAAS4xB,GAAQxc;;;IAM7CxX,GACNoC,GACA4xB,GACAxc;QAEA,MAAM8lC,IAAWp9C,KAAKgxD,GACpB;YACEhP,yBAAwB;YACxBsP,KAAuB;WAEzB;YACEhpD,MAAO6G;;;gBAGLiuC,KAGEjuC,EAAOoiD,SAASpvC,aAChB7K,KACmB,aAAnBA,EAAQymC,SAERjqB,EACE,IAAIzyB,EACFlB,EAAKgB,aACL,mLAOJe,EAAQiN;;YAGZohB,OAAOuD;;;;;;;WAULh0B,GAAqBi0D;QAC3B,IAA+B,mBAApBA,GAA8B;YACvC,IAAwB,OAApBA,GACF,MAAM,IAAI1yD,EACRlB,EAAKI,kBACL;YAIJ,KACGP,KAAKiyD,GAAOp8B,SACqB,MAAlCk+B,EAAgBvsD,QAAQ,MAExB,MAAM,IAAInG,EACRlB,EAAKI,kBACL,oHAEE,IAAIwzD;YAGV,MAAMxsD,IAAOvH,KAAKiyD,GAAO1qD,KAAKyP,MAC5B7P,EAAaqB,EAAWurD;YAE1B,KAAKxrD,EAAY0P,GAAc1Q,IAC7B,MAAM,IAAIlG,EACRlB,EAAKI,kBACL,yIAEE,QAAQgH,uDAA0DA,EAAK3B;YAG7E,OAAOqK,EAASjQ,KAAKwsC,UAAUC,IAAa,IAAIlkC,EAAYhB;;QACvD,IAAIwsD,aAA2B1nB,IAAmB;YACvD,MAAM1I,IAAMowB;YACZ,OAAO9jD,EAASjQ,KAAKwsC,UAAUC,IAAa9I,EAAI+I;;QAEhD,MAAM,IAAIrrC,EACRlB,EAAKI,kBACL,mIAEE,GAAGmlC,GAAiBquB;;;;;WASpBj0D,GACN+B,GACAmyD;QAEA,KAAKvuB,MAAMn1B,QAAQzO,MAA2B,MAAjBA,EAAM+D,QACjC,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,sDACE,IAAIyzD,EAAStyD;QAGnB,IAAIG,EAAM+D,SAAS,IACjB,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,mBAAmByzD,EAAStyD,mCAC1B;QAGN,IAAIG,EAAM2F,QAAQ,SAAS,GACzB,MAAM,IAAInG,EACRlB,EAAKI,kBACL,mBAAmByzD,EAAStyD,+CAC1B;QAGN,IAAIG,EAAM6F,OAAO4J,KAAW1B,OAAOzD,MAAMmF,IAAU1L,SAAS,GAC1D,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,mBAAmByzD,EAAStyD,8CAC1B;;IAKA5B,GAAkB4H;QACxB,IAAIA,aAAkBoS,IAAa;YACjC,MAAMm6C,IAAW,2FACXC,IAAiB,mEACjBC,IAAYF,EAASzsD,QAAQE,EAAOuS,OAAO,GAC3Cm6C,IAAkBF,EAAe1sD,QAAQE,EAAOuS,OAAO;YAE7D,IAAIvS,EAAOqS,MAAgB;gBACzB,MAAMs6C,IAAgBr0D,KAAKiyD,GAAOz5C;gBAClC,IAAsB,SAAlB67C,MAA2BA,EAAc7uD,QAAQkC,EAAOgL,QAC1D,MAAM,IAAIrR,EACRlB,EAAKI,kBACL,kHAEE,2BAA2B8zD,EAAc3yD,gBACzC,SAASgG,EAAOgL,MAAMhR;gBAI5B,MAAM+W,IAAoBzY,KAAKiyD,GAAOv5C;gBACZ,SAAtBD,KACFzY,KAAKs0D,GACH5sD,EAAOgL,OACP+F;mBAGC,IAAI27C,KAAmBD,GAAW;;;gBAGvC,IAAII,IAAiC;gBAOrC,IANIH,MACFG,IAAgBv0D,KAAKiyD,GAAOuC,GAAmBN,KAE3B,SAAlBK,KAA0BJ,MAC5BI,IAAgBv0D,KAAKiyD,GAAOuC,GAAmBP,KAE5B,QAAjBM;;gBAEF,MAAIA,MAAkB7sD,EAAOuS,KACrB,IAAI5Y,EACRlB,EAAKI,kBACL,iDACE,IAAImH,EAAOuS,GAAGvY,yBAGZ,IAAIL,EACRlB,EAAKI,kBACL,kCAAkCmH,EAAOuS,GAAGvY,yBAC1C,SAAS6yD,EAAc7yD;;;;IAQ7B5B,GAAmB6X;QACzB,IAA2C,SAAvC3X,KAAKiyD,GAAOv5C,MAAiC;;YAE/C,MAAMH,IAAkBvY,KAAKiyD,GAAOz5C;YACZ,SAApBD,KACFvY,KAAKs0D,GAAkC/7C,GAAiBZ,EAAQjF;;;IAK9D5S,GACN20D,GACA98C;QAEA,KAAKA,EAAQnS,QAAQivD,IACnB,MAAM,IAAIpzD,EACRlB,EAAKI,kBACL,+DACE,+BAA+Bk0D,EAAW/yD,iBAC1C,6BAA6B+yD,EAAW/yD,iBACxC,mEACA,gBAAgBiW,EAAQjW;;;;MAMrBkyD;IAOX9zD,YACmByvD,GACAmF,GACAC,GACA9E;kBAHAN,aACAmF,aACAC,aACA9E,GATnB7vD,UAAoE,MACpEA,UAA+D;QAU7DA,KAAKuxD,WAAW,IAAIC,GAClBmD,EAAUn9C,kBACVm9C,EAAUxyC;;IAIdJ;QACE,MAAM5S,IAAoD;QAE1D,OADAnP,KAAKoG,QAAQ4O,KAAO7F,EAAO7I,KAAK0O,KACzB7F;;IAGTmF;QACE,OAAOtU,KAAK20D,GAAU5yC,KAAK1Y;;IAG7B7C;QACE,OAAOxG,KAAK20D,GAAU5yC,KAAKvb;;IAG7B1G,QACEuzB,GACAuhC;QAEAhwB,GAA4B,yBAAyB4C,WAAW,GAAG,IACnE1C,GAAgB,yBAAyB,YAAY,GAAGzR;QACxDrzB,KAAK20D,GAAU5yC,KAAK3b,QAAQ4O;YAC1Bqe,EAASjqB,KAAKwrD,GAAS50D,KAAK60D,GAAsB7/C;;;IAItD8M;QACE,OAAO,IAAI5J,GAAMlY,KAAK00D,IAAgB10D,KAAKuvD,IAAYvvD,KAAK6vD;;IAG9D/vD,WACEwX;QAEIA,MACFuvB,GAAoB,4BAA4BvvB,GAAS,EACvD,6BAEF+tB,GACE,4BACA,WACA,0BACA/tB,EAAQ0qC;QAIZ,MAAMA,OACJ1qC,MAAWA,EAAQ0qC;QAGrB,IAAIA,KAA0BhiD,KAAK20D,GAAUtyC,IAC3C,MAAM,IAAIhhB,EACRlB,EAAKI,kBACL;QAkBJ,OAZGP,KAAK80D,MACN90D,KAAK+0D,OAAyC/S,MAE9ChiD,KAAK80D;;;;;;iBAmNTtoB,GACAwV,GACA7S,GACA4c;YAEA,IAAI5c,EAASntB,GAAQ3Y,KAAW;;;gBAG9B,IAAI2rD,GACAvuD,IAAQ;gBACZ,OAAO0oC,EAASltB,WAAWna,IAAI0Z;oBAC7B,MAAMxM,IAAM,IAAI68C,GACdrlB,GACAhrB,EAAOxM,IAAIhM,KACXwY,EAAOxM,KACPm6B,EAAShtB,WACTgtB,EAASjtB,GAAYrC,IAAI2B,EAAOxM,IAAIhM,MACpC+iD;oBAWF,OADAiJ,IAAUxzC,EAAOxM,KACV;wBACL2M,MAAM;wBACN3M,KAAAA;wBACAigD,WAAW;wBACXC,UAAUzuD;;;;YAGT;;;gBAGL,IAAI0uD,IAAehmB,EAASntB;gBAC5B,OAAOmtB,EAASltB,WACbva,OACC8Z,KAAUwgC,0BAA0BxgC,EAAOG,MAE5C7Z,IAAI0Z;oBACH,MAAMxM,IAAM,IAAI68C,GACdrlB,GACAhrB,EAAOxM,IAAIhM,KACXwY,EAAOxM,KACPm6B,EAAShtB,WACTgtB,EAASjtB,GAAYrC,IAAI2B,EAAOxM,IAAIhM,MACpC+iD;oBAEF,IAAIkJ,KAAY,GACZC,KAAY;oBAUhB,yBATI1zC,EAAOG,SACTszC,IAAWE,EAAa3tD,QAAQga,EAAOxM,IAAIhM,MAE3CmsD,IAAeA,EAAavgD,OAAO4M,EAAOxM,IAAIhM;wCAE5CwY,EAAOG,SACTwzC,IAAeA,EAAar1C,IAAI0B,EAAOxM,MACvCkgD,IAAWC,EAAa3tD,QAAQga,EAAOxM,IAAIhM,OAEtC;wBAAE2Y,MAAMyzC,GAAiB5zC,EAAOG;wBAAO3M,KAAAA;wBAAKigD,UAAAA;wBAAUC,UAAAA;;;;SAlRzCG,CACpBr1D,KAAKuvD,IACLvN,GACAhiD,KAAK20D,IACL30D,KAAK6vD,KAEP7vD,KAAK+0D,KAAuC/S,IAGvChiD,KAAK80D;;kEAIdh1D,QAAQkF;QACN,MAAMA,aAAiB4uD,KACrB,MAAM7sB,GAAkB,WAAW,iBAAiB,GAAG/hC;QAGzD,OACEhF,KAAKuvD,OAAevqD,EAAMuqD,MAC1BvvD,KAAK00D,GAAelvD,QAAQR,EAAM0vD,OAClC10D,KAAK20D,GAAUnvD,QAAQR,EAAM2vD,OAC7B30D,KAAK6vD,OAAe7qD,EAAM6qD;;IAItB/vD,GAAsBkV;QAC5B,OAAO,IAAI68C,GACT7xD,KAAKuvD,IACLv6C,EAAIhM,KACJgM,GACAhV,KAAKuxD,SAASpvC,WACdniB,KAAK20D,GAAUzyC,GAAYrC,IAAI7K,EAAIhM,MACnChJ,KAAK6vD;;;;MAKEZ,WAAwD/2C;IAEnEpY,YACWw1D,GACT9oB,GACAqjB;QAGA,IADApuD,MAAM0tD,GAAc7iC,GAAOgpC,IAAQ9oB,GAAWqjB,cAJrCyF,GAKLA,EAAM1vD,SAAS,KAAM,GACvB,MAAM,IAAIvE,EACRlB,EAAKI,kBACL,kGAEE,GAAG+0D,EAAMhuD,WAAyBguD,EAAM1vD;;IAKhDksB;QACE,OAAO9xB,KAAKiyD,GAAO1qD,KAAKwO;;IAG1B0W;QACE,MAAM4M,IAAar5B,KAAKiyD,GAAO1qD,KAAKqlB;QACpC,OAAIyM,EAAWhwB,MACN,OAEA,IAAIgjC,GACT,IAAI9jC,EAAY8wB,IAChBr5B,KAAKwsC;;IAKXjlC;QACE,OAAOvH,KAAKiyD,GAAO1qD,KAAKD;;IAG1BxH,IAAIkvD;QAaF,IAZApqB,GAA4B,2BAA2B4C,WAAW,GAAG;;;QAG5C,MAArBA,UAAU5hC,WACZopD,IAAal+B,GAAOg4B,OAEtBhkB,GACE,2BACA,oBACA,GACAkqB;QAEiB,OAAfA,GACF,MAAM,IAAI3tD,EACRlB,EAAKI,kBACL;QAGJ,MAAMgH,IAAOJ,EAAaqB;QAC1B,OAAO6jC,GAAkB6iB,GACvBlvD,KAAKiyD,GAAO1qD,KAAKyP,MAAMzP,IACvBvH,KAAKwsC,WACLxsC,KAAK6vD;;IAIT/vD,IAAI+B;QACF2iC,GAA0B,2BAA2BgD,WAAW,IAIhE1C,GAAgB,2BAA2B,UAAU,GAH9B9kC,KAAK6vD,KACxB7vD,KAAK6vD,GAAW0F,YAAY1zD,KAC5BA;QAEJ,MAAM2zD,IAASx1D,KAAKgV;QACpB,OAAOwgD,EAAO7gD,IAAI9S,GAAOmB,KAAK,MAAMwyD;;IAGtC11D,cACEisD;QAEA,OAAO,IAAIkD,GAAuBjvD,KAAKs1D,IAAOt1D,KAAKwsC,WAAWuf;;;;AAIlE,SAAS+D,GACPlnB,GACAtxB;IAEA,SAAgBzR,MAAZyR,GACF,OAAO;QACL44C,QAAO;;IAeX,IAXArpB,GAAoB+B,GAAYtxB,GAAS,EAAC,SAAS,kBACnD+tB,GAA0BuD,GAAY,WAAW,SAAStxB,EAAQ44C,QAClE5qB,GACEsD,GACA,eACA,2BACAtxB,EAAQ64C,aACR7+C,KACqB,mBAAZA,KAAwBA,aAAmBu5B;SAG1BhlC,MAAxByR,EAAQ64C,oBAA+CtqD,MAAlByR,EAAQ44C,OAC/C,MAAM,IAAI7uD,EACRlB,EAAKI,kBACL,sCAAsCqoC,0CACpC;IAIN,OAAOtxB;;;AAGT,SAASs6C,GACPhpB,GACAtxB;IAEA,YAAgBzR,MAAZyR,IACK,MAGTuvB,GAAoB+B,GAAYtxB,GAAS,EAAC,uBAC1CsuB,GACEgD,GACA,GACA,oBACAtxB,EAAQ06C,kBACR,EAAC,YAAY,YAAY;IAEpB16C;;;AAGT,SAAS65C,GACPvoB,GACAtxB;IAEA4tB,GAAwB0D,GAAY,UAAU,GAAGtxB,IAC7CA,MACFuvB,GAAoB+B,GAAYtxB,GAAS,EAAC,aAC1CsuB,GACEgD,GACA,GACA,UACAtxB,EAAQymC,QACR,EAAC,WAAW,UAAU;;;AAK5B,SAAS2R,GACP9mB,GACA6mB,GACAjjB;IAEA,IAAMijB,aAAuBpjB,IAEtB;QAAA,IAAIojB,EAAYjjB,cAAcA,GACnC,MAAM,IAAInrC,EACRlB,EAAKI,kBACL;QAGF,OAAOkvD;;IAPP,MAAM1oB,GAAkB6B,GAAY,qBAAqB,GAAG6mB;;;AAqFhE,SAAS2F,GAAiBzzC;IACxB,QAAQA;MACN;QACE,OAAO;;MACT;MACA;QACE,OAAO;;MACT;QACE,OAAO;;MACT;QACE,OAp8EsB7b;;;;;;;;;;;;GAi9E5B,UAASkqD,GACPjE,GACAlqD,GACAyiC;IAEA,IAAIyrB;IAOJ,OANIhE,KACFgE,IAAiBhE,EAAUwJ,YAAY1zD,IACvCyiC,IAAe,sBAAsBA,KAErCyrB,IAAiBluD,GAEZ,EAACkuD,GAAgBzrB;;;SCv9EVmxB,GACdC,GACAC;IAEA,SAASC;QACP,IAAIrlC,IAAQ;QAKZ,MAJIolC,MACFplC,KAAS,KACTA,KAASolC,IAEL,IAAIt0D,EAAelB,EAAKI,kBAAkBgwB;;;;;IAWlD,OANAqlC,EAAkB1sD,YAAYwsD,EAAIxsD;;IAGlCD,OAAO4sD,OAAOD,GAAmBF,IAG1BE;;;;;;;;;;;;;;;;;;;;4BCTF;MAAME,KAAkBL,GAC7BzI,IACA,sCAEW+I,KAAoBN,GAC/B7kB,IACA,uDAEWolB,KAAmBP,GAC9BrG,IACA,8CAEW6G,KAA0BR,GACrCppB,IACA,4CAEW6pB,KAAyBT,GAAuB7F,KAChDuG,KAA8BV,GACzC5D,KAEWuE,KAAcX,GAAuBv9C,KACrCm+C,KAAsBZ,GAAuB7B,KAC7C0C,KAA4Bb,GACvCxG,IACA,mDAEWsH,KAAmBd;I1ByH9B31D;QAEE,OADAukC,GAAe,qBAAqBmD,YAC7B,IAAIU;;IAGbpoC;QAEE,OADAukC,GAAe,8BAA8BmD,YACtC,IAAIc;;IAGbxoC,qBAAqBmR;;;QAInB,OAHAyzB,GAA4B,yBAAyB8C,WAAW,IAGzD,IAAIe,GAAyBt3B;;IAGtCnR,sBAAsBmR;;;QAIpB,OAHAyzB,GAA4B,0BAA0B8C,WAAW,IAG1D,IAAI0B,GAA0Bj4B;;IAGvCnR,iBAAiBqe;QAGf,OAFA2mB,GAAgB,wBAAwB,UAAU,GAAG3mB,IACrDqmB,GAA0B,wBAAwBgD,WAAW;QACtD,IAAI2B,GAA+BhrB;;IAG5Cre,QAAQkF;QACN,OAAOhF,SAASgF;;G0BtJlB,sCAEWwxD,KAAaf,GACxBpuB,IACA,kEAGIovB,KAAqB;IACzBzJ,WAAW8I;IACXxsB,UAAAA;IACAllC,WAAAA;IACAijC,MAAMmvB;IACN5lB,aAAamlB;IACb3G,YAAY4G;IACZ3pB,mBAAmB4pB;IACnBrG,kBAAkBsG;IAClBh+C,OAAOk+C;IACPvE,uBAAuBsE;IACvBvC,eAAeyC;IACfpH,qBAAqBqH;eACrB1uD;IACA8uD,YAAYH;IACZzmC,aAAak9B,GAAUl9B;IACvB68B,sBAAAA;;;;;;;;;;;;;;;;;;;MC1EWgK;IACX72D,GAAYuzB;;;IAIZvzB;;;;;;;;;;;;;;;;;;;;;;;;;;MCQW82D;IAOX92D;QANAE,UAA4C,MAC1CA,KAAK62D,MACP72D,UAA8C,MAC5CA,KAAK82D,MACP92D,UAAmD,IAGjDA,KAAK+2D;;IAGPj3D,GAAYuzB;QACVrzB,KAAK8+C,GAAUx4C,KAAK+sB;;IAGtBvzB;QACEg7B,OAAOQ,oBAAoB,UAAUt7B,KAAKg3D,KAC1Cl8B,OAAOQ,oBAAoB,WAAWt7B,KAAKi3D;;IAGrCn3D;QACNg7B,OAAOC,iBAAiB,UAAU/6B,KAAKg3D,KACvCl8B,OAAOC,iBAAiB,WAAW/6B,KAAKi3D;;IAGlCn3D;QACNunB,GA/BY,uBA+BM;QAClB,KAAK,MAAMgM,KAAYrzB,KAAK8+C,IAC1BzrB;;IAIIvzB;QACNunB,GAtCY,uBAsCM;QAClB,KAAK,MAAMgM,KAAYrzB,KAAK8+C,IAC1BzrB;;;;;IAOJvzB;QACE,OACoB,sBAAXg7B,eACqBj1B,MAA5Bi1B,OAAOC,yBACwBl1B,MAA/Bi1B,OAAOQ;;;;;;;;;;;;;;;;;;;;;;;;UCnDA47B;IAQXp3D,YAAYqwB;QACVnwB,KAAKm3D,KAAShnC,EAAKgnC,IACnBn3D,KAAKo3D,KAAUjnC,EAAKinC;;IAGtBt3D,GAAOuzB;QAELrzB,KAAKq3D,KAAgBhkC;;IAGvBvzB,GAAQuzB;QAENrzB,KAAKs3D,KAAiBjkC;;IAGxBvzB,UAAUuzB;QAERrzB,KAAKu3D,KAAmBlkC;;IAG1BvzB;QACEE,KAAKo3D;;IAGPt3D,KAAKkwB;QACHhwB,KAAKm3D,GAAOnnC;;IAGdlwB;QAKEE,KAAKq3D;;IAGPv3D,GAAYq0B;QAKVn0B,KAAKs3D,GAAenjC;;IAGtBr0B,GAAckwB;QAKZhwB,KAAKu3D,GAAiBvnC;;;;;;;;;;;;;;;;;;;GChC1B,OASMwnC,KAAmD;IACzDC,mBAA6C;IAC7CC,QAAkC;GAK5BC,KAA0B,iBAAiBh4D;;MAIpCi4D;IAKX93D,YAAY+3D;QACV73D,KAAKkQ,KAAa2nD,EAAK3nD;QACvB,MAAM4F,IAAQ+hD,EAAKrmC,MAAM,UAAU;QACnCxxB,KAAK83D,KAAUhiD,IAAQ,QAAQ+hD,EAAKtmC,MACpCvxB,KAAKyxB,mBAAmBomC,EAAKpmC;;;;;WAOvB3xB,GACNiE,GACAuqC;QAEA,IAAIA,GACF,KAAK,MAAMypB,KAAUzpB,EAAMvsC,GACrBusC,EAAMvsC,EAAYoH,eAAe4uD,OACnCh0D,EAAQg0D,KAAUzpB,EAAMvsC,EAAYg2D;QAI1Ch0D,EAAQ,uBAAuB4zD;;IAGjC73D,GACE2wC,GACAnB,GACAhB;QAEA,MAAM0pB,IAAMh4D,KAAKi4D,GAAQxnB;QAEzB,OAAO,IAAIxuC,QAAQ,CAACC,GAAyB4xB;YAC3C,MAAMokC,IAAM,IAAIC;YAChBD,EAAIE,WAAWC,EAAUC,UAAU;gBACjC;oBACE,QAAQJ,EAAIK;sBACV,KAAKC,EAAUC;wBACb,MAAMC,IAAOR,EAAIS;wBACjBtxC,GAhEE,cAgEgB,iBAAiBmf,KAAKC,UAAUiyB,KAClDx2D,EAAQw2D;wBACR;;sBACF,KAAKF,EAAUI;wBACbvxC,GApEE,cAoEgB,UAAUopB,IAAU,gBACtC3c,EACE,IAAIzyB,EAAelB,EAAKK,mBAAmB;wBAE7C;;sBACF,KAAKg4D,EAAUK;wBACb,MAAMvwC,IAAS4vC,EAAIY;wBAQnB,IAPAzxC,GA3EE,cA6EA,UAAUopB,IAAU,yBACpBnoB,GACA,kBACA4vC,EAAIa;wBAEFzwC,IAAS,GAAG;4BACd,MAAM0wC,IAAiBd,EAAIS,kBACxBpoC;4BACH,IACIyoC,KACAA,EAAc1wC,UACd0wC,EAAcx3D,SAChB;gCACA,MAAMy3D,a/DwK2B3wC;oCACjD,MAAM4wC,IAAc5wC,EAAO6wC,cAAcnxD,QAAQ,KAAK;oCACtD,OAAOiB,OAAOsD,OAAOpM,GAAMqH,QAAQ0xD,MAAwB,IACtDA,IACD/4D,EAAKG;iC+D5KkC84D,CACzBJ,EAAc1wC;gCAEhBwL,EACE,IAAIzyB,EACF43D,GACAD,EAAcx3D;mCAIlBsyB,EACE,IAAIzyB,EACFlB,EAAKG,SACL,kCAAkC43D,EAAIY;;;;wBAO5CzxC,GA9GA,cA8GkB,UAAUopB,IAAU,aACtC3c,EACE,IAAIzyB,EAAelB,EAAKgB,aAAa;wBAGzC;;sBACF;wBACE2E;;;oBAYJuhB,GAjIM,cAiIY,UAAUopB,IAAU;;;;;;YAO1C,MAAM4oB,IAAWpwD,kBAAKqmC;mBACf+pB,EAAQjpD;YAEf,MAAMkpD,IAAgB9yB,KAAKC,UAAU4yB;YACrChyC,GA5IU,cA4IQ,iBAAiB2wC,IAAM,MAAMsB;;;;;;YAM/C,MAAMv1D,IAAqB;gBAAEw1D,gBAAgB;;YAE7Cv5D,KAAKw5D,GAAwBz1D,GAASuqC,IAEtC4pB,EAAIrqB,KAAKmqB,GAAK,QAAQsB,GAAev1D,GApIlB;;;IAwIvBjE,GACE2wC,GACAnB,GACAhB;;;QAIA,OAAOtuC,KAAK0wC,GAAuBD,GAASnB,GAAShB;;IAGvDxuC,GACE2wC,GACAnC;QAEA,MAAMmrB,IAAW,EACfz5D,KAAK83D,IACL,KAxKqB,iCA0KrB,KACArnB,GACA,cAEIipB,IAAsBC,KACtBrqB,IAA6B;;;YAGjCsqB,oBAAoB;YACpBC,oBAAoB;YACpBC,kBAAkB;;;gBAGhB1pD,UAAU,YAAYpQ,KAAKkQ,GAAWC,uBAAuBnQ,KAAKkQ,GAAWE;;YAE/E2pD,cAAa;YACbC,yBAAwB;YACxBC,uBAAuB;;;;;;;gBAOrBC,gCAAgC;;YAElCzoC,kBAAkBzxB,KAAKyxB;;QAGzBzxB,KAAKw5D,GAAwBlqB,EAA2B,oBAAEhB;;;;;;;;;;;;;;;;QAoBvD6rB,OACAC,OACAC,OACAC,OACAC,OACAC,QAEDlrB,EAAQmrB,4BAA4B;QAGtC,MAAMzC,IAAMyB,EAASpyD,KAAK;QAC1BggB,GAxOY,cAwOM,0BAA0B2wC,IAAM,MAAM1oB;QACxD,MAAMorB,IAAUhB,EAAoBiB,iBAAiB3C,GAAK1oB;;;;;;gBAO1D,IAAIsrB,KAAS,GAKTC,KAAS;;;;gBAEb,MAAMC,IAAe,IAAI5D,GAAwB;YAC/C6D,IAAS/qC;gBACF6qC,IASHxzC,GAlQM,cAkQY,6CAA6C2I,MAR1D4qC,MACHvzC,GA3PI,cA2Pc;gBAClBqzC,EAAQM,QACRJ,KAAS,IAEXvzC,GA/PM,cA+PY,uBAAuB2I,IACzC0qC,EAAQ7sB,KAAK7d;;YAKjBirC,IAAS,MAAMP,EAAQntB;YAOnB2tB,IAAuB,CAC3Bv5C,GACA/a;;;YAIA8zD,EAAQje,OAAO96B,GAAOw5C;gBACpB;oBACEv0D,EAAGu0D;kBACH,OAAO1qC;oBACP6J,WAAW;wBACT,MAAM7J;uBACL;;;;;;;;gBAuFT,OAlFAyqC,EAAqBE,EAAW/C,UAAUgD,MAAM;YACzCR,KACHxzC,GA/RQ,cA+RU;YAItB6zC,EAAqBE,EAAW/C,UAAUiD,OAAO;YAC1CT,MACHA,KAAS,GACTxzC,GAtSQ,cAsSU,gCAClByzC,EAAaS;YAIjBL,EAA4BE,EAAW/C,UAAU/nC,OAAO6D;YACjD0mC,MACHA,KAAS,GACTxzC,GA9SQ,cA8SU,iCAAiC8M,IACnD2mC,EAAaS,GACX,IAAIl6D,EACFlB,EAAKgB,aACL;YAaR+5D,EACEE,EAAW/C,UAAUmD,SACrBxrC;;YACE,KAAK6qC,GAAQ;gBACX,MAAMY,IAAUzrC,EAAKhc,KAAK;gBAjUc3Q,KAkU3Bo4D;;;;;;gBAMb,MAAMC,IAA2CD,GAC3ClrC,IACJmrC,EAAenrC,wBACdmrC,EAAqC,iCAAInrC;gBAC5C,IAAIA,GAAO;oBACTlJ,GA/UI,cA+Uc,8BAA8BkJ;;oBAEhD,MAAMjI,IAAiBiI,EAAMjI;oBAC7B,IAAI/mB,a/DvRqB+mB;;;wBAGnC,MAAM/mB,IAAgB8a,GAAQiM;wBAC9B,SAAaziB,MAATtE,GAIJ,OAAOgb,GAAmBhb;qB+D+QLo6D,CAAqBrzC,IAC5B9mB,IAAU+uB,EAAM/uB;yBACPqE,MAATtE,MACFA,IAAOpB,EAAKe,UACZM,IACE,2BACA8mB,IACA,mBACAiI,EAAM/uB;;oBAGVq5D,KAAS,GACTC,EAAaS,GAAY,IAAIl6D,EAAeE,GAAMC,KAClDk5D,EAAQntB;uBAERlmB,GAjWI,cAiWc,wBAAwBo0C,IAC1CX,EAAac,GAAcH;;YAMnCnhC,WAAW;;;;;YAKTwgC,EAAae;WACZ,IACIf;;;IAITh7D,GAAQ2wC;QACN,MAAMqrB,IAAatE,GAAsB/mB;QAKzC,OACEzwC,KAAK83D,KACL,kBAGA93D,KAAKkQ,GAAWC,YAChB,gBACAnQ,KAAKkQ,GAAWE,WAChB,gBACA0rD;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AC3ZNryD,GAAgBsyD,GAAY;ICM1Bj8D;QACEE,KAAKonC,KAAkC,sBAATz9B;;IAGhCid;;;QAGE,OAA2B,sBAAbA,WAA2BA,WAAW;;IAGtDkU;;;QAGE,OAAyB,sBAAXA,SAAyBA,SAAS;;IAGlDh7B,GAAe8oD;QACb,OAAO3mD,QAAQC,QAAQ,IAAI01D,GAAqBhP;;IAGlD9oD;QACE,OAAI82D,GAA2BoF,OACtB,IAAIpF,KAEJ,IAAID;;IAIf72D,GAAcoQ;QACZ,OAAO,IAAImY,GAAoBnY,GAAY;YAAE+rD,KAAe;;;IAG9Dn8D,GAAW+B;QACT,OAAO2kC,KAAKC,UAAU5kC;;IAGxB/B,KAAKo8D;QACH,OAAOvyD,KAAKuyD;;IAGdp8D,KAAKq8D;QACH,OAAOpyD,KAAKoyD;;IAGdr8D,GAAYs8D;;QAIV,MAAMC;;QAEY,sBAATC,SAAyBA,KAAKD,UAAWC,KAAuB,WACnEzzC,IAAQ,IAAI5e,WAAWmyD;QAC7B,IAAIC,GACFA,EAAOE,gBAAgB1zC;;QAGvB,KAAK,IAAIniB,IAAI,GAAGA,IAAI01D,GAAQ11D,KAC1BmiB,EAAMniB,KAAK7B,KAAKC,MAAsB,MAAhBD,KAAKm0B;QAG/B,OAAOnQ;;;;;;;;;;;;;;;;;;;;;;;SC7DK2zC,GAAkBtwC;;;;;;;;;aPwEhCtsB,GACA68D;QAKC78D,EAAgCsB,SAASw7D,kBACxC,IAAIC,EACF,aACAC;YACE,MAAMxP,IAAMwP,EAAUC,YAAY,OAAOj6D;YACzC,OAAO65D,EAAiBrP,GAAKwP,EAAUC,YAAY;kCAGrDC,kCAAqBrG;KOrFzBsG,CACE7wC,GACA,CAACkhC,GAAKzqD,MAAS,IAAIqqD,GAAUI,GAAKzqD,GAAM,IAAImlD,MAE9C57B,EAAS8wC;;;AAGXR,GAAkB58D;;"} |