2020-08-27 19:12:06 +02:00
|
|
|
/*
|
|
|
|
Copyright 2020 The Matrix.org Foundation C.I.C.
|
|
|
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
you may not use this file except in compliance with the License.
|
|
|
|
You may obtain a copy of the License at
|
|
|
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
limitations under the License.
|
|
|
|
*/
|
|
|
|
|
2021-10-05 17:51:10 +02:00
|
|
|
import anotherjson from "another-json";
|
2023-02-27 18:13:53 +01:00
|
|
|
import {SESSION_E2EE_KEY_PREFIX, OLM_ALGORITHM, MEGOLM_ALGORITHM} from "./common";
|
2020-08-27 19:12:06 +02:00
|
|
|
|
2020-08-28 13:52:27 +02:00
|
|
|
// use common prefix so it's easy to clear properties that are not e2ee related during session clear
|
2021-09-29 11:49:58 +02:00
|
|
|
const ACCOUNT_SESSION_KEY = SESSION_E2EE_KEY_PREFIX + "olmAccount";
|
|
|
|
const DEVICE_KEY_FLAG_SESSION_KEY = SESSION_E2EE_KEY_PREFIX + "areDeviceKeysUploaded";
|
|
|
|
const SERVER_OTK_COUNT_SESSION_KEY = SESSION_E2EE_KEY_PREFIX + "serverOTKCount";
|
2020-08-27 19:12:06 +02:00
|
|
|
|
2021-10-26 18:47:46 +02:00
|
|
|
async function initiallyStoreAccount(account, pickleKey, areDeviceKeysUploaded, serverOTKCount, storage) {
|
|
|
|
const pickledAccount = account.pickle(pickleKey);
|
|
|
|
const txn = await storage.readWriteTxn([
|
|
|
|
storage.storeNames.session
|
|
|
|
]);
|
|
|
|
try {
|
|
|
|
// add will throw if the key already exists
|
|
|
|
// we would not want to overwrite olmAccount here
|
|
|
|
txn.session.add(ACCOUNT_SESSION_KEY, pickledAccount);
|
|
|
|
txn.session.add(DEVICE_KEY_FLAG_SESSION_KEY, areDeviceKeysUploaded);
|
|
|
|
txn.session.add(SERVER_OTK_COUNT_SESSION_KEY, serverOTKCount);
|
|
|
|
} catch (err) {
|
|
|
|
txn.abort();
|
|
|
|
throw err;
|
|
|
|
}
|
|
|
|
await txn.complete();
|
|
|
|
}
|
|
|
|
|
2020-08-27 19:12:06 +02:00
|
|
|
export class Account {
|
2020-09-11 10:43:17 +02:00
|
|
|
static async load({olm, pickleKey, hsApi, userId, deviceId, olmWorker, txn}) {
|
2020-08-27 19:12:06 +02:00
|
|
|
const pickledAccount = await txn.session.get(ACCOUNT_SESSION_KEY);
|
|
|
|
if (pickledAccount) {
|
|
|
|
const account = new olm.Account();
|
|
|
|
const areDeviceKeysUploaded = await txn.session.get(DEVICE_KEY_FLAG_SESSION_KEY);
|
|
|
|
account.unpickle(pickleKey, pickledAccount);
|
2020-08-28 13:56:44 +02:00
|
|
|
const serverOTKCount = await txn.session.get(SERVER_OTK_COUNT_SESSION_KEY);
|
|
|
|
return new Account({pickleKey, hsApi, account, userId,
|
2020-09-11 10:43:17 +02:00
|
|
|
deviceId, areDeviceKeysUploaded, serverOTKCount, olm, olmWorker});
|
2020-08-27 19:12:06 +02:00
|
|
|
}
|
|
|
|
}
|
2021-10-27 10:26:36 +02:00
|
|
|
|
2021-10-26 18:47:46 +02:00
|
|
|
static async adoptDehydratedDevice({olm, dehydratedDevice, pickleKey, hsApi, userId, olmWorker, storage}) {
|
|
|
|
const account = dehydratedDevice.adoptUnpickledOlmAccount();
|
2021-10-27 10:26:36 +02:00
|
|
|
const oneTimeKeys = JSON.parse(account.one_time_keys());
|
2021-10-26 18:47:46 +02:00
|
|
|
// only one algorithm supported by olm atm, so hardcode its name
|
|
|
|
const oneTimeKeysEntries = Object.entries(oneTimeKeys.curve25519);
|
|
|
|
const serverOTKCount = oneTimeKeysEntries.length;
|
2021-10-27 10:26:36 +02:00
|
|
|
const areDeviceKeysUploaded = true;
|
2021-10-26 18:47:46 +02:00
|
|
|
await initiallyStoreAccount(account, pickleKey, areDeviceKeysUploaded, serverOTKCount, storage);
|
|
|
|
return new Account({
|
|
|
|
pickleKey, hsApi, account, userId,
|
|
|
|
deviceId: dehydratedDevice.deviceId,
|
|
|
|
areDeviceKeysUploaded, serverOTKCount, olm, olmWorker
|
|
|
|
});
|
|
|
|
}
|
2020-08-27 19:12:06 +02:00
|
|
|
|
2020-09-11 10:43:17 +02:00
|
|
|
static async create({olm, pickleKey, hsApi, userId, deviceId, olmWorker, storage}) {
|
2020-08-27 19:12:06 +02:00
|
|
|
const account = new olm.Account();
|
2020-09-11 10:43:17 +02:00
|
|
|
if (olmWorker) {
|
|
|
|
await olmWorker.createAccountAndOTKs(account, account.max_number_of_one_time_keys());
|
|
|
|
} else {
|
|
|
|
account.create();
|
|
|
|
account.generate_one_time_keys(account.max_number_of_one_time_keys());
|
|
|
|
}
|
2021-10-27 10:26:36 +02:00
|
|
|
const areDeviceKeysUploaded = false;
|
|
|
|
const serverOTKCount = 0;
|
2021-10-26 18:47:46 +02:00
|
|
|
if (storage) {
|
2021-10-27 10:26:36 +02:00
|
|
|
await initiallyStoreAccount(account, pickleKey, areDeviceKeysUploaded, serverOTKCount, storage);
|
2020-09-11 10:43:17 +02:00
|
|
|
}
|
2020-08-28 13:56:44 +02:00
|
|
|
return new Account({pickleKey, hsApi, account, userId,
|
2021-10-27 10:26:36 +02:00
|
|
|
deviceId, areDeviceKeysUploaded, serverOTKCount, olm, olmWorker});
|
2020-08-27 19:12:06 +02:00
|
|
|
}
|
|
|
|
|
2020-09-11 10:43:17 +02:00
|
|
|
constructor({pickleKey, hsApi, account, userId, deviceId, areDeviceKeysUploaded, serverOTKCount, olm, olmWorker}) {
|
2020-09-01 17:59:39 +02:00
|
|
|
this._olm = olm;
|
2020-08-27 19:12:06 +02:00
|
|
|
this._pickleKey = pickleKey;
|
|
|
|
this._hsApi = hsApi;
|
|
|
|
this._account = account;
|
|
|
|
this._userId = userId;
|
|
|
|
this._deviceId = deviceId;
|
|
|
|
this._areDeviceKeysUploaded = areDeviceKeysUploaded;
|
2020-08-28 13:56:44 +02:00
|
|
|
this._serverOTKCount = serverOTKCount;
|
2020-09-11 10:43:17 +02:00
|
|
|
this._olmWorker = olmWorker;
|
2020-09-01 17:59:39 +02:00
|
|
|
this._identityKeys = JSON.parse(this._account.identity_keys());
|
|
|
|
}
|
|
|
|
|
|
|
|
get identityKeys() {
|
|
|
|
return this._identityKeys;
|
2020-08-27 19:12:06 +02:00
|
|
|
}
|
2020-08-27 19:13:24 +02:00
|
|
|
|
2021-10-27 18:08:50 +02:00
|
|
|
setDeviceId(deviceId) {
|
|
|
|
this._deviceId = deviceId;
|
|
|
|
}
|
|
|
|
|
|
|
|
async uploadKeys(storage, isDehydratedDevice, log) {
|
2020-08-27 19:13:24 +02:00
|
|
|
const oneTimeKeys = JSON.parse(this._account.one_time_keys());
|
|
|
|
// only one algorithm supported by olm atm, so hardcode its name
|
|
|
|
const oneTimeKeysEntries = Object.entries(oneTimeKeys.curve25519);
|
|
|
|
if (oneTimeKeysEntries.length || !this._areDeviceKeysUploaded) {
|
|
|
|
const payload = {};
|
|
|
|
if (!this._areDeviceKeysUploaded) {
|
2021-02-23 19:22:25 +01:00
|
|
|
log.set("identity", true);
|
2020-08-27 19:13:24 +02:00
|
|
|
const identityKeys = JSON.parse(this._account.identity_keys());
|
2021-10-27 18:08:50 +02:00
|
|
|
payload.device_keys = this._deviceKeysPayload(identityKeys);
|
2020-08-27 19:13:24 +02:00
|
|
|
}
|
|
|
|
if (oneTimeKeysEntries.length) {
|
2021-02-23 19:22:25 +01:00
|
|
|
log.set("otks", true);
|
2020-08-27 19:13:24 +02:00
|
|
|
payload.one_time_keys = this._oneTimeKeysPayload(oneTimeKeysEntries);
|
|
|
|
}
|
2021-10-27 18:08:50 +02:00
|
|
|
const dehydratedDeviceId = isDehydratedDevice ? this._deviceId : undefined;
|
2021-10-26 18:47:46 +02:00
|
|
|
const response = await this._hsApi.uploadKeys(dehydratedDeviceId, payload, {log}).response();
|
2020-08-28 13:56:44 +02:00
|
|
|
this._serverOTKCount = response?.one_time_key_counts?.signed_curve25519;
|
2021-02-23 19:22:25 +01:00
|
|
|
log.set("serverOTKCount", this._serverOTKCount);
|
2020-08-28 13:56:44 +02:00
|
|
|
// TODO: should we not modify this in the txn like we do elsewhere?
|
|
|
|
// we'd have to pickle and unpickle the account to clone it though ...
|
|
|
|
// and the upload has succeed at this point, so in-memory would be correct
|
|
|
|
// but in-storage not if the txn fails.
|
2020-08-27 19:13:24 +02:00
|
|
|
await this._updateSessionStorage(storage, sessionStore => {
|
|
|
|
if (oneTimeKeysEntries.length) {
|
|
|
|
this._account.mark_keys_as_published();
|
2021-10-26 18:47:46 +02:00
|
|
|
sessionStore?.set(ACCOUNT_SESSION_KEY, this._account.pickle(this._pickleKey));
|
|
|
|
sessionStore?.set(SERVER_OTK_COUNT_SESSION_KEY, this._serverOTKCount);
|
2020-08-27 19:13:24 +02:00
|
|
|
}
|
|
|
|
if (!this._areDeviceKeysUploaded) {
|
|
|
|
this._areDeviceKeysUploaded = true;
|
2021-10-26 18:47:46 +02:00
|
|
|
sessionStore?.set(DEVICE_KEY_FLAG_SESSION_KEY, this._areDeviceKeysUploaded);
|
2020-08-27 19:13:24 +02:00
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-23 19:22:25 +01:00
|
|
|
async generateOTKsIfNeeded(storage, log) {
|
2021-03-09 12:33:31 +01:00
|
|
|
// We need to keep a pool of one time public keys on the server so that
|
|
|
|
// other devices can start conversations with us. But we can only store
|
|
|
|
// a finite number of private keys in the olm Account object.
|
2022-01-05 14:25:42 +01:00
|
|
|
// To complicate things further there can be a delay between a device
|
2021-03-09 12:33:31 +01:00
|
|
|
// claiming a public one time key from the server and it sending us a
|
|
|
|
// message. We need to keep the corresponding private key locally until
|
|
|
|
// we receive the message.
|
|
|
|
// But that message might never arrive leaving us stuck with duff
|
|
|
|
// private keys clogging up our local storage.
|
|
|
|
// So we need some kind of engineering compromise to balance all of
|
|
|
|
// these factors.
|
|
|
|
|
|
|
|
// Check how many keys we can store in the Account object.
|
2020-08-28 13:58:17 +02:00
|
|
|
const maxOTKs = this._account.max_number_of_one_time_keys();
|
2021-03-09 12:33:31 +01:00
|
|
|
// Try to keep at most half that number on the server. This leaves the
|
|
|
|
// rest of the slots free to hold keys that have been claimed from the
|
|
|
|
// server but we haven't recevied a message for.
|
|
|
|
// If we run out of slots when generating new keys then olm will
|
|
|
|
// discard the oldest private keys first. This will eventually clean
|
|
|
|
// out stale private keys that won't receive a message.
|
|
|
|
const keyLimit = Math.floor(maxOTKs / 2);
|
|
|
|
// does the server have insufficient OTKs?
|
|
|
|
if (this._serverOTKCount < keyLimit) {
|
2020-08-28 13:58:17 +02:00
|
|
|
const oneTimeKeys = JSON.parse(this._account.one_time_keys());
|
|
|
|
const oneTimeKeysEntries = Object.entries(oneTimeKeys.curve25519);
|
|
|
|
const unpublishedOTKCount = oneTimeKeysEntries.length;
|
2021-03-09 12:33:31 +01:00
|
|
|
// we want to end up with maxOTKs / 2 key on the server,
|
|
|
|
// so generate any on top of the remaining ones on the server and the unpublished ones
|
|
|
|
// (we have generated before but haven't uploaded yet for some reason)
|
|
|
|
// to get to that number.
|
|
|
|
const newKeyCount = keyLimit - unpublishedOTKCount - this._serverOTKCount;
|
|
|
|
if (newKeyCount > 0) {
|
|
|
|
await log.wrap("generate otks", log => {
|
2021-02-23 19:22:25 +01:00
|
|
|
log.set("max", maxOTKs);
|
|
|
|
log.set("server", this._serverOTKCount);
|
|
|
|
log.set("unpublished", unpublishedOTKCount);
|
|
|
|
log.set("new", newKeyCount);
|
2021-03-09 12:33:31 +01:00
|
|
|
log.set("limit", keyLimit);
|
2020-08-28 13:58:17 +02:00
|
|
|
this._account.generate_one_time_keys(newKeyCount);
|
2021-03-09 12:33:31 +01:00
|
|
|
this._updateSessionStorage(storage, sessionStore => {
|
|
|
|
sessionStore.set(ACCOUNT_SESSION_KEY, this._account.pickle(this._pickleKey));
|
|
|
|
});
|
|
|
|
});
|
2020-08-28 13:58:17 +02:00
|
|
|
}
|
2021-03-09 12:33:31 +01:00
|
|
|
// even though we didn't generate any keys, we still have some unpublished ones that should be published
|
|
|
|
return true;
|
2020-08-28 13:58:17 +02:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2020-09-01 17:59:39 +02:00
|
|
|
createInboundOlmSession(senderKey, body) {
|
|
|
|
const newSession = new this._olm.Session();
|
2020-09-02 17:37:13 +02:00
|
|
|
try {
|
|
|
|
newSession.create_inbound_from(this._account, senderKey, body);
|
|
|
|
return newSession;
|
|
|
|
} catch (err) {
|
|
|
|
newSession.free();
|
|
|
|
throw err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-11-10 11:04:09 +01:00
|
|
|
async createOutboundOlmSession(theirIdentityKey, theirOneTimeKey) {
|
2020-09-02 17:37:13 +02:00
|
|
|
const newSession = new this._olm.Session();
|
|
|
|
try {
|
2020-11-10 11:04:09 +01:00
|
|
|
if (this._olmWorker) {
|
|
|
|
await this._olmWorker.createOutboundOlmSession(this._account, newSession, theirIdentityKey, theirOneTimeKey);
|
|
|
|
} else {
|
|
|
|
newSession.create_outbound(this._account, theirIdentityKey, theirOneTimeKey);
|
|
|
|
}
|
2020-09-02 17:37:13 +02:00
|
|
|
return newSession;
|
|
|
|
} catch (err) {
|
|
|
|
newSession.free();
|
|
|
|
throw err;
|
|
|
|
}
|
2020-09-01 17:59:39 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
writeRemoveOneTimeKey(session, txn) {
|
|
|
|
// this is side-effecty and will have applied the change if the txn fails,
|
|
|
|
// but don't want to clone the account for now
|
|
|
|
// and it is not the worst thing to think we have used a OTK when
|
|
|
|
// decrypting the message that actually used it threw for some reason.
|
|
|
|
this._account.remove_one_time_keys(session);
|
|
|
|
txn.session.set(ACCOUNT_SESSION_KEY, this._account.pickle(this._pickleKey));
|
|
|
|
}
|
|
|
|
|
2021-02-17 18:45:04 +01:00
|
|
|
writeSync(deviceOneTimeKeysCount, txn, log) {
|
2020-08-28 13:56:44 +02:00
|
|
|
// we only upload signed_curve25519 otks
|
2022-01-05 14:26:15 +01:00
|
|
|
const otkCount = deviceOneTimeKeysCount.signed_curve25519;
|
2020-08-28 13:56:44 +02:00
|
|
|
if (Number.isSafeInteger(otkCount) && otkCount !== this._serverOTKCount) {
|
|
|
|
txn.session.set(SERVER_OTK_COUNT_SESSION_KEY, otkCount);
|
2021-02-17 18:45:04 +01:00
|
|
|
log.set("otkCount", otkCount);
|
2020-08-28 13:56:44 +02:00
|
|
|
return otkCount;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
afterSync(otkCount) {
|
|
|
|
// could also be undefined
|
|
|
|
if (Number.isSafeInteger(otkCount)) {
|
|
|
|
this._serverOTKCount = otkCount;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-14 16:25:32 +01:00
|
|
|
_keysAsSignableObject(identityKeys) {
|
2020-08-27 19:13:24 +02:00
|
|
|
const obj = {
|
|
|
|
user_id: this._userId,
|
2021-10-27 18:08:50 +02:00
|
|
|
device_id: this._deviceId,
|
2020-08-28 13:51:58 +02:00
|
|
|
algorithms: [OLM_ALGORITHM, MEGOLM_ALGORITHM],
|
2020-08-27 19:13:24 +02:00
|
|
|
keys: {}
|
|
|
|
};
|
|
|
|
for (const [algorithm, pubKey] of Object.entries(identityKeys)) {
|
|
|
|
obj.keys[`${algorithm}:${this._deviceId}`] = pubKey;
|
|
|
|
}
|
2023-02-14 16:25:32 +01:00
|
|
|
return obj;
|
|
|
|
}
|
|
|
|
|
2023-03-02 15:02:42 +01:00
|
|
|
getUnsignedDeviceKey() {
|
2023-02-14 16:25:32 +01:00
|
|
|
const identityKeys = JSON.parse(this._account.identity_keys());
|
|
|
|
return this._keysAsSignableObject(identityKeys);
|
|
|
|
}
|
|
|
|
|
|
|
|
_deviceKeysPayload(identityKeys) {
|
|
|
|
const obj = this._keysAsSignableObject(identityKeys);
|
2020-08-27 19:13:24 +02:00
|
|
|
this.signObject(obj);
|
|
|
|
return obj;
|
|
|
|
}
|
|
|
|
|
|
|
|
_oneTimeKeysPayload(oneTimeKeysEntries) {
|
|
|
|
const obj = {};
|
|
|
|
for (const [keyId, pubKey] of oneTimeKeysEntries) {
|
|
|
|
const keyObj = {
|
|
|
|
key: pubKey
|
|
|
|
};
|
|
|
|
this.signObject(keyObj);
|
|
|
|
obj[`signed_curve25519:${keyId}`] = keyObj;
|
|
|
|
}
|
|
|
|
return obj;
|
|
|
|
}
|
|
|
|
|
|
|
|
async _updateSessionStorage(storage, callback) {
|
2021-10-26 18:47:46 +02:00
|
|
|
if (storage) {
|
|
|
|
const txn = await storage.readWriteTxn([
|
|
|
|
storage.storeNames.session
|
|
|
|
]);
|
|
|
|
try {
|
|
|
|
await callback(txn.session);
|
|
|
|
} catch (err) {
|
|
|
|
txn.abort();
|
|
|
|
throw err;
|
|
|
|
}
|
|
|
|
await txn.complete();
|
|
|
|
} else {
|
|
|
|
await callback(undefined);
|
2020-08-27 19:13:24 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
signObject(obj) {
|
|
|
|
const sigs = obj.signatures || {};
|
|
|
|
const unsigned = obj.unsigned;
|
|
|
|
|
|
|
|
delete obj.signatures;
|
|
|
|
delete obj.unsigned;
|
|
|
|
|
|
|
|
sigs[this._userId] = sigs[this._userId] || {};
|
|
|
|
sigs[this._userId]["ed25519:" + this._deviceId] =
|
|
|
|
this._account.sign(anotherjson.stringify(obj));
|
|
|
|
obj.signatures = sigs;
|
|
|
|
if (unsigned !== undefined) {
|
|
|
|
obj.unsigned = unsigned;
|
|
|
|
}
|
|
|
|
}
|
2021-10-26 18:47:46 +02:00
|
|
|
|
|
|
|
pickleWithKey(key) {
|
|
|
|
return this._account.pickle(key);
|
|
|
|
}
|
|
|
|
|
|
|
|
dispose() {
|
|
|
|
this._account.free();
|
2021-10-28 11:48:25 +02:00
|
|
|
this._account = undefined;
|
2021-10-26 18:47:46 +02:00
|
|
|
}
|
2020-08-27 19:12:06 +02:00
|
|
|
}
|