The problem here is that sequence embeddings should have tons of side-channels which should convey non-semantic information (like, say, frequencies of tokens in sequence) and you can come a long way with this sort of information.
What would be really interesting is to train embedding models in different languages and check whether you can translate highly metaphorical sentences with no correspondence other than semantic, or train embedding models on different representations of the same math (for example, matrix mechanics vs wave mechanics formulations of quantum mechanics) and see if they recognize equivalent theorems.
For your information, Ukraine seems to have attacked airfields in Murmansk and Irkutsk Oblast’s. It’s approximately 1800 and 4500 km from Ukraine border respectively. Suspected method of attack is drones, transported on truck.