synapse-old/synapse/rest/media/v1/media_storage.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

376 lines
13 KiB
Python
Raw Normal View History

# Copyright 2018-2021 The Matrix.org Foundation C.I.C.
2018-01-08 10:07:30 -07:00
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
2018-07-09 00:09:20 -06:00
import contextlib
import logging
import os
import shutil
from types import TracebackType
from typing import (
IO,
TYPE_CHECKING,
Any,
Awaitable,
BinaryIO,
Callable,
Generator,
Optional,
Sequence,
Tuple,
Type,
)
import attr
from twisted.internet.defer import Deferred
from twisted.internet.interfaces import IConsumer
2018-07-09 00:09:20 -06:00
from twisted.protocols.basic import FileSender
2018-01-08 10:07:30 -07:00
import synapse
from synapse.api.errors import NotFoundError
from synapse.logging.context import defer_to_thread, make_deferred_yieldable
from synapse.util import Clock
2018-01-17 09:56:23 -07:00
from synapse.util.file_consumer import BackgroundFileConsumer
2018-01-08 10:07:30 -07:00
from ._base import FileInfo, Responder
from .filepath import MediaFilePaths
if TYPE_CHECKING:
from synapse.server import HomeServer
from .storage_provider import StorageProviderWrapper
2018-01-17 09:56:23 -07:00
2018-01-08 10:07:30 -07:00
logger = logging.getLogger(__name__)
2020-09-04 04:54:56 -06:00
class MediaStorage:
2018-01-08 10:07:30 -07:00
"""Responsible for storing/fetching files from local sources.
2018-01-12 04:15:31 -07:00
Args:
hs
local_media_directory: Base path where we store media on disk
filepaths
storage_providers: List of StorageProvider that are used to fetch and store files.
2018-01-08 10:07:30 -07:00
"""
def __init__(
self,
hs: "HomeServer",
local_media_directory: str,
filepaths: MediaFilePaths,
storage_providers: Sequence["StorageProviderWrapper"],
):
self.hs = hs
self.reactor = hs.get_reactor()
2018-01-08 10:07:30 -07:00
self.local_media_directory = local_media_directory
self.filepaths = filepaths
2018-01-08 10:19:55 -07:00
self.storage_providers = storage_providers
self.spam_checker = hs.get_spam_checker()
self.clock = hs.get_clock()
2018-01-08 10:07:30 -07:00
async def store_file(self, source: IO, file_info: FileInfo) -> str:
2018-01-08 10:07:30 -07:00
"""Write `source` to the on disk media store, and also any other
configured storage providers
Args:
source: A file like object that should be written
file_info: Info about the file to store
2018-01-08 10:07:30 -07:00
Returns:
the file path written to in the primary media store
2018-01-08 10:07:30 -07:00
"""
2018-02-14 10:55:18 -07:00
with self.store_into_file(file_info) as (f, fname, finish_cb):
# Write to the main repository
await self.write_to_file(source, f)
await finish_cb()
return fname
2018-01-08 10:07:30 -07:00
async def write_to_file(self, source: IO, output: IO) -> None:
"""Asynchronously write the `source` to `output`."""
await defer_to_thread(self.reactor, _write_file_synchronously, source, output)
2018-01-08 10:07:30 -07:00
@contextlib.contextmanager
def store_into_file(
self, file_info: FileInfo
) -> Generator[Tuple[BinaryIO, str, Callable[[], Awaitable[None]]], None, None]:
2018-01-08 10:07:30 -07:00
"""Context manager used to get a file like object to write into, as
described by file_info.
2018-01-12 04:15:31 -07:00
Actually yields a 3-tuple (file, fname, finish_cb), where file is a file
like object that can be written to, fname is the absolute path of file
on disk, and finish_cb is a function that returns an awaitable.
2018-01-12 04:15:31 -07:00
fname can be used to read the contents from after upload, e.g. to
generate thumbnails.
finish_cb must be called and waited on after the file has been
successfully been written to. Should not be called if there was an
error.
2018-01-08 10:07:30 -07:00
Args:
file_info: Info about the file to store
2018-01-08 10:07:30 -07:00
Example:
with media_storage.store_into_file(info) as (f, fname, finish_cb):
# .. write into f ...
await finish_cb()
2018-01-08 10:07:30 -07:00
"""
path = self._file_info_to_path(file_info)
fname = os.path.join(self.local_media_directory, path)
dirname = os.path.dirname(fname)
os.makedirs(dirname, exist_ok=True)
2018-01-08 10:07:30 -07:00
finished_called = [False]
try:
with open(fname, "wb") as f:
async def finish() -> None:
# Ensure that all writes have been flushed and close the
# file.
f.flush()
f.close()
spam_check = await self.spam_checker.check_media_file_for_spam(
ReadableFileWrapper(self.clock, fname), file_info
)
if spam_check != synapse.module_api.NOT_SPAM:
logger.info("Blocking media due to spam checker")
# Note that we'll delete the stored media, due to the
# try/except below. The media also won't be stored in
# the DB.
# We currently ignore any additional field returned by
# the spam-check API.
raise SpamMediaException(errcode=spam_check[0])
for provider in self.storage_providers:
await provider.store_file(path, file_info)
finished_called[0] = True
2018-01-08 10:07:30 -07:00
yield f, fname, finish
except Exception as e:
2018-01-08 10:07:30 -07:00
try:
os.remove(fname)
except Exception:
pass
raise e from None
2018-01-08 10:07:30 -07:00
if not finished_called:
2018-01-12 04:15:31 -07:00
raise Exception("Finished callback not called")
2018-01-08 10:07:30 -07:00
async def fetch_media(self, file_info: FileInfo) -> Optional[Responder]:
2018-01-08 10:07:30 -07:00
"""Attempts to fetch media described by file_info from the local cache
and configured storage providers.
Args:
file_info
2018-01-08 10:07:30 -07:00
Returns:
Returns a Responder if the file was found, otherwise None.
2018-01-08 10:07:30 -07:00
"""
paths = [self._file_info_to_path(file_info)]
2018-01-08 10:07:30 -07:00
# fallback for remote thumbnails with no method in the filename
if file_info.thumbnail and file_info.server_name:
paths.append(
self.filepaths.remote_media_thumbnail_rel_legacy(
server_name=file_info.server_name,
file_id=file_info.file_id,
width=file_info.thumbnail.width,
height=file_info.thumbnail.height,
content_type=file_info.thumbnail.type,
)
)
for path in paths:
local_path = os.path.join(self.local_media_directory, path)
if os.path.exists(local_path):
logger.debug("responding with local file %s", local_path)
return FileResponder(open(local_path, "rb"))
logger.debug("local file %s did not exist", local_path)
2018-01-08 10:19:55 -07:00
for provider in self.storage_providers:
for path in paths:
res: Any = await provider.fetch(path, file_info)
if res:
logger.debug("Streaming %s from %s", path, provider)
return res
logger.debug("%s not found on %s", path, provider)
2018-01-08 10:19:55 -07:00
return None
2018-01-08 10:07:30 -07:00
async def ensure_media_is_in_local_cache(self, file_info: FileInfo) -> str:
"""Ensures that the given file is in the local cache. Attempts to
download it from storage providers if it isn't.
Args:
file_info
Returns:
Full path to local file
"""
path = self._file_info_to_path(file_info)
local_path = os.path.join(self.local_media_directory, path)
if os.path.exists(local_path):
return local_path
# Fallback for paths without method names
# Should be removed in the future
if file_info.thumbnail and file_info.server_name:
legacy_path = self.filepaths.remote_media_thumbnail_rel_legacy(
server_name=file_info.server_name,
file_id=file_info.file_id,
width=file_info.thumbnail.width,
height=file_info.thumbnail.height,
content_type=file_info.thumbnail.type,
)
legacy_local_path = os.path.join(self.local_media_directory, legacy_path)
if os.path.exists(legacy_local_path):
return legacy_local_path
2018-01-17 09:56:23 -07:00
dirname = os.path.dirname(local_path)
os.makedirs(dirname, exist_ok=True)
2018-01-17 09:56:23 -07:00
for provider in self.storage_providers:
res: Any = await provider.fetch(path, file_info)
if res:
with res:
consumer = BackgroundFileConsumer(
open(local_path, "wb"), self.reactor
2019-06-20 03:32:02 -06:00
)
await res.write_to_consumer(consumer)
await consumer.wait()
return local_path
raise NotFoundError()
def _file_info_to_path(self, file_info: FileInfo) -> str:
2018-01-08 10:07:30 -07:00
"""Converts file_info into a relative path.
2018-01-12 04:15:31 -07:00
The path is suitable for storing files under a directory, e.g. used to
store files on local FS under the base media repository directory.
2018-01-08 10:07:30 -07:00
"""
if file_info.url_cache:
if file_info.thumbnail:
return self.filepaths.url_cache_thumbnail_rel(
media_id=file_info.file_id,
width=file_info.thumbnail.width,
height=file_info.thumbnail.height,
content_type=file_info.thumbnail.type,
method=file_info.thumbnail.method,
)
2018-01-08 10:07:30 -07:00
return self.filepaths.url_cache_filepath_rel(file_info.file_id)
if file_info.server_name:
if file_info.thumbnail:
return self.filepaths.remote_media_thumbnail_rel(
server_name=file_info.server_name,
file_id=file_info.file_id,
width=file_info.thumbnail.width,
height=file_info.thumbnail.height,
content_type=file_info.thumbnail.type,
method=file_info.thumbnail.method,
2018-01-08 10:07:30 -07:00
)
return self.filepaths.remote_media_filepath_rel(
file_info.server_name, file_info.file_id
)
if file_info.thumbnail:
return self.filepaths.local_media_thumbnail_rel(
media_id=file_info.file_id,
width=file_info.thumbnail.width,
height=file_info.thumbnail.height,
content_type=file_info.thumbnail.type,
method=file_info.thumbnail.method,
2018-01-08 10:07:30 -07:00
)
return self.filepaths.local_media_filepath_rel(file_info.file_id)
def _write_file_synchronously(source: IO, dest: IO) -> None:
2018-02-14 10:55:18 -07:00
"""Write `source` to the file like `dest` synchronously. Should be called
2018-01-08 10:07:30 -07:00
from a thread.
Args:
2018-02-14 10:55:18 -07:00
source: A file like object that's to be written
dest: A file like object to be written to
2018-01-08 10:07:30 -07:00
"""
source.seek(0) # Ensure we read from the start of the file
2018-02-14 10:55:18 -07:00
shutil.copyfileobj(source, dest)
2018-01-08 10:07:30 -07:00
class FileResponder(Responder):
"""Wraps an open file that can be sent to a request.
Args:
open_file: A file like object to be streamed ot the client,
2018-01-08 10:07:30 -07:00
is closed when finished streaming.
"""
2019-06-20 03:32:02 -06:00
def __init__(self, open_file: IO):
2018-01-08 10:07:30 -07:00
self.open_file = open_file
def write_to_consumer(self, consumer: IConsumer) -> Deferred:
return make_deferred_yieldable(
FileSender().beginFileTransfer(self.open_file, consumer)
)
2018-01-08 10:07:30 -07:00
def __exit__(
self,
exc_type: Optional[Type[BaseException]],
exc_val: Optional[BaseException],
exc_tb: Optional[TracebackType],
) -> None:
2018-01-08 10:07:30 -07:00
self.open_file.close()
class SpamMediaException(NotFoundError):
"""The media was blocked by a spam checker, so we simply 404 the request (in
the same way as if it was quarantined).
"""
@attr.s(slots=True, auto_attribs=True)
class ReadableFileWrapper:
"""Wrapper that allows reading a file in chunks, yielding to the reactor,
and writing to a callback.
This is simplified `FileSender` that takes an IO object rather than an
`IConsumer`.
"""
CHUNK_SIZE = 2**14
clock: Clock
path: str
async def write_chunks_to(self, callback: Callable[[bytes], None]) -> None:
"""Reads the file in chunks and calls the callback with each chunk."""
with open(self.path, "rb") as file:
while True:
chunk = file.read(self.CHUNK_SIZE)
if not chunk:
break
callback(chunk)
# We yield to the reactor by sleeping for 0 seconds.
await self.clock.sleep(0)