1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
media / gpu / test / raw_video.cc [blame]
// Copyright 2023 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifdef UNSAFE_BUFFERS_BUILD
// TODO(crbug.com/40285824): Remove this and convert code to safer constructs.
#pragma allow_unsafe_buffers
#endif
#include "media/gpu/test/raw_video.h"
#include "base/files/file_util.h"
#include "base/files/memory_mapped_file.h"
#include "base/functional/bind.h"
#include "base/functional/callback_helpers.h"
#include "base/json/json_reader.h"
#include "base/logging.h"
#include "base/memory/ptr_util.h"
#include "base/memory/raw_ptr_exclusion.h"
#include "base/numerics/safe_conversions.h"
#include "base/synchronization/waitable_event.h"
#include "base/threading/thread.h"
#include "media/base/media.h"
#include "media/base/media_serializers.h"
#include "media/base/video_decoder_config.h"
#include "media/base/video_frame.h"
#include "media/base/video_types.h"
#include "media/ffmpeg/ffmpeg_common.h"
#include "media/ffmpeg/scoped_av_packet.h"
#include "media/filters/ffmpeg_glue.h"
#include "media/filters/in_memory_url_protocol.h"
#include "media/filters/offloading_video_decoder.h"
#include "media/filters/vpx_video_decoder.h"
#include "media/gpu/test/video_frame_helpers.h"
#include "media/parsers/vp9_parser.h"
#include "testing/gtest/include/gtest/gtest.h"
#include "third_party/ffmpeg/libavformat/avformat.h"
#include "third_party/ffmpeg/libavutil/avutil.h"
#include "third_party/libyuv/include/libyuv/convert.h"
#include "third_party/libyuv/include/libyuv/planar_functions.h"
namespace media::test {
namespace {
// Suffix appended to the video file path to get the metadata file path, if no
// explicit metadata file path was specified.
constexpr const base::FilePath::CharType* kMetadataSuffix =
FILE_PATH_LITERAL(".json");
std::unique_ptr<base::MemoryMappedFile> CreateMemoryMappedFile(size_t size) {
base::FilePath tmp_file_path;
if (!base::CreateTemporaryFile(&tmp_file_path)) {
LOG(ERROR) << "Failed to create a temporary file";
return nullptr;
}
auto mmapped_file = std::make_unique<base::MemoryMappedFile>();
bool success = mmapped_file->Initialize(
base::File(
tmp_file_path, base::File::FLAG_CREATE_ALWAYS |
base::File::FLAG_READ | base::File::FLAG_WRITE
// On Windows FLAG_CREATE_ALWAYS will require FLAG_WRITE, and FLAG_APPEND
// must not be specified.
#if BUILDFLAG(IS_CHROMEOS) || BUILDFLAG(IS_LINUX)
| base::File::FLAG_APPEND
#endif // BUILDFLAG(IS_CHROMEOS) || BUILDFLAG(IS_LINUX)
),
base::MemoryMappedFile::Region{0, size},
base::MemoryMappedFile::READ_WRITE_EXTEND);
base::DeleteFile(tmp_file_path);
return success ? std::move(mmapped_file) : nullptr;
}
std::unique_ptr<base::MemoryMappedFile> LoadRawData(
const base::FilePath& data_file_path,
size_t video_frame_size,
size_t num_read_frames) {
auto memory_mapped_file = std::make_unique<base::MemoryMappedFile>();
if (!memory_mapped_file->Initialize(
base::File(data_file_path,
base::File::FLAG_OPEN | base::File::FLAG_READ),
base::MemoryMappedFile::Region{0, video_frame_size * num_read_frames},
base::MemoryMappedFile::READ_ONLY)) {
LOG(ERROR) << "Failed to read the file: " << data_file_path;
return nullptr;
}
CHECK_EQ(memory_mapped_file->length(), video_frame_size * num_read_frames);
return memory_mapped_file;
}
} // namespace
class RawVideo::VP9Decoder {
public:
static std::unique_ptr<VP9Decoder> Create(
const base::FilePath& vp9_webm_data_file_path,
const VideoFrameLayout& layout,
size_t num_read_frames);
static std::unique_ptr<VP9Decoder> Duplicate(const VideoFrameLayout& layout,
const VP9Decoder& vp9_decoder);
std::vector<uint8_t> DecodeFrame(size_t index) {
// If this is the first decode, then starts the thread.
if (!decoder_thread_.IsRunning()) {
LOG_IF(FATAL, !decoder_thread_.Start())
<< "Failed to start decoder thread";
DecoderStatus result;
base::WaitableEvent event;
// base::Unretained(this) is safe because this is blocking call.
decoder_thread_.task_runner()->PostTask(
FROM_HERE, base::BindOnce(&VP9Decoder::InitializeTask,
base::Unretained(this), &result, &event));
event.Wait();
LOG_ASSERT(result.is_ok())
<< "Failed to initialize VpxVideoDecoder: "
<< MediaSerializeForTesting(result)
<< "with config=" << config_.AsHumanReadableString();
}
std::vector<uint8_t> decoded_frame_buffer;
base::WaitableEvent done;
decoder_thread_.task_runner()->PostTask(
FROM_HERE,
base::BindOnce(&VP9Decoder::DecodeFrameTask, base::Unretained(this),
index, &decoded_frame_buffer, &done));
done.Wait();
CHECK(!decoded_frame_buffer.empty());
return decoded_frame_buffer;
}
~VP9Decoder() {
if (decoder_thread_.IsRunning()) {
decoder_thread_.task_runner()->DeleteSoon(FROM_HERE,
std::move(vpx_decoder_));
}
}
private:
struct VP9Data : public base::RefCountedThreadSafe<VP9Data> {
VP9Data(std::unique_ptr<base::MemoryMappedFile> mmap_file,
const std::vector<base::span<const uint8_t>>& chunks,
const std::vector<size_t>& keyframe_indices)
: chunks(chunks),
keyframe_indices(keyframe_indices),
mmap_file_(std::move(mmap_file)) {}
// TODO(367764863) Rewrite to base::raw_span.
RAW_PTR_EXCLUSION const std::vector<base::span<const uint8_t>> chunks;
const std::vector<size_t> keyframe_indices;
protected:
friend class base::RefCountedThreadSafe<VP9Data>;
virtual ~VP9Data() = default;
private:
std::unique_ptr<base::MemoryMappedFile> mmap_file_;
};
VP9Decoder(scoped_refptr<VP9Data> vp9_data,
const VideoDecoderConfig& config,
const VideoFrameLayout& layout)
: vp9_data_(std::move(vp9_data)),
config_(config),
layout_(layout),
video_frame_size_(layout_.planes().back().offset +
layout_.planes().back().size),
decoder_thread_("VP9DecoderThread") {
DETACH_FROM_SEQUENCE(decoder_sequence_);
}
void OnFrameDecoded(scoped_refptr<VideoFrame> frame) {
DCHECK_CALLED_ON_VALID_SEQUENCE(decoder_sequence_);
last_decoded_frame_ = std::move(frame);
}
void InitializeTask(DecoderStatus* result, base::WaitableEvent* event) {
DCHECK_CALLED_ON_VALID_SEQUENCE(decoder_sequence_);
vpx_decoder_ = std::make_unique<VpxVideoDecoder>(
OffloadableVideoDecoder::OffloadState::kOffloaded);
vpx_decoder_->Initialize(
config_,
/*low_delay*/ false,
/*CdmContext*/ nullptr,
base::BindOnce([](DecoderStatus* save_to,
DecoderStatus save_from) { *save_to = save_from; },
result),
// base::Unretained(this) is safe because |vpx_decoder_| is owned by
// this.
base::BindRepeating(&VP9Decoder::OnFrameDecoded,
base::Unretained(this)),
/*waiting_cb=*/base::NullCallback());
if (result->is_ok()) {
std::fill(std::begin(cached_frame_indices_),
std::end(cached_frame_indices_), -1);
}
event->Signal();
}
void DecodeFrameTask(size_t target_index,
std::vector<uint8_t>* decoded_frame_buffer,
base::WaitableEvent* done) {
DCHECK_CALLED_ON_VALID_SEQUENCE(decoder_sequence_);
if (size_t cached_index = target_index % kNumCachedFrames;
cached_frame_indices_[cached_index] == target_index) {
*decoded_frame_buffer = cached_frames_[cached_index];
done->Signal();
return;
}
auto upper =
std::upper_bound(vp9_data_->keyframe_indices.begin(),
vp9_data_->keyframe_indices.end(), target_index);
const size_t next_keyframe_index =
upper == vp9_data_->keyframe_indices.end() ? vp9_data_->chunks.size()
: *upper;
const size_t before_keyframe_index = *(--upper);
DecoderStatus decode_status{DecoderStatus::Codes::kOk};
for (size_t i = before_keyframe_index; i < next_keyframe_index; ++i) {
base::span<const uint8_t> chunk = vp9_data_->chunks[i];
vpx_decoder_->Decode(
DecoderBuffer::CopyFrom(chunk),
base::BindOnce([](DecoderStatus* out_status,
DecoderStatus status) { *out_status = status; },
&decode_status));
LOG_ASSERT(decode_status.is_ok())
<< "Failed to decode the " << i
<< "-th vp9 chunk: " << MediaSerializeForTesting(decode_status);
LOG_ASSERT(!!last_decoded_frame_)
<< "|last_decoded_frame_| is not filled";
auto buffer = CreateBufferFromFrame(*last_decoded_frame_);
last_decoded_frame_.reset();
const size_t cached_index = i % kNumCachedFrames;
cached_frame_indices_[cached_index] = i;
cached_frames_[cached_index] = std::move(buffer);
if (i == target_index) {
*decoded_frame_buffer = cached_frames_[cached_index];
done->Signal();
}
}
}
std::vector<uint8_t> CreateBufferFromFrame(
const VideoFrame& i420_frame) const {
LOG_ASSERT(i420_frame.format() == VideoPixelFormat::PIXEL_FORMAT_I420);
std::vector<uint8_t> buffer(video_frame_size_);
if (layout_.format() == PIXEL_FORMAT_NV12) {
uint8_t* nv12_frame = buffer.data();
int ret = libyuv::I420ToNV12(
i420_frame.data(0), i420_frame.stride(0), i420_frame.data(1),
i420_frame.stride(1), i420_frame.data(2), i420_frame.stride(2),
nv12_frame + layout_.planes()[0].offset, layout_.planes()[0].stride,
nv12_frame + layout_.planes()[1].offset, layout_.planes()[1].stride,
layout_.coded_size().width(), layout_.coded_size().height());
LOG_ASSERT(ret == 0) << "Failed converting from I420 to NV12";
} else {
CHECK_EQ(layout_.format(), PIXEL_FORMAT_I420);
// Copy the resolution area.
uint8_t* dst_plane = buffer.data();
for (size_t plane = 0; plane < 3; ++plane) {
const int stride = i420_frame.stride(plane);
const int rows = VideoFrame::Rows(plane, i420_frame.format(),
layout_.coded_size().height());
const int row_bytes = VideoFrame::RowBytes(
plane, i420_frame.format(), layout_.coded_size().width());
// VideoFrame::PlaneSize() cannot be used because it computes the
// plane size with resolutions aligned by two while our test code
// works with a succinct buffer size.
const uint8_t* src = i420_frame.data(plane);
libyuv::CopyPlane(src, stride, dst_plane, row_bytes, row_bytes, rows);
dst_plane += (rows * row_bytes);
}
}
return buffer;
}
const scoped_refptr<VP9Data> vp9_data_;
const VideoDecoderConfig config_;
const VideoFrameLayout layout_;
const size_t video_frame_size_;
base::Thread decoder_thread_;
std::unique_ptr<VpxVideoDecoder> vpx_decoder_
GUARDED_BY_CONTEXT(decoder_sequence_);
scoped_refptr<VideoFrame> last_decoded_frame_
GUARDED_BY_CONTEXT(decoder_sequence_);
// frame_index -> file index
static constexpr size_t kNumCachedFrames = 30;
size_t cached_frame_indices_[kNumCachedFrames];
std::vector<uint8_t> cached_frames_[kNumCachedFrames];
SEQUENCE_CHECKER(decoder_sequence_);
};
// static
std::unique_ptr<RawVideo::VP9Decoder> RawVideo::VP9Decoder::Create(
const base::FilePath& vp9_webm_data_file_path,
const VideoFrameLayout& layout,
size_t num_read_frames) {
base::MemoryMappedFile vp9_webm_data_mmap_file;
if (!vp9_webm_data_mmap_file.Initialize(vp9_webm_data_file_path,
base::MemoryMappedFile::READ_ONLY)) {
LOG(ERROR) << "Failed to read file: " << vp9_webm_data_file_path;
return nullptr;
}
base::span<const uint8_t> vp9_webm_data(vp9_webm_data_mmap_file.data(),
vp9_webm_data_mmap_file.length());
InitializeMediaLibrary();
// Initialize ffmpeg with the compressed video data.
InMemoryUrlProtocol protocol(vp9_webm_data.data(), vp9_webm_data.size(),
/*streaming=*/false);
FFmpegGlue glue(&protocol);
LOG_ASSERT(glue.OpenContext()) << "Failed to open AVFormatContext";
// Find the first VP9 stream in the file.
std::optional<size_t> vp9_stream_index;
VideoDecoderConfig config;
for (size_t i = 0; i < glue.format_context()->nb_streams; ++i) {
AVStream* stream = glue.format_context()->streams[i];
const AVCodecParameters* codec_parameters = stream->codecpar;
const AVMediaType codec_type = codec_parameters->codec_type;
const AVCodecID codec_id = codec_parameters->codec_id;
if (codec_type == AVMEDIA_TYPE_VIDEO && codec_id == AV_CODEC_ID_VP9 &&
AVStreamToVideoDecoderConfig(stream, &config) &&
config.IsValidConfig()) {
vp9_stream_index = i;
break;
}
}
if (!vp9_stream_index) {
return nullptr;
}
auto vp9_data_mmap_file = CreateMemoryMappedFile(vp9_webm_data.size());
uint8_t* const vp9_data = vp9_data_mmap_file->data();
size_t vp9_data_size = 0;
auto packet = ScopedAVPacket::Allocate();
size_t num_packets = 0;
Vp9Parser vp9_parser(/*parsing_compressed_header=*/false);
std::vector<size_t> keyframe_indices;
std::vector<base::span<const uint8_t>> vp9_data_chunks(num_read_frames);
while (av_read_frame(glue.format_context(), packet.get()) >= 0 &&
num_packets < num_read_frames) {
if (base::checked_cast<size_t>(packet->stream_index) ==
(*vp9_stream_index)) {
LOG_ASSERT(vp9_data_size + packet->size <= vp9_data_mmap_file->length())
<< "The vp9 data size must be less than webm file size";
std::memcpy(vp9_data + vp9_data_size, packet->data, packet->size);
vp9_data_chunks[num_packets] = base::span<const uint8_t>(
vp9_data + vp9_data_size, base::checked_cast<size_t>(packet->size));
vp9_data_size += packet->size;
Vp9FrameHeader header;
gfx::Size allocate_size;
vp9_parser.SetStream(packet->data, packet->size,
/*stream_config=*/nullptr);
if (vp9_parser.ParseNextFrame(&header, &allocate_size, nullptr) ==
Vp9Parser::kInvalidStream) {
LOG(ERROR) << "Failed parsing vp9 data";
return nullptr;
}
if (header.IsKeyframe()) {
keyframe_indices.push_back(num_packets);
}
num_packets++;
}
av_packet_unref(packet.get());
}
CHECK_EQ(num_read_frames, num_packets);
CHECK(!keyframe_indices.empty());
CHECK_EQ(keyframe_indices[0], 0u);
CHECK(std::is_sorted(keyframe_indices.begin(), keyframe_indices.end()));
if (num_read_frames > kLimitedReadFrames &&
(num_read_frames / keyframe_indices.size()) > 10) {
LOG(WARNING) << "Getting video frames by decoding on demand takes a long "
<< "time for a video with infrequent keyframes";
return nullptr;
}
return base::WrapUnique<VP9Decoder>(new VP9Decoder(
base::MakeRefCounted<VP9Data>(std::move(vp9_data_mmap_file),
vp9_data_chunks, keyframe_indices),
config, layout));
}
// static
std::unique_ptr<RawVideo::VP9Decoder> RawVideo::VP9Decoder::Duplicate(
const VideoFrameLayout& layout,
const RawVideo::VP9Decoder& vp9_decoder) {
return base::WrapUnique(
new VP9Decoder(vp9_decoder.vp9_data_, vp9_decoder.config_, layout));
}
RawVideo::RawVideo(std::unique_ptr<VP9Decoder> vp9_decoder,
const Metadata& metadata,
size_t video_frame_size)
: vp9_decoder_(std::move(vp9_decoder)),
metadata_(metadata),
video_frame_size_(video_frame_size) {}
RawVideo::RawVideo(std::unique_ptr<base::MemoryMappedFile> memory_mapped_file,
const Metadata& metadata,
size_t video_frame_size)
: memory_mapped_file_(std::move(memory_mapped_file)),
metadata_(metadata),
video_frame_size_(video_frame_size) {}
RawVideo::~RawVideo() = default;
RawVideo::Metadata::Metadata() = default;
RawVideo::Metadata::~Metadata() = default;
RawVideo::Metadata::Metadata(const Metadata&) = default;
RawVideo::Metadata& RawVideo::Metadata::operator=(const Metadata&) = default;
RawVideo::FrameData::FrameData(const std::vector<const uint8_t*>& plane_addrs,
const std::vector<size_t>& strides,
std::vector<uint8_t> buffer)
: plane_addrs(plane_addrs), strides(strides), buffer(std::move(buffer)) {}
RawVideo::FrameData::FrameData(FrameData&& frame_data)
: plane_addrs(frame_data.plane_addrs),
strides(frame_data.strides),
buffer(std::move(frame_data.buffer)) {}
RawVideo::FrameData::~FrameData() = default;
// Load the metadata from |json_file_path|. The read metadata is filled into
// |metadata| and compressed_data is set to true if the metadata denotes the
// |video| is vp9 video.
// static
bool RawVideo::LoadMetadata(const base::FilePath& json_file_path,
Metadata& metadata,
bool& is_vp9_data) {
std::string json_data;
if (!base::ReadFileToString(json_file_path, &json_data)) {
LOG(ERROR) << "Failed to read video metadata file: " << json_file_path;
return false;
}
auto metadata_result =
base::JSONReader::ReadAndReturnValueWithError(json_data);
if (!metadata_result.has_value()) {
LOG(ERROR) << "Failed to parse video metadata: " << json_file_path << ": "
<< metadata_result.error().message;
return false;
}
base::Value::Dict& metadata_dict = metadata_result->GetDict();
// The json must have either "profile" or "pixel_format".
// If it has "profile", then the data file is vp9 webm.
// If it has "pixel_format", then the data file is I420.
const std::string* profile = metadata_dict.FindString("profile");
const std::string* pixel_format = metadata_dict.FindString("pixel_format");
if (!!profile == !!pixel_format) {
LOG(ERROR) << "Metadata file must have either profile or pixel_format";
return false;
}
if (profile && *profile != "VP9PROFILE_PROFILE0") {
LOG(ERROR) << "The compressed video data file must be VP9 profile 0";
return false;
}
if (pixel_format && *pixel_format != "I420") {
LOG(ERROR) << "The raw video data file must be I420";
return false;
}
is_vp9_data = !!profile;
std::optional<int> frame_rate = metadata_dict.FindInt("frame_rate");
if (!frame_rate.has_value()) {
LOG(ERROR) << "Key \"frame_rate\" is not found in " << json_file_path;
return false;
}
metadata.frame_rate = base::checked_cast<uint32_t>(*frame_rate);
std::optional<int> num_frames = metadata_dict.FindInt("num_frames");
if (!num_frames.has_value()) {
LOG(ERROR) << "Key \"num_frames\" is not found in " << json_file_path;
return false;
}
metadata.num_frames = base::checked_cast<size_t>(*num_frames);
std::optional<int> width = metadata_dict.FindInt("width");
if (!width.has_value()) {
LOG(ERROR) << "Key \"width\" is not found in " << json_file_path;
return false;
}
std::optional<int> height = metadata_dict.FindInt("height");
if (!height) {
LOG(ERROR) << "Key \"height\" is not found in " << json_file_path;
return false;
}
const gfx::Size resolution(static_cast<uint32_t>(*width),
static_cast<uint32_t>(*height));
metadata.frame_layout =
CreateVideoFrameLayout(PIXEL_FORMAT_I420, resolution, 1u /* alignment */);
// The default visible rectangle is (0, 0, |resolution_|). Expand() needs to
// be called to change the visible rectangle.
metadata.visible_rect = gfx::Rect(resolution);
return true;
}
// static
std::unique_ptr<RawVideo> RawVideo::Create(
const base::FilePath& file_path,
const base::FilePath& metadata_file_path,
bool read_all_frames) {
CHECK(!file_path.empty());
const base::FilePath data_file_path = ResolveFilePath(file_path);
if (data_file_path.empty()) {
LOG(ERROR) << "Video file not found: " << file_path;
return nullptr;
}
const base::FilePath json_file_path = ResolveFilePath(
metadata_file_path.empty() ? file_path.AddExtension(kMetadataSuffix)
: metadata_file_path);
if (json_file_path.empty()) {
LOG(ERROR) << "Metadata file not found: " << file_path;
return nullptr;
}
bool is_vp9_data;
RawVideo::Metadata metadata;
if (!LoadMetadata(json_file_path, metadata, is_vp9_data)) {
LOG(ERROR) << "Invalid metadata file: " << json_file_path;
return nullptr;
}
std::vector<size_t> plane_offsets;
size_t video_frame_size = 0;
constexpr VideoPixelFormat kPixelFormat = PIXEL_FORMAT_I420;
const gfx::Size& resolution = metadata.frame_layout->coded_size();
for (size_t i = 0; i < VideoFrame::NumPlanes(kPixelFormat); ++i) {
video_frame_size +=
VideoFrame::RowBytes(i, kPixelFormat, resolution.width()) *
VideoFrame::Rows(i, kPixelFormat, resolution.height());
}
LOG_ASSERT(video_frame_size ==
metadata.frame_layout->planes().back().offset +
metadata.frame_layout->planes().back().size)
<< " video frame size computed by media::VideoFrame is different from"
<< " one computed by media::VideoFrameLayout";
if (!read_all_frames && metadata.num_frames > kLimitedReadFrames) {
DLOG(WARNING) << "Limit video length to " << kLimitedReadFrames
<< " frames";
metadata.num_frames = kLimitedReadFrames;
}
std::unique_ptr<base::MemoryMappedFile> memory_mapped_file;
if (is_vp9_data) {
// If the given data is compressed video (i.e. vp9 webm), then we decode.
auto vp9_decoder = RawVideo::VP9Decoder::Create(
data_file_path, *metadata.frame_layout, metadata.num_frames);
CHECK(vp9_decoder);
if (metadata.num_frames > kLimitedReadFrames) {
return base::WrapUnique(
new RawVideo(std::move(vp9_decoder), metadata, video_frame_size));
}
memory_mapped_file =
CreateMemoryMappedFile(video_frame_size * metadata.num_frames);
for (size_t i = 0; i < metadata.num_frames; ++i) {
auto buffer = vp9_decoder->DecodeFrame(i);
memcpy(memory_mapped_file->data() + i * video_frame_size, buffer.data(),
buffer.size());
}
} else {
memory_mapped_file =
LoadRawData(data_file_path, video_frame_size, metadata.num_frames);
}
if (!memory_mapped_file) {
return nullptr;
}
return base::WrapUnique(
new RawVideo(std::move(memory_mapped_file), metadata, video_frame_size));
}
std::unique_ptr<RawVideo> RawVideo::CreateNV12Video() const {
LOG_ASSERT(FrameLayout().format() == PIXEL_FORMAT_I420)
<< "The pixel format of source video is not I420";
auto nv12_layout = CreateVideoFrameLayout(PIXEL_FORMAT_NV12, Resolution(),
1u /* alignment*/);
LOG_ASSERT(nv12_layout) << "Failed creating VideoFrameLayout";
Metadata new_metadata = metadata_;
new_metadata.frame_layout = nv12_layout;
if (vp9_decoder_) {
return base::WrapUnique(new RawVideo(
RawVideo::VP9Decoder::Duplicate(*nv12_layout, *vp9_decoder_),
new_metadata, video_frame_size_));
}
auto new_memory_mapped_file =
CreateMemoryMappedFile(NumFrames() * video_frame_size_);
LOG_ASSERT(new_memory_mapped_file) << "Failed creating memory mapped file";
for (size_t i = 0; i < NumFrames(); ++i) {
const FrameData i420_frame = GetFrame(i);
uint8_t* const nv12_frame =
new_memory_mapped_file->data() + i * video_frame_size_;
int ret =
libyuv::I420ToNV12(i420_frame.plane_addrs[0], i420_frame.strides[0],
i420_frame.plane_addrs[1], i420_frame.strides[1],
i420_frame.plane_addrs[2], i420_frame.strides[2],
nv12_frame + nv12_layout->planes()[0].offset,
nv12_layout->planes()[0].stride,
nv12_frame + nv12_layout->planes()[1].offset,
nv12_layout->planes()[1].stride,
Resolution().width(), Resolution().height());
LOG_ASSERT(ret == 0) << "Failed converting from I420 to NV12";
}
return base::WrapUnique(new RawVideo(std::move(new_memory_mapped_file),
new_metadata, video_frame_size_));
}
std::unique_ptr<RawVideo> RawVideo::CreateExpandedVideo(
const gfx::Size& resolution,
const gfx::Rect& visible_rect) const {
LOG_ASSERT(PixelFormat() == VideoPixelFormat::PIXEL_FORMAT_NV12)
<< "The pixel format of source video is not NV12";
LOG_ASSERT(visible_rect.size() == Resolution())
<< "The resolution is different from the copied-into area of visible "
<< "rectangle";
LOG_ASSERT(gfx::Rect(resolution).Contains(visible_rect))
<< "The resolution doesn't contain visible rectangle";
LOG_ASSERT(visible_rect.x() % 2 == 0 && visible_rect.y() % 2 == 0)
<< "An odd origin point is not supported";
LOG_ASSERT(!vp9_decoder_ && !!memory_mapped_file_);
const std::optional<VideoFrameLayout> dst_layout =
CreateVideoFrameLayout(PIXEL_FORMAT_NV12, resolution, 1u /* alignment*/);
LOG_ASSERT(dst_layout) << "Failed creating VideoFrameLayout";
const auto& dst_planes = dst_layout->planes();
auto compute_dst_visible_data_offset = [&dst_layout,
&visible_rect](size_t plane) {
const size_t stride = dst_layout->planes()[plane].stride;
const size_t bytes_per_pixel =
VideoFrame::BytesPerElement(dst_layout->format(), plane);
gfx::Point origin = visible_rect.origin();
LOG_ASSERT(dst_layout->format() == VideoPixelFormat::PIXEL_FORMAT_NV12)
<< "The pixel format of destination video is not NV12";
if (plane == 1) {
origin.SetPoint(origin.x() / 2, origin.y() / 2);
}
return stride * origin.y() + bytes_per_pixel * origin.x();
};
const size_t dst_y_visible_offset = compute_dst_visible_data_offset(0);
const size_t dst_uv_visible_offset = compute_dst_visible_data_offset(1);
const size_t new_video_frame_size =
dst_planes.back().offset + dst_planes.back().size;
auto new_memory_mapped_file =
CreateMemoryMappedFile(new_video_frame_size * NumFrames());
CHECK(new_memory_mapped_file);
for (size_t i = 0; i < NumFrames(); i++) {
uint8_t* const dst_frame =
new_memory_mapped_file->data() + (i * new_video_frame_size);
uint8_t* const dst_y_plane_visible_data =
dst_frame + dst_planes[0].offset + dst_y_visible_offset;
uint8_t* const dst_uv_plane_visible_data =
dst_frame + dst_planes[1].offset + dst_uv_visible_offset;
FrameData src_frame = GetFrame(i);
libyuv::NV12Copy(src_frame.plane_addrs[0], src_frame.strides[0],
src_frame.plane_addrs[1], src_frame.strides[1],
dst_y_plane_visible_data, dst_planes[1].stride,
dst_uv_plane_visible_data, dst_planes[1].stride,
visible_rect.width(), visible_rect.height());
}
Metadata new_metadata = metadata_;
new_metadata.frame_layout = *dst_layout;
new_metadata.visible_rect = visible_rect;
return base::WrapUnique(new RawVideo(std::move(new_memory_mapped_file),
new_metadata, new_video_frame_size));
}
RawVideo::FrameData RawVideo::GetFrame(size_t frame_index) const {
CHECK_LT(frame_index, NumFrames());
std::vector<uint8_t> buffer;
const uint8_t* frame_addr;
if (vp9_decoder_) {
buffer = vp9_decoder_->DecodeFrame(frame_index);
frame_addr = buffer.data();
} else {
frame_addr = memory_mapped_file_->data() + video_frame_size_ * frame_index;
}
const auto& plane_layouts = FrameLayout().planes();
const size_t num_planes = plane_layouts.size();
std::vector<const uint8_t*> plane_addrs(num_planes);
std::vector<size_t> strides(num_planes);
for (size_t i = 0; i < num_planes; ++i) {
plane_addrs[i] = frame_addr + plane_layouts[i].offset;
strides[i] = plane_layouts[i].stride;
}
return RawVideo::FrameData(plane_addrs, strides, std::move(buffer));
}
// static
base::FilePath RawVideo::test_data_path_;
// static
void RawVideo::SetTestDataPath(const base::FilePath& test_data_path) {
test_data_path_ = test_data_path;
}
// static
base::FilePath RawVideo::ResolveFilePath(const base::FilePath& file_path) {
base::FilePath resolved_path = file_path;
// Try to resolve the path into an absolute path. If the path doesn't exist,
// it might be relative to the test data dir.
if (!resolved_path.IsAbsolute()) {
resolved_path = base::MakeAbsoluteFilePath(
PathExists(resolved_path) ? resolved_path
: test_data_path_.Append(resolved_path));
}
return base::PathExists(resolved_path) ? resolved_path : base::FilePath();
}
} // namespace media::test