/* * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #ifndef WEBRTC_MEDIA_BASE_FAKEVIDEORENDERER_H_ #define WEBRTC_MEDIA_BASE_FAKEVIDEORENDERER_H_ #include "webrtc/base/logging.h" #include "webrtc/base/sigslot.h" #include "webrtc/media/base/videoframe.h" #include "webrtc/media/base/videosinkinterface.h" namespace cricket { // Faked video renderer that has a callback for actions on rendering. class FakeVideoRenderer : public rtc::VideoSinkInterface { public: FakeVideoRenderer() : errors_(0), width_(0), height_(0), rotation_(webrtc::kVideoRotation_0), timestamp_(0), num_rendered_frames_(0), black_frame_(false) {} virtual void OnFrame(const VideoFrame& frame) { rtc::CritScope cs(&crit_); // TODO(zhurunz) Check with VP8 team to see if we can remove this // tolerance on Y values. Some unit tests produce Y values close // to 16 rather than close to zero, for supposedly black frames. // Largest value observed is 34, e.g., running // P2PTestConductor.LocalP2PTest16To9 (peerconnection_unittests). black_frame_ = CheckFrameColorYuv(0, 48, 128, 128, 128, 128, &frame); // Treat unexpected frame size as error. ++num_rendered_frames_; width_ = frame.width(); height_ = frame.height(); rotation_ = frame.rotation(); timestamp_ = frame.GetTimeStamp(); SignalRenderFrame(&frame); } int errors() const { return errors_; } int width() const { rtc::CritScope cs(&crit_); return width_; } int height() const { rtc::CritScope cs(&crit_); return height_; } webrtc::VideoRotation rotation() const { rtc::CritScope cs(&crit_); return rotation_; } int64_t timestamp() const { rtc::CritScope cs(&crit_); return timestamp_; } int num_rendered_frames() const { rtc::CritScope cs(&crit_); return num_rendered_frames_; } bool black_frame() const { rtc::CritScope cs(&crit_); return black_frame_; } sigslot::signal3 SignalSetSize; sigslot::signal1 SignalRenderFrame; private: static bool CheckFrameColorYuv(uint8_t y_min, uint8_t y_max, uint8_t u_min, uint8_t u_max, uint8_t v_min, uint8_t v_max, const cricket::VideoFrame* frame) { if (!frame || !frame->video_frame_buffer()) { return false; } // Y int y_width = frame->width(); int y_height = frame->height(); const uint8_t* y_plane = frame->video_frame_buffer()->DataY(); const uint8_t* y_pos = y_plane; int32_t y_pitch = frame->video_frame_buffer()->StrideY(); for (int i = 0; i < y_height; ++i) { for (int j = 0; j < y_width; ++j) { uint8_t y_value = *(y_pos + j); if (y_value < y_min || y_value > y_max) { return false; } } y_pos += y_pitch; } // U and V int chroma_width = (frame->width() + 1)/2; int chroma_height = (frame->height() + 1)/2; const uint8_t* u_plane = frame->video_frame_buffer()->DataU(); const uint8_t* v_plane = frame->video_frame_buffer()->DataV(); const uint8_t* u_pos = u_plane; const uint8_t* v_pos = v_plane; int32_t u_pitch = frame->video_frame_buffer()->StrideU(); int32_t v_pitch = frame->video_frame_buffer()->StrideV(); for (int i = 0; i < chroma_height; ++i) { for (int j = 0; j < chroma_width; ++j) { uint8_t u_value = *(u_pos + j); if (u_value < u_min || u_value > u_max) { return false; } uint8_t v_value = *(v_pos + j); if (v_value < v_min || v_value > v_max) { return false; } } u_pos += u_pitch; v_pos += v_pitch; } return true; } int errors_; int width_; int height_; webrtc::VideoRotation rotation_; int64_t timestamp_; int num_rendered_frames_; bool black_frame_; rtc::CriticalSection crit_; }; } // namespace cricket #endif // WEBRTC_MEDIA_BASE_FAKEVIDEORENDERER_H_