1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
gpu / command_buffer / common / command_buffer_shared.h [blame]
// Copyright 2012 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifdef UNSAFE_BUFFERS_BUILD
// TODO(crbug.com/351564777): Remove this and convert code to safer constructs.
#pragma allow_unsafe_buffers
#endif
#ifndef GPU_COMMAND_BUFFER_COMMON_COMMAND_BUFFER_SHARED_H_
#define GPU_COMMAND_BUFFER_COMMON_COMMAND_BUFFER_SHARED_H_
#include <atomic>
#include "command_buffer.h"
#include "base/atomicops.h"
namespace gpu {
// This is a standard 4-slot asynchronous communication mechanism, used to
// ensure that the reader gets a consistent copy of what the writer wrote.
template<typename T>
class SharedState {
T states_[2][2];
base::subtle::Atomic32 reading_;
base::subtle::Atomic32 latest_;
base::subtle::Atomic32 slots_[2];
public:
void Initialize() {
for (int i = 0; i < 2; ++i) {
for (int j = 0; j < 2; ++j) {
states_[i][j] = T();
}
}
base::subtle::NoBarrier_Store(&reading_, 0);
base::subtle::NoBarrier_Store(&latest_, 0);
base::subtle::NoBarrier_Store(&slots_[0], 0);
base::subtle::Release_Store(&slots_[1], 0);
std::atomic_thread_fence(std::memory_order_seq_cst);
}
void Write(const T& state) {
int towrite = !base::subtle::Acquire_Load(&reading_);
int index = !base::subtle::Acquire_Load(&slots_[towrite]);
states_[towrite][index] = state;
base::subtle::Release_Store(&slots_[towrite], index);
base::subtle::Release_Store(&latest_, towrite);
std::atomic_thread_fence(std::memory_order_seq_cst);
}
// Attempt to update the state, updating only if the generation counter is
// newer.
void Read(T* state) {
std::atomic_thread_fence(std::memory_order_seq_cst);
int toread = !!base::subtle::Acquire_Load(&latest_);
base::subtle::Release_Store(&reading_, toread);
std::atomic_thread_fence(std::memory_order_seq_cst);
int index = !!base::subtle::Acquire_Load(&slots_[toread]);
if (states_[toread][index].generation - state->generation < 0x80000000U)
*state = states_[toread][index];
}
};
typedef SharedState<CommandBuffer::State> CommandBufferSharedState;
} // namespace gpu
#endif // GPU_COMMAND_BUFFER_COMMON_COMMAND_BUFFER_SHARED_H_