angler-fishThe Vulnerability History Project

CVE-2015-1234

This vulnerability was a consequence of how the GPU and the native client application interact with each other. Essentially, they communicate using threads and shared memory to send commands and receive return values. The problem was in a specific block of code that accessed a specific resource in untrusted shared memory. There was a race condition where multiple threads were trying to access and/or modify the size of a shared resource, so if the timing between the threads was not perfect, it would be possible to access this resource, validate it, and then later modify it to cause a buffer overflow, which means that an attacker could potentially execute their own piece of code. So while there is a buffer overlow issue here, it is mostly the race condition that allows for this vulnerability.


I feel that this vulnerability was more of a design mistake, followed by maintainability issues. There is quite a bit of reliance on untrusted shared memory here because of the way that the different processes communicate with each other, and this is part of the architecture for this application, which means it is very difficult to change. This reliance, combined with the complexity that comes naturally with concurrent processes, exposes the application to a myriad of issues that may be very difficult to identify. Moreover, there were multiple code design flaws that overlooked the atomicity of certain operations in the code. In other words, by looking at the VCC and comparing it with the fix, it seemed like the original code was using the "easiest" path to retrieve data in shared memory, even when these operations were not atomic and introduced the possibility of race conditions. And since this code had been present for years before the fix, it propagated to other parts of the system. In fact, there was even some discussion in the bug report about volatile memory in other parts of the code that was not identified as such, so this vulnerability exposed other maintainability issues that required more time and effort from the developers. Finally, the fix seems to implement some proper ways to deal with the race condition issue, specifically the use of volatile type modifiers in the code to indicate the use of untrusted shared memory. In addition, the fix also introduces a better use of atomic operations by avoiding having to calculate, access, or write to shared memory more times than is necessary.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
CVE: CVE-2015-1234
CWE:
- 362
bugs:
- 468936
repo: https://chromium.googlesource.com/chromium/src
vccs:
- note: |
    They were adding a wrapper to support BOOL types. There is a specific line that
    was particularly responsible for the CVE: GLsizei num_values = result->GetNumResults();
  commit: 939e7367d5d984efe1f402c7e5135503beaebfae
fixes:
- note: |
    Main fix was to use a computed value instead of a value in shared memory. This ensures that
    the value being used can be trusted.
  commit: 181c7400b2bf50ba02ac77149749fb419b4d4797
bounty:
  date: 
  amount: 
  references: []
lessons:
  yagni:
    note: 
    applies: 
  question: |
    Are there any common lessons we have learned from class that apply to this
    vulnerability? In other words, could this vulnerability serve as an example
    of one of those lessons?

    Leave "applies" blank or put false if you did not see that lesson (you do
    not need to put a reason). Put "true" if you feel the lesson applies and put
    a quick explanation of how it applies.

    Don't feel the need to claim that ALL of these apply, but it's pretty likely
    that one or two of them apply.

    If you think of another lesson we covered in class that applies here, feel
    free to give it a small name and add one in the same format as these.
  serial_killer:
    note: 
    applies: 
  complex_inputs:
    note: 
    applies: 
  distrust_input:
    note: |
      I feel that this applies in terms of input between processes. The vulnerability was
      largely possible due to shared memory between processes, and the fix was mostly about
      validating limits and data that was transferred between processes.
    applies: true
  least_privilege:
    note: 
    applies: 
  native_wrappers:
    note: 
    applies: 
  defense_in_depth:
    note: 
    applies: 
  DoS_in_many_forms:
    note: |
      This lesson applies because a DoS was extremely likely to occur if the vulnerability was exploited.
      This would have happened through a race condition and overflow weakness, which is not the common way
      to produce a DoS attack.
    applies: true
  secure_by_default:
    note: 
    applies: 
  environment_variables:
    note: 
    applies: 
  security_by_obscurity:
    note: 
    applies: 
  frameworks_are_optional:
    note: 
    applies: 
reviews:
- 1016193003
- 1021803003
- 1034093002
upvotes: 5
mistakes:
  answer: |
    I feel that this vulnerability was more of a design mistake, followed by maintainability
    issues. There is quite a bit of reliance on untrusted shared memory here because of the
    way that the different processes communicate with each other, and this is part of the
    architecture for this application, which means it is very difficult to change. This
    reliance, combined with the complexity that comes naturally with concurrent processes,
    exposes the application to a myriad of issues that may be very difficult to identify.

    Moreover, there were multiple code design flaws that overlooked the atomicity of certain
    operations in the code. In other words, by looking at the VCC and comparing it with the
    fix, it seemed like the original code was using the "easiest" path to retrieve data in
    shared memory, even when these operations were not atomic and introduced the possibility
    of race conditions. And since this code had been present for years before the fix, it
    propagated to other parts of the system. In fact, there was even some discussion in the
    bug report about volatile memory in other parts of the code that was not identified as such,
    so this vulnerability exposed other maintainability issues that required more time and effort
    from the developers.

    Finally, the fix seems to implement some proper ways to deal with the race condition issue,
    specifically the use of volatile type modifiers in the code to indicate the use of untrusted
    shared memory. In addition, the fix also introduces a better use of atomic operations by avoiding
    having to calculate, access, or write to shared memory more times than is necessary.
  question: |
    In your opinion, after all of this research, what mistakes were made that
    led to this vulnerability? Coding mistakes? Design mistakes?
    Maintainability? Requirements? Miscommunications?

    Look at the CWE entry for this vulnerability and examine the mitigations
    they have written there. Are they doing those? Does the fix look proper?

    Use those questions to inspire your answer. Don't feel obligated to answer
    every one. Write a thoughtful entry here that those in the software
    engineering industry would find interesting.
announced: '2015-04-01'
subsystem:
  name: GPU
  answer: Based on the "Components" label in the bug report.
  question: |
    What subsystems was the mistake in?

    Look at the path of the source code files code that were fixed to get
    directory names. Look at comments in the code. Look at the bug reports how
    the bug report was tagged. Examples: "clipboard", "gpu", "ssl", "speech", "renderer"
discovered:
  date: '2015-03-19'
  answer: |
    Seems like a hacker with the alias of lokihardt found this vulnerability during a contest
    (evidence here https://securitytracker.com/id/1032012).
  google: false
  contest: Pwn2Own
  question: |
    How was this vulnerability discovered?

    Go to the bug report and read the conversation to find out how this was
    originally found. Answer in longform below in "answer", fill in the date in
    YYYY-MM-DD, and then determine if the vulnerability was found by a Google
    employee (you can tell from their email address). If it's clear that the
    vulenrability was discovered by a contest, fill in the name there.

    The "automated" flag can be true, false, or nil.
    The "google" flag can be true, false, or nil.

    If there is no evidence as to how this vulnerability was found, then you may
    leave the entries blank except for "answer". Write down where you looked in "answer".
  automated: false
description: |
  This vulnerability was a consequence of how the GPU and the native client application
  interact with each other. Essentially, they communicate using threads and shared
  memory to send commands and receive return values. The problem was in a specific
  block of code that accessed a specific resource in untrusted shared memory.

  There was a race condition where multiple threads were trying to access and/or
  modify the size of a shared resource, so if the timing between the threads was
  not perfect, it would be possible to access this resource, validate it, and then
  later modify it to cause a buffer overflow, which means that an attacker could
  potentially execute their own piece of code. So while there is a buffer overlow
  issue here, it is mostly the race condition that allows for this vulnerability.
unit_tested:
  fix: false
  code: true
  answer: |
    There are unit tests for the code in the file in question, but it seems like
    no tests were modified or added because of the vulnerability.
  question: |
    Were automated unit tests involved in this vulnerability?
    Was the original code unit tested, or not unit tested? Did the fix involve
    improving the automated tests?

    For the "code" answer below, look not only at the fix but the surrounding
    code near the fix and determine if and was there were unit tests involved
    for this module.

    For the "fix" answer below, check if the fix for the vulnerability involves
    adding or improving an automated test to ensure this doesn't happen again.
major_events:
  answer: |
    No specific event was found, but it seems like the file in question was being worked on by
    multiple people across some big companies (e.g. Intel, Samsung, Nvidia). Too many developers
    could have lead to multiple issues. For example, a lack of good communication between developers
    can increase the chances of introducing bugs or vulnerabilities, especially when the developers
    are spread across multiple companies.
  events:
  - date: 
    name: 
  - date: 
    name: 
  question: |
    Please record any major events you found in the history of this
    vulnerability. Was the code rewritten at some point? Was a nearby subsystem
    changed? Did the team change?

    The event doesn't need to be directly related to this vulnerability, rather,
    we want to capture what the development team was dealing with at the time.
curation_level: 1
CWE_instructions: |
  Please go to cwe.mitre.org and find the most specific, appropriate CWE entry
  that describes your vulnerability. (Tip: this may not be a good one to start
  with - spend time understanding this vulnerability before making your choice!)
bounty_instructions: |
  If you came across any indications that a bounty was paid out for this
  vulnerability, fill it out here. Or correct it if the information already here
  was wrong. Otherwise, leave it blank.
interesting_commits:
  answer: |
    A lot of the work leading up to the fix was related to buffers or to texture displays. It seems like the
    vulnerability was difficult to find and was fixed immediately after it was reported.
  commits:
  - note: |
      They were experiencing multiple minor bugs, so they preemptively added logging functionality
      to multiple files, including the file that introduced the vulnerability. While this is a
      safe and effective way to diagnose future issues, it seems like they were logging a bit too
      much, just by looking at the amount of log commands in each file within this commit.
    commit: d049874acef2be3c17612d4a06b480f3a45ea6e9
  - note: |
      They seemed to be doing a lot of work with GPU buffers, which is heavily related to the
      vulnerability. Even still, the vulnerability was only found years after the VCC was introduced.
      This leads me to believe that, even though the vulnerability was critical, it was mostly
      hidden behind a race condition that was not really being tested. While this vulnerability
      was easy to reproduce, it was very difficult to find.
    commit: 0140c18b7e341dd2c4e81f9323e830c69574cda1
  question: |
    Are there any interesting commits between your VCC(s) and fix(es)?

    Write a brief (under 100 words) description of why you think this commit was
    interesting in light of the lessons learned from this vulnerability. Any
    emerging themes?

    If there are no interesting commits, demonstrate that you completed this section by explaining what happened between the VCCs and the fix.
curated_instructions: |
  If you are manually editing this file, then you are "curating" it. Set the
  entry below to "true" as soon as you start. This will enable additional
  integrity checks on this file to make sure you fill everything out properly.
  If you are a student, we cannot accept your work as finished unless curated is
  set to true.
upvotes_instructions: |
  For the first round, ignore this upvotes number.

  For the second round of reviewing, you will be giving a certain amount of
  upvotes to each vulnerability you see. Your peers will tell you how
  interesting they think this vulnerability is, and you'll add that to the
  upvotes score on your branch.
announced_instructions: |
  Was there a date that this vulnerability was announced to the world? You can
  find this in changelogs, blogs, bug reports, or perhaps the CVE date. A good
  source for this is Chrome's Stable Release Channel
  (https://chromereleases.googleblog.com/).
  Please enter your date in YYYY-MM-DD format.
fixes_vcc_instructions: |
  Please put the commit hash in "commit" below (see my example in
  CVE-2011-3092.yml). Fixes and VCCs follow the same format.
description_instructions: |
  You can get an initial description from the CVE entry on cve.mitre.org. These
  descriptions are a fine start, but they can be kind of jargony.

  Rewrite this description in your own words. Make it interesting and easy to
  read to anyone with some programming experience. We can always pull up the NVD
  description later to get more technical.

  Try to still be specific in your description, but remove Chromium-specific
  stuff. Remove references to versions, specific filenames, and other jargon
  that outsiders to Chromium would not understand. Technology like "regular
  expressions" is fine, and security phrases like "invalid write" are fine to
  keep too.

See a mistake? Is something missing from our story? We welcome contributions! All of our work is open-source and version-controlled on GitHub. You can curate using our Curation Wizard.

Use our Curation Wizard

Or go to GitHub

  • There are no articles here... yet

Timeline

Hover over an event to see its title.
Click on the event to learn more.
Filter by event type with the buttons below.

expand_less