viewer -> lttvwindow name change
[lttv.git] / ltt / branches / poly / doc / developer / lttvwindow_events_delivery.txt
CommitLineData
6ea2aecb 1Linux Trace Toolkit
2
3Mathieu Desnoyers 17-05-2004
4
5
6This document explains how the lttvwindow API could process the event requests
7of the viewers, merging event requests and hook lists to benefit from the fact
8that process_traceset can call multiple hooks for the same event.
9
10First, we will explain the detailed process of event delivery in the current
11framework. We will then study its strengths and weaknesses.
12
13In a second time, a framework where the events requests are dealt by the main
14window with fine granularity will be described. We will then discussed the
15advantages and inconvenients over the first framework.
16
17
181. (Actual) Boundaryless event reading
19
20Actually, viewers request events in a time interval from the main window. They
21also specify a (not so) maximum number of events to be delivered. In fact, the
22number of events to read only gives a stop point, from where only events with
23the same timestamp will be delivered.
24
25Viewers register hooks themselves in the traceset context. When merging read
26requests in the main window, all hooks registered by viewers will be called for
27the union of all the read requests, because the main window has no control on
28hook registration.
29
30The main window calls process_traceset on its own for all the intervals
31requested by all the viewers. It must not duplicate a read of the same time
32interval : it could be very hard to filter by viewers. So, in order to achieve
33this, time requests are sorted by start time, and process_traceset is called for
34each time request. We keep the last event time between each read : if the start
35time of the next read is lower than the time reached, we continue the reading
36from the actual position.
37
38We deal with specific number of events requests (infinite end time) by
39garantying that, starting from the time start of the request, at least that
40number of events will be read. As we can't do it efficiently without interacting
41very closely with process_traceset, we always read the specified number of
42events requested starting from the current position when we answer to a request
43based on the number of events.
44
45The viewers have to filter events delivered by traceset reading, because they
46can be asked by another viewer for a totally (or partially) different time
47interval.
48
49
50Weaknesses
51
52- process_middle does not guarantee the number of events read
53
54First of all, a viewer that requests events to process_traceset has no garantee
55that it will get exactly what it asked for. For example, a direct call to
56traceset_middle for a specific number of events will delived _at least_ that
57quantity of events, plus the ones that have the same timestamp that the last one
58has.
59
60- Border effects
61
62Viewer's writers will have to deal with a lot of border effects caused by the
318585ee 63particularities of the reading. They will be required to select the information
64they need from their input by filtering.
6ea2aecb 65
318585ee 66- Lack of encapsulation and difficulty of testing
6ea2aecb 67
68The viewer's writer will have to take into account all the border effects caused
69by the interaction with other modules. This means that event if a viewer works
70well alone or with another viewer, it's possible that new bugs arises when a new
318585ee 71viewer comes around. So, even if a perfect testbench works well for a viewer, it
72does not confirm that no new bug will arise when another viewer is loaded at the
73same moment asking for different time intervals.
6ea2aecb 74
75
76- Duplication of the work
77
78Time based filters and counters of events will have to be implemented at the
79viewer's side, which is a duplication of the functionnalities that would
80normally be expected from the tracecontext API.
81
82- Lack of control over the data input
83
84As we expect module's writers to prefer to be as close as possible from the raw
85datas, making them interact with a lower level library that gives them a data
86input that they only control by further filtering of the input is not
87appropriated. We should expect some reluctancy from them about using this API
88because of this lack of control on the input.
89
90- Speed cost
91
92All hooks of all viewers will be called for all the time intervals. So, if we
93have a detailed events list and a control flow view, asking both for different
94time intervals, the detailed events list will have to filter all the events
95delivered originally to the control flow view. This can be a case occuring quite
96often.
97
98
99
100Strengths
101
102- Simple concatenation of time intervals at the main window level.
103
104Having the opportunity of delivering more events than necessary to the viewers
105means that we can concatenate time intervals and number of events requested
106fairly easily, while being hard to determine if some specific cases will be
318585ee 107wrong, in depth testing being impossible.
6ea2aecb 108
109- No duplication of the tracecontext API
110
111Viewers deal directly with the tracecontext API for registering hooks, removing
112a layer of encapsulation.
113
114
115
116
117
1182. (Proposed) Strict boundaries events reading
119
120The idea behind this method is to provide exactly the events requested by the
121viewers to them, no more, no less.
122
6ea2aecb 123It uses the new API for process traceset suggested in the document
124process_traceset_strict_boundaries.txt.
125
126It also means that the lttvwindow API will have to deal with viewer's hooks.
127Those will not be allowed to add them directly in the context. They will give
128them to the lttvwindow API, along with the time interval or the position and
129number of events. The lttvwindow API will have to take care of adding and
130removing hooks for the different time intervals requested. That means that hooks
131insertion and removal will be done between each traceset processing based on
132the time intervals and event positions related to each hook. We must therefore
133provide a simple interface for hooks passing between the viewers and the main
318585ee 134window, make them easier to manage from the main window. A modification to the
135LttvHooks type solves this problem.
6ea2aecb 136
137
138Architecture
139
140Added to the lttvwindow API :
141
142
8646cedb 143void lttvwindow_events_request
81f3eebf 144( Tab *tab,
145 const EventsRequest *events_request);
6ea2aecb 146
8646cedb 147void lttvwindow_events_request_remove_all
81f3eebf 148( Tab *tab,
149 gconstpointer viewer);
6ea2aecb 150
6ea2aecb 151
8646cedb 152Internal functions :
6ea2aecb 153
154- lttvwindow_process_pending_requests
155
318585ee 156
8646cedb 157Events Requests Removal
318585ee 158
8646cedb 159A new API function will be necessary to let viewers remove all event requests
160they have made previously. By allowing this, no more out of bound requests will
161be serviced : a viewer that sees its time interval changed before the first
162servicing is completed can clear its previous events requests and make a new
163one for the new interval needed, considering the finished chunks as completed
164area.
318585ee 165
8646cedb 166It is also very useful for dealing with the viewer destruction case : the viewer
167just has to remove its events requests from the main window before it gets
168destroyed.
318585ee 169
318585ee 170
8646cedb 171Permitted GTK Events Between Chunks
318585ee 172
8646cedb 173All GTK Events will be enabled between chunks. This is due to the fact that the
174background processing and a high priority request are seen as the same case.
175While a background processing is in progress, the whole graphical interface must
176be enabled.
318585ee 177
8646cedb 178We needed to deal with the coherence of background processing and diverse GTK
179events anyway. This algorithm provides a generalized way to deal with any type
180of request and any GTK events.
318585ee 181
318585ee 182
8646cedb 183Background Computation Request
318585ee 184
8646cedb 185The types of background computation that can be requested by a viewer : state
186computation (main window scope) or viewer specific background computation.
318585ee 187
6ea2aecb 188
206ea1f4 189A New "Redraw" Button
190
191It will be used to redraw the viewers entirely. It is useful to restart the
192servicing after a "stop" action.
193
194A New "Continue" Button
195
196It will tell the viewers to send requests for damaged areas. It is useful to
197complete the servicing after a "stop" action.
198
199
6ea2aecb 200
65bc0600 201Tab change
202
203If a tab change occurs, we still want to do background processing.
204Events requests must be stocked in a list located in the same scope than the
205traceset context. Right now, this is tab scope. All functions called from the
206request servicing function must _not_ use the current_tab concept, as it may
207change. The idle function must the take a tab, and not the main window, as
208parameter.
209
210If a tab is removed, its associated idle events requests servicing function must
211also be removed.
212
4a97e95b 213It now looks a lot more useful to give a Tab* to the viewer instead of a
214MainWindow*, as all the information needed by the viewer is located at the tab
215level. It will diminish the dependance upon the current tab concept.
216
65bc0600 217
218
219Idle function (lttvwindow_process_pending_requests)
220
221The idle function must return FALSE to be removed from the idle functions when
222no more events requests are pending. Otherwise, it returns TRUE. It will service
223requests until there is no more request left.
224
225
226
8646cedb 227Implementation
6ea2aecb 228
6ea2aecb 229
8646cedb 230- Type LttvHooks
6ea2aecb 231
8646cedb 232see hook_prio.txt
69381fc7 233
8646cedb 234The viewers will just have to pass hooks to the main window through this type,
235using the hook.h interface to manipulate it. Then, the main window will add
236them and remove them from the context to deliver exactly the events requested by
237each viewer through process traceset.
69381fc7 238
239
8646cedb 240- lttvwindow_events_request
69381fc7 241
e9057d65 242It adds the an EventsRequest struct to the list of events requests
8646cedb 243pending and registers a pending request for the next g_idle if none is
244registered. The viewer can access this structure during the read as its
245hook_data. Only the stop_flag can be changed by the viewer through the
246event hooks.
69381fc7 247
69381fc7 248typedef struct _EventsRequest {
63b8a718 249 gpointer owner; /* Owner of the request */
250 gpointer viewer_data; /* Unset : NULL */
69381fc7 251 gboolean servicing; /* service in progress: TRUE */
63b8a718 252 LttTime start_time;/* Unset : { G_MAXUINT, G_MAXUINT }*/
253 LttvTracesetContextPosition *start_position; /* Unset : NULL */
69381fc7 254 gboolean stop_flag; /* Continue:TRUE Stop:FALSE */
63b8a718 255 LttTime end_time;/* Unset : { G_MAXUINT, G_MAXUINT } */
69381fc7 256 guint num_events; /* Unset : G_MAXUINT */
63b8a718 257 LttvTracesetContextPosition *end_position; /* Unset : NULL */
69381fc7 258 LttvHooks *before_traceset; /* Unset : NULL */
259 LttvHooks *before_trace; /* Unset : NULL */
260 LttvHooks *before_tracefile;/* Unset : NULL */
261 LttvHooks *event; /* Unset : NULL */
262 LttvHooksById *event_by_id; /* Unset : NULL */
263 LttvHooks *after_tracefile; /* Unset : NULL */
264 LttvHooks *after_trace; /* Unset : NULL */
265 LttvHooks *after_traceset; /* Unset : NULL */
e9057d65 266 LttvHooks *before_request; /* Unset : NULL */
267 LttvHooks *after_request /* Unset : NULL */
69381fc7 268} EventsRequest;
269
8646cedb 270- lttvwindow_events_request_remove_all
271
63b8a718 272It removes all the events requests from the pool that has their "owner" field
273maching the owner pointer given as argument.
8646cedb 274
e9057d65 275It calls the traceset/trace/tracefile end hooks for each request removed if
276they are currently serviced.
8646cedb 277
278
279- lttvwindow_process_pending_requests
280
281This internal function gets called by g_idle, taking care of the pending
282requests. It is responsible for concatenation of time intervals and position
283requests. It does it with the following algorithm organizing process traceset
284calls. Here is the detailed description of the way it works :
285
286
287
288- Revised Events Requests Servicing Algorithm (v2)
289
69381fc7 290The reads are splitted in chunks. After a chunk is over, we want to check if
291there is a GTK Event pending and execute it. It can add or remove events
292requests from the event requests list. If it happens, we want to start over
e9057d65 293the algorithm from the beginning. The after traceset/trace/tracefile hooks are
63b8a718 294called after each chunk, and before traceset/trace/tracefile are
e9057d65 295called when the request processing resumes. Before and after request hooks are
296called respectively before and after the request processing.
69381fc7 297
69381fc7 298
299Data structures necessary :
300
301List of requests added to context : list_in
302List of requests not added to context : list_out
303
304Initial state :
305
306list_in : empty
307list_out : many events requests
308
309
63b8a718 310// NOT A. While (list_in !empty or list_out !empty) and !GTK Event pending
311
312We do this once, go back to GTK, then get called again...
313We remove the g_idle function when list_in and list_out are empty
314
69381fc7 315 1. If list_in is empty (need a seek)
316 1.1 Add requests to list_in
63b8a718 317 1.1.1 Find all time requests with lowest start time in list_out (ltime)
318 1.1.2 Find all position requests with lowest position in list_out (lpos)
319 1.1.3 If lpos.start time < ltime
69381fc7 320 - Add lpos to list_in, remove them from list_out
63b8a718 321 1.1.4 Else, (lpos.start time >= ltime)
69381fc7 322 - Add ltime to list_in, remove them from list_out
323 1.2 Seek
324 1.2.1 If first request in list_in is a time request
e6359327 325 - If first req in list_in start time != current time
326 - Seek to that time
69381fc7 327 1.2.2 Else, the first request in list_in is a position request
e6359327 328 - If first req in list_in pos != current pos
63b8a718 329 - seek to that position
e9057d65 330 1.3 Add hooks and call before request for all list_in members
69381fc7 331 1.3.1 If !servicing
e9057d65 332 - begin request hooks called
69381fc7 333 - servicing = TRUE
9087aaaa 334 1.3.2 call before chunk
69381fc7 335 1.3.3 events hooks added
336 2. Else, list_in is not empty, we continue a read
337 2.1 For each req of list_out
338 - if req.start time == current context time
339 - Add to list_in, remove from list_out
340 - If !servicing
9087aaaa 341 - Call before request
69381fc7 342 - servicing = TRUE
9087aaaa 343 - Call before chunk
69381fc7 344 - events hooks added
345 - if req.start position == current position
346 - Add to list_in, remove from list_out
347 - If !servicing
9087aaaa 348 - Call before request
69381fc7 349 - servicing = TRUE
9087aaaa 350 - Call before_chunk
69381fc7 351 - events hooks added
352
353 3. Find end criterions
354 3.1 End time
355 3.1.1 Find lowest end time in list_in
356 3.1.2 Find lowest start time in list_out (>= than current time*)
63b8a718 357 * To eliminate lower prio requests (not used)
69381fc7 358 3.1.3 Use lowest of both as end time
359 3.2 Number of events
360 3.2.1 Find lowest number of events in list_in
361 3.2.2 Use min(CHUNK_NUM_EVENTS, min num events in list_in) as num_events
362 3.3 End position
363 3.3.1 Find lowest end position in list_in
364 3.3.2 Find lowest start position in list_out (>= than current
365 position)
366 3.3.3 Use lowest of both as end position
367
368 4. Call process traceset middle
369 4.1 Call process traceset middle (Use end criterion found in 3)
370 * note : end criterion can also be viewer's hook returning TRUE
371 5. After process traceset middle
372 - if current context time > traceset.end time
373 - For each req in list_in
69381fc7 374 - Remove events hooks for req
9087aaaa 375 - Call end chunk for req
e9057d65 376 - Call end request for req
69381fc7 377 - remove req from list_in
378 5.1 For each req in list_in
379 - req.num -= count
380 - if req.num == 0
69381fc7 381 - Remove events hooks for req
9087aaaa 382 - Call end chunk for req
e9057d65 383 - Call end request for req
69381fc7 384 - remove req from list_in
385 - if current context time > req.end time
69381fc7 386 - Remove events hooks for req
9087aaaa 387 - Call end chunk for req
e9057d65 388 - Call end request for req
69381fc7 389 - remove req from list_in
390 - if req.end pos == current pos
69381fc7 391 - Remove events hooks for req
9087aaaa 392 - Call end_chunk for req
e9057d65 393 - Call end request for req
69381fc7 394 - remove req from list_in
395 - if req.stop_flag == TRUE
69381fc7 396 - Remove events hooks for req
9087aaaa 397 - Call end chunk for req
e9057d65 398 - Call end request for req
69381fc7 399 - remove req from list_in
69381fc7 400
9087aaaa 401B. Between each chunk (interrupted or not)
69381fc7 402 1. for each request in list_in
403 1.1. Use current postition as start position
404 1.2. Remove start time
9087aaaa 405 1.3. Call after_chunk_traceset
69381fc7 406 1.4. Remove event hooks
407 1.5. Put it back in list_out
69381fc7 408
409
410
411Notes :
412End criterions for process traceset middle :
413If the criterion is reached, event is out of boundaries and we return.
414Current time >= End time
415Event count > Number of events
416Current position >= End position
417Last hook list called returned TRUE
418
419The >= for position is necessary to make ensure consistency between start time
420requests and positions requests that happens to be at the exact same start time
421and position.
422
8646cedb 423
424
425
426Weaknesses
427
63b8a718 428- ?
8646cedb 429
430Strengths
431
432- Removes the need for filtering of information supplied to the viewers.
433
434- Viewers have a better control on their data input.
435
436- Solves all the weaknesses idenfied in the actual boundaryless traceset
437reading.
438
This page took 0.042884 seconds and 4 git commands to generate.