-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathdb_ssi4.py
979 lines (812 loc) · 35 KB
/
db_ssi4.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
import collections, random, sys, traceback
from dbsim import *
class EagerFailWrite(AbortTransaction): pass
class EagerFailWriteDep(AbortTransaction): pass
class EagerFailReadDep(AbortTransaction): pass
class EagerFailReadAntiDep(AbortTransaction): pass
class EarlyPrecommitFail(AbortTransaction): pass
class PrecommitFail(AbortTransaction): pass
class PrecommitBackedgeFail(AbortTransaction): pass
class ReadSkew(AbortTransaction):
'''Read and overwrote different versions'''
pass
class WriteSkew(AbortTransaction):
'''Overwrote a version created after my snapshot'''
def make_db(tracker, stats, db_size, tid_watch, rid_watch, read_only_opt,
version_reader, safe_snaps,
si_relax_writes, si_relax_reads, verbose, **extra_kwargs):
'''For purposes of Test #1, which examines T1 x:x T r:w T2, we mostly
have to worry about anti-dependency edges of the form (TA r:w TB):
• If TB commits < TA reads, then TA is aware of TB, but chooses
not to "see" the write for some reason.
o TA updates s(TA) with s(TB) at the time of the read.
o TB does nothing, TA arrives too late to affect p(TB).
• If TB overwrites < TA reads < TB commits, then TA is aware of TB
but is not not allowed to "see" the write.
o TA updates s(TA) with s(TB) at pre-commit iff TB has
committed by then.
o Either TA will set t(V)=TA at read (allowing TB to update
p(TB) at pre-commit if TA has committed), or TA will adjust
p(TB) directly at pre-commit.
• If TA reads < TB overwrites < TA commits, then TA is *not* aware
of TB at read time and will have to verify its read set during
pre-commit unless it gets some help from TB.
o If t(V) is inactive, TA sets t(V)=TA
o If t(V) is active, TA updates s(TA) at pre-commit iff TB has
already committed.
o TB updates s(TA) at pre-commit iff TA has not yet committed.
o TB updates r(V) at pre-commit if t(V) is active and
committed. TB also updates p(TB) using r(V) at pre-commit.
o Note that TA must still follow this protocol if it sees an
uncommitted overwrite, because that overwrite could abort
and some later write could still arrive in time to matter.
• If TA commits < TB overwrites, then p(TB) must account for
TA. However, we only care about the latest such TA, because
p(TB) is a maximum.
o TB does not impact s(TA). TA must make c(TA) available to TB
either by setting t(V) (at read) or r(V) (at pre-commit).
o TB updates p(TB) with both r(V) and c(t(V)), as appropriate,
at write time.
'''
q,e = not verbose, errlog
# interaction: read-only optimization only works under SI
read_only_opt &= not si_relax_reads
tx_sstamps = {1:1}
tx_pstamps = collections.defaultdict(lambda:0)
tx_reads = collections.defaultdict(dict)
tx_writes = collections.defaultdict(set)
safe_snapper = safe_snap_maker(tracker, safe_snaps)
'''Each version V is associated with several pieces of information:
- TID of its creator (the tracker manages this for us). Note that
the TID may belong to an "old" transaction that the system has
forgotten, in which case we won't be able to extract any further
information about it.
- Creation time. The tracker manages this also. Crucially, the
timestamp remains in the version even after the transaction that
creates it has been forgotten.
- Read stamp. The most recent commit stamp of any transaction that
read this version. Every reader must update this stamp on any
version that has not been overwritten before the read
commits. The read stamp is used to update p(T) of the
transaction that eventually overwrites this version.
- Active reader. Each version can store the TID of one transaction
that is currently reading it. A transaction does not have to
update the read stamp if it can leave its TID instead, reducing
the amount of work it performs at pre-commit. Any reader or
overwrite that comes later will update the read stamp with the
active reader's commit stamp. If an active reader is forgotten
before another transaction accesses the version, its commit time
is not available; in these cases, the current safe point is used
instead. This may cause some extra false positives in cycle
detection, but not many (it would require a chain of
anti-dependencies to span several epochs. If the loss of
precision is unacceptable, the active reader slot should be
eliminated, forcing all readers to set read stamps.
- NOTE: We do *NOT* track s(T) of V's creator. T will not be
forgotten until V dies, so any TR that is able to access V can
still look up s(T).
'''
'''Life cycle of a version:
Invisible: the creating transaction has not yet committed, nobody
else can see this version. No locks are necessary, but the creator
must record it in a write set.
Visible: the creating transaction has committed and no delete has
committed yet. Any read or delete that arrives before the next
safe point must account for c(xmin) in its p(T). A reader that
commits before any deletion cannot have a dangerous r:w conflict,
but must still acquire an SIREAD lock. If the reader delays lock
acquisigion until pre-commit, it can take over any existing SIREAD
lock (which by definition must be older) instead of acquiring a
new one. The lock can be released at the next safe point, and a
writer that arrives before that will use the reader's c(T) to
update its own p(T); a late-arriving writer will use the safe
point to update its p(T) instead.
Deleted: the version has been deleted, but some readers may still
see it. The writer can release the SIREAD lock (if any), and no
further SIREAD locks are necessary. The writer's identity is
embedded in the version as xmax, and any reader that follows will
use s(xmax) to update its own s(T).
Dead: a new safe point has arrived since the version was deleted,
and nobody can see the version any more. The system can safely
forget c(xmax) and s(xmax).
'''
# we need to track all in-flight readers and the most
# recently-committed reader of each unclobbered version. These are
# the equivalent to P&G's SIREAD locks. Our version is less
# memory-intensive, though, because each version retains at most
# one committed lock---that of the most recent reader---and even
# that one lock can be discarded at the next safe point.
v_readers,v_rstamps = {},{}
def get_rstamp(rid, dep):
return v_rstamps.get((rid,dep), 0)
def get_version_reader(rid, dep):
reader = v_readers.get((rid,dep), 0)
if not version_reader:
return reader,None
rc = tracker.get_safe_point()
if tracker.is_known(reader):
commit = tracker.is_committed(reader)
if commit:
# read-only optimization: use begin time of committed
# read-only reader
commit = safe_snapper.get_snap(reader)
elif commit is False:
reader = 0 # aborted!
rc = commit
return reader,rc
def tx_read(pid, rid, for_update=False):
Q = q and pid not in tid_watch and rid not in rid_watch
'''Postgresql does not track records as sequences of
versions. Instead, each record is more or less independent and
contains two transaction ids (xids): xmin and xmax, which are,
respectively, the transaction that created and deleted the
version in question.
We achieve the same effect here (since our simulated
microbenchmark does not have inserts or deletes) by treating
the xmin of the clobbering version as xmax.
'''
y,xprev,xmax = None,None,None
pt,st = tx_pstamps[pid],tx_sstamps.get(pid, None)
def read_filter(it):
# take the newest version that precedes our snapshot.
nonlocal xmax
snap = safe_snapper.get_snap(pid) or tracker.get_begin(pid)
for x,xmin,_ in it:
#e('x=%d xmin=%d', x, xmin)
if not tracker.is_known(xmin):
cxmin = tracker.get_safe_point()
else:
cxmin = tracker.is_committed(xmin)
assert cxmin
r2 = not safe_snapper.get_snap(pid) and si_relax_reads == 2
r1 = not safe_snapper.get_snap(pid) and si_relax_reads == 1
#e('r1=%d r2=%d pid=%d cxmin=%d snap=%d', r1, r2, pid, cxmin, snap)
#if (r2 and pid == xmin) or cxmin < snap:
# Return the latest committed for RC (r2=true)
if (r2) or cxmin < snap:
if r1 and pid != xmin and xmax:
# check for and avoid fatal anti-dep
assert pid != xmin
assert tracker.is_known(xmax)
sxmax = tx_sstamps.get(xmax, None)
if sxmax and not (pt < sxmax):
xmax=xprev
return y
if st and not (cxmin < st):
xmax=xprev
return y
return x
# only committed clobbers are visible to us here
y,xprev,xmax = x,xmax,xmin
errlog('Oops! T=%d finds no suitable version of R=%d', pid, rid)
errlog('\t(last version was %d, visible since %s, T started at %d)',
xmin, cxmin, snap)
assert not 'reachable'
# do not rely on version commit stamps
xmin,_ = yield from tracker.on_access(pid, rid, read_filter)
Q or e('read: pid=%s rid=%s', pid, rid)
# deal with inbound dependency edge
if not safe_snapper.get_snap(pid):
if not tracker.is_known(xmin):
cxmin = tracker.get_safe_point()
else:
cxmin = tracker.is_committed(xmin)
assert cxmin
Q or e('\tInbound dependency: T=%d p(T)=%d X=%d c(X)=%s',
pid, pt, xmin, cxmin)
if cxmin > pt:
pt = tx_pstamps[pid] = cxmin
if st and not (pt < st):
raise EagerFailReadDep
# check for a committed outbound anti-dependency edge. The
# clobber (if any) cannot have yet been finalized, else we
# would have used that version instead, so we can grab s(T) of
# the overwriter (which must exist)
if safe_snapper.get_snap(pid):
pass
elif xmax:
# reading under a committed overwrite
'''WARNING:
Because pgsql does not track records as sequences of
versions, and because it allows deletions, it could
encounter a finalized xmax. In that case the version is
simply not visible, and so we cannot acquire a r:w
conflict through it.
In our case, though, we should have seen the newer version
and it would be an error not to have chosen it.
'''
assert tracker.is_known(xmax)
sxmax = tx_sstamps[xmax]
assert sxmax
if not safe_snapper.get_snap(pid):
Q or e('\tOutbound anti-dependency: T=%d s(T)=%s X=%d s(X)=%s',
pid, st, xmax, sxmax)
if not st or sxmax < st:
st = tx_sstamps[pid] = sxmax
if not (pt < st):
raise EagerFailReadAntiDep
else:
Q or e('\tNo committed overwrite found -> sxmax=None')
if xmin != pid:
# reading somebody else's write, clobber yet to come
x = tx_reads[pid].setdefault(rid, xmin)
if x != xmin:
assert x != pid
# oops, already read a different version, and
# neither was created by this transaction
raise ReadSkew
# have to worry about a future overwrite
rstamp = get_rstamp(rid,xmin)
reader,rc = get_version_reader(rid,xmin)
if rc is None:
# have to revisit during pre-commit
Q or e('\tT=%d adds version to read footprint', pid)
if xmin != pid:
x = tx_reads[pid].setdefault(rid, xmin)
if x != xmin:
assert x != pid
# oops, already read a different version, and
# neither was created by this transaction
raise ReadSkew
else:
# update rstamp as necessary and claim the slot
if rstamp < rc:
Q or e('\tUpdate rstamp r(V)=%d c(reader)=%d', rstamp, rc)
rstamp = v_rstamps[rid,xmin] = rc
Q or e('\tT=%d marks version', pid)
reader = v_readers[rid,xmin] = pid
yield from sys_busy(random.randint(ONE_TICK, 2*ONE_TICK),
color='green', title='%s=db[%s]' % (xmin, rid))
return xmin
def tx_write(pid, rid):
assert not safe_snapper.get_snap(pid)
Q = q and pid not in tid_watch and rid not in rid_watch
Q or e('write: pid=%s rid=%d', pid, rid)
# do the write (don't depend on the version stamp)
dep,_ = yield from tracker.on_access(pid, rid, False)
# have I written this before?
if dep == pid:
q or e('\talready wrote to this record')
return
# Did we read a different version than this? We can't know for
# sure, because we don't actually track the full read set, but
# we can at least check the reads we did record.
rdep = tx_reads.get(pid, {}).get(rid, dep)
if rdep != dep:
raise ReadSkew
# can I actually see the version I'm overwriting?
if not tracker.is_known(dep):
cdep = tracker.get_safe_point()
else:
cdep = tracker.get_end(dep)
snap = safe_snapper.get_snap(pid) or tracker.get_begin(pid)
if not si_relax_writes and cdep > snap:
raise WriteSkew
# record the write
tx_writes[pid].add((rid,dep))
# deal with inbound dependency edge
pt,st = tx_pstamps[pid],tx_sstamps.get(pid,None)
Q or e('\tInbound dependency: T=%d p(T)=%d X=%d c(X)=%d',
pid, pt, dep, cdep)
if pt < cdep:
pt = tx_pstamps[pid] = cdep
if st and not (pt < st):
raise EagerFailWriteDep
# check for inbound anti-dependency edges
rstamp = get_rstamp(rid,dep)
reader,rc = get_version_reader(rid,dep)
if reader == pid:
Q or e('\tI previously read this version!')
reader = v_readers[rid,dep] = 0
elif rc and rstamp < rc:
Q or e('\tT=%d found committed read mark: r(V)=%s X=%d c(X)=%s',
pid, rstamp, reader, rc)
rstamp = v_rstamps[rid,dep] = rc
Q or e('\tDeal with committed inbound anti-deps: T=%d p(T)=%d r(V)=%d',
pid, pt, rstamp)
if pt < rstamp:
pt = tx_pstamps[pid] = rstamp
if st and not (pt < st):
q or e('EagerFailWrite pid=%d', pid)
raise EagerFailWrite
last_snap = safe_snapper.get_last_snap()
if pt < last_snap:
# don't update pt, it diffuses blame
#pt = tx_pstamps[pid] = last_snap
if st and not (last_snap < st):
raise SafeSnapKill
# no outbound edges possible
yield from sys_busy(random.randint(ONE_TICK, 2*ONE_TICK),
color='blue', title='%s=db[%s]' % (dep, rid))
tid_watched = set()
def tx_create(pid, is_readonly=False):
if tid_watch and len(tid_watch) == len(tid_watched):
hanging = set()
for x in tid_watched:
if tracker.is_known(x):
hanging.add(x)
if hanging:
errlog('Unfinalized pids: %s', ','.join(map(str, hanging)))
yield from sys_exit(0)
#t.begin = yield from sys_now()
tracker.on_begin(pid)
if is_readonly:
yield from safe_snapper.request_safe_snap(pid)
def tx_commit(pid):
Q = q and pid not in tid_watch
yield from sys_busy(random.randint(ONE_TICK, 2*ONE_TICK), color='yellow')
Q or e('Commit %s', pid)
if pid in tid_watch:
tid_watched.add(pid)
# get a serialization point, not reused even if we abort
end = tracker.on_precommit(pid)
pt,st = tx_pstamps[pid],tx_sstamps.setdefault(pid, end)
if st == end:
Q or e('\tT=%d initialized s(T)=%s at pre-commit', pid, end)
else:
Q or e('\tT=%d has s(T)=%s at start of pre-commit', pid, end)
# /// BEGIN CRITICAL SECTION ///
# read-only optimization: I have to update rstamps of all
# remembered reads, but if I finished read-only then I can use
# my begin timestamp instead of my end timestamp, thus
# reducing the probability that I violate some overwriter's
# exclusion window. In case of forgotten reads, they will get
# the right value by calling get_version_reader().
ct = safe_snapper.get_snap(pid)
if read_only_opt and not ct and pid not in tx_writes:
ct = tracker.get_begin(pid)
else:
ct = end
# check for reads that commit under an overwrite
safe_point = tracker.get_safe_point()
for rid,dep in tx_reads.get(pid, {}).items():
#print (rid)
rstamp = get_rstamp(rid,dep)
Q or e('\tT=%d verifying read of R%d/T%d', pid, rid, dep)
# skip versions we overwrote
if (rid,dep) in tx_writes.get(pid, ()):
continue
# Check for overwrites. If no clobber has shown up yet,
# pretend this transaction is the overwriter (guaranteed
# to be reported as in-flight) to force an rstamp update.
x,_ = tracker.get_overwriter(rid, dep)
if not x:
# no committed overwrite, have to update rstamp
Q or e('\tUpdate r(V): T=%d c(T)=%d r(V)=%d',
pid, ct, rstamp)
if rstamp < ct:
rstamp = v_rstamps[rid,dep] = ct
else:
# overwrite committed after we read, so update s(T)
if not tracker.is_known(x):
sx = -1
else:
sx = tx_sstamps.get(x, None)
Q or e('\tCommitted clobber: T=%d s(T)=%d X=%d s(X)=%s',
pid, st, x, sx)
if sx and st > sx:
st = tx_sstamps[pid] = sx
for rid,dep in tx_writes.get(pid,()):
Q or e('\tT=%d verifying overwrite of R%d/T%d', pid, rid, dep)
rstamp = get_rstamp(rid,dep)
reader,rc = get_version_reader(rid,dep)
if rc:
Q or e('\tActive reader committed: V=%d, X=%d c(X)=%d r(V)=%d',
rid, reader, rc, rstamp)
if rstamp < rc:
rstamp = v_rstamps[rid,dep] = rc
Q or e('\tAccount for committed readers: T=%d p(T)=%d r(V)=%d',
pid, pt, rstamp)
if pt < rstamp:
pt = tx_pstamps[pid] = rstamp
if not (pt < st):
if pt == st:
Q or e('\tPrecommitBackedgeFail pid=%d', pid)
raise PrecommitBackedgeFail
Q or e('\tPrecommitFail pid=%d', pid)
raise PrecommitFail
last_snap = safe_snapper.get_last_snap()
if pid in tx_writes and pt < last_snap:
Q or e('\tT=%d follows safe snapshot %d', pid, last_snap)
pt = tx_pstamps[pid] = last_snap
if not (pt < st):
Q or e('\tPrecommitFail pid=%d', pid)
raise SafeSnapKill
for rid,dep in tx_writes.get(pid,()):
Q or e('\tT=%d verifying overwrite of R%d/T%d', pid, rid, dep)
reader,rc = get_version_reader(rid,dep)
if not rc:
sr = tx_sstamps.get(reader,None)
Q or e('\tUpdate s(X) for in-flight reader: T=%d s(T)=%s X=%d s(X)=%s',
pid, st, reader, sr)
if not sr or sr > st:
sr = tx_sstamps[reader] = st
Q or e('\tT=%d at commit: p(T)=%d s(T)=%d', pid, pt, st)
# /// END CRITICAL SECTION ///
tx_reads.pop(pid,None)
tx_pstamps.pop(pid,None)
yield from tracker.on_finish(pid, True, finish_callback)
yield from sys_sleep(random.randint(5*ONE_TICK, 10*ONE_TICK))
yield from sys_busy(random.randint(ONE_TICK, 2*ONE_TICK), color='orange')
def tx_abort(pid):
Q = q and pid not in tid_watch
if pid in tid_watch:
tid_watched.add(pid)
tx_reads.pop(pid,None)
tx_pstamps.pop(pid,None)
Q or e('Abort %d', pid)
yield from sys_busy(random.randint(ONE_TICK, 2*ONE_TICK), color='red')
yield from tracker.on_finish(pid, False, finish_callback)
def finish_callback(pid):
Q = q and pid not in tid_watch
Q or e('Forget pid=%d', pid)
tx_sstamps.pop(pid,None)
tx_writes.pop(pid,None)
safe_snapper.forget(pid)
def fini():
print_general_stats(stats)
print_failure_causes(stats)
return NamedTuple(db_size=db_size, tx_begin=tx_create,
tx_read=tx_read, tx_write=tx_write,
tx_commit=tx_commit, tx_abort=tx_abort,
fini=fini,
begin_tracking=tracker.begin_tracking,
end_tracking=tracker.end_tracking)
def test_ssi3_db():
R,U,X = 1,2,3
def test_fini(db):
done = False
def callback():
nonlocal done
done = True
def nop():
pid = yield from sys_getpid()
yield from db.tx_begin(pid)
yield from db.tx_abort(pid)
yield from sys_sleep(1000*ONE_TICK)
db.end_tracking(callback)
yield from sys_spawn(nop())
yield from sys_sleep(1000*ONE_TICK)
yield from sys_spawn(nop())
yield from sys_sleep(10000*ONE_TICK)
assert done
db.fini()
yield from sys_exit()
def access(db, pid, rid, mode, delay):
yield from sys_sleep(delay*ONE_TICK)
yield from db.tx_write(pid, rid) if mode == X else db.tx_read(pid, rid, mode == U)
def commit(db, pid, delay):
if not isinstance(delay, int):
errlog('bad delay: %s', delay)
yield from sys_sleep(delay*ONE_TICK)
yield from db.tx_commit(pid)
def tx_one(db, rid, mode, delay1, delay2):
def thunk():
pid = yield from sys_getpid()
try:
yield from db.tx_begin(pid)
yield from access(db, pid, rid, mode, delay1)
yield from commit(db, pid, delay2)
except AbortTransaction:
yield from db.tx_abort(pid)
return (yield from sys_spawn(thunk()))
def tx_two(db, rid1, mode1, rid2=None, mode2=None, delay1=0, delay2=0, delay3=0):
def thunk():
pid = yield from sys_getpid()
try:
yield from db.tx_begin(pid)
yield from access(db, pid, rid1, mode1, delay1)
yield from access(db, pid, rid2 or rid1, mode2 or mode1, delay2)
yield from commit(db, pid, delay3)
except AbortTransaction:
yield from db.tx_abort(pid)
return (yield from sys_spawn(thunk()))
def tx_n(db, commit_delay, *args):
# accept (rid,mode,delay) triples
def thunk():
pid = yield from sys_getpid()
try:
yield from db.tx_begin(pid)
for rid,mode,delay in args:
yield from access(db, pid, rid, mode, delay)
yield from commit(db, pid, commit_delay)
except AbortTransaction:
yield from db.tx_abort(pid)
return (yield from sys_spawn(thunk()))
def test1(db):
'''reads coexist peacefully'''
# R-lock at t=0, S-locks at t=2 and t=2
yield from tx_one(db, 1, R, 0, 10)
yield from tx_one(db, 1, R, 1, 8)
yield from tx_one(db, 1, R, 2, 6)
yield from test_fini(db)
def test2(db):
'''incompatible requests block until the holder leaves'''
yield from tx_one(db, 1, R, 0, 10)
yield from tx_one(db, 1, X, 2, 0)
yield from test_fini(db)
def test3(db):
'''reads can coexist with one upgrade lock, but a second upgrader
blocks until the first leaves. Also make sure that the first
upgrader can still upgrade when the lock mode is W.
'''
yield from tx_one(db, 1, R, 0, 10)
yield from tx_two(db, rid1=1, mode1=U, mode2=X, delay1=2, delay2=8)
yield from tx_two(db, rid1=1, mode1=U, mode2=X, delay1=3, delay2=6)
yield from test_fini(db)
def test4(db):
'''new readers can coexist with an upgrade lock but not with an
in-progress upgrade.
'''
yield from tx_one(db, 1, R, 0, 4)
yield from tx_two(db, rid1=1, mode1=U, mode2=X, delay1=1, delay2=1)
yield from tx_one(db, 1, R, 2, 4)
yield from tx_one(db, 1, R, 4, 0)
yield from test_fini(db)
def test5(db):
'''reader cannot upgrade if an upgrade lock has been granted'''
yield from tx_two(db, rid1=1, mode1=R, mode2=X, delay2=5, delay3=1)
yield from tx_one(db, 1, U, 1, 10)
yield from test_fini(db)
def test6(db):
'''simple two-party deadlock detected'''
yield from tx_two(db, rid1=1, mode1=R, rid2=2, mode2=X, delay2=5)
yield from tx_two(db, rid1=2, mode1=R, rid2=1, mode2=X, delay1=1)
yield from test_fini(db)
def test7(db):
'''three-party deadlock detected'''
yield from tx_two(db, rid1=1, mode1=R, rid2=2, mode2=X, delay2=5)
yield from tx_two(db, rid1=2, mode1=R, rid2=3, mode2=X, delay1=1)
yield from tx_two(db, rid1=3, mode1=R, rid2=1, mode2=X, delay1=2)
yield from test_fini(db)
def test8(db):
'''SI read-only anomaly detected:
T1 T2 T3
read A
read B
write B' (T2)
D-commit (T2)
read A
read B' (T3)
read C
D-commit (T2 T3)
write C' (T1)
!! abort !!
I-commit
I-commit
'''
yield from tx_n(db, 0, (0, R, 200), (1, R, 0), (2, R, 0))
yield from tx_n(db, 0, (0, R, 0), (1, R, 0), (2, X, 300))
yield from tx_n(db, 0, (1, X, 100))
yield from test_fini(db)
A,B,C,D = 0,1,2,3
def test9(db):
'''Complex scenario that SI and 2PL both reject:
T1 T2 T3 T4 T5
read A
write A' (T1)
D-commit (T1)
read B
write B' (T3)
write C'
D-commit (T3)
write C'' (T2)
D-commit (T2, T3)
read C'' (T5)
read C (**)
write D'
DI-commit
I-commit
I-commit
write D''
DI-commit
I-commit
'''
yield from tx_n(db, 0, (A, R, 0), (C, R, 500), (D, X, 300))
yield from tx_n(db, 0, (B, X, 300), (C, X, 0))
yield from tx_n(db, 0, (B, R, 200), (C, R, 400), (D, X, 0))
yield from tx_n(db, 0, (A, X, 100))
yield from tx_n(db, 0, (C, X, 400))
yield from test_fini(db)
def test10(db):
'''Schedule found during a measurement run. The Dangerous Structure
has not yet formed when T2 commits first; when
T3 commits second it sees T1 rw(A) T2 rw(B) T3:
T1 T2 T3
read B
write B'
write A'
read A
D-commit
write C'
D-commit
read C'
D-commit
'''
yield from tx_n(db, 100, (A, R, 0), (C, X, 200))
yield from tx_n(db, 0, (B, X, 100), (A, X, 400))
yield from tx_n(db, 0, (C, R, 300), (B, R, 300))
yield from test_fini(db)
def test11(db):
'''Schedule found during a measurement run. The Dangerous Structure
has not yet formed when T2 commits first; when
T3 commits second it sees T1 rw(A) T2 rw(B) T3:
T1 T2 T3
read B
write A'
write C'
read A
commit
write B'
commit
write C''
commit
'''
yield from tx_n(db, 0, (A, R, 200), (C, X, 300))
yield from tx_n(db, 300, (B, R, 0), (A, X, 0))
yield from tx_n(db, 0, (C, X, 200), (B, X, 300))
yield from test_fini(db)
# Reproduces the scenario in Figure 4d of the paper
def testrc(db):
yield from tx_n(db, 0, (1, R, 100), (0, X, 400))
yield from tx_n(db, 0, (1, X, 200))
yield from tx_n(db, 0, (0, R, 0), (1, R, 300), (2, X, 500))
yield from test_fini(db)
def test12(db):
'''Schedule found during a measurement run. A dependency cycle forms
without manifesting any Dangerous Structures:
T1 T2 T3 T4
read A
write A'
write B'
read C
D-commit
write C'
write D'
D-commit
read D'
read B'
D-commit
D-commit
Adversarial commit ordering can make it even nastier. Here, T1
is arguably the problem transaction, since its read of D' is
impossible under SI, but it commits before T3 even reads
C'. That means it's not enough for T1 to detect that it has
gone awry; T3 must know to abort itself even though it has
done nothing "wrong."
T1 T2 T3 T4
read A
write A'
write C'
D-commit
read B
write B'
write D'
D-commit
read D'
D-commit
read C'
D-commit
Note that there's still a shadow of the Dangerous Structure
here. Every cycle must include (at least) two RW dependencies:
one leaks information out of an uncommitted transaction and
the other circumvents the commit-time dependency tracking that
normally prevents the leakage from causing isolation
failures. The RW deps just don't have to be adjacent any more
after we give up SI.
Thought: perhaps we can check D-commit times: it's a bad sign
if I read a version that was clobbered before 1+ of the
transactions in my dependency set. In the above example, T3
depends on {T1 T2}, and gives a RW dep to T4; T4 D-committed
before T1, and so T1 could potentially be poisoned by
T4. There could be a lot of false positives, though: if T4 did
not write D' at all (writing only B', for example), then there
is no cycle. This looks suspiciously similar to wound-wait,
which has a painfully high false positive rate.
'''
yield from tx_n(db, 300, (A, X, 0), (B, X, 0))
yield from tx_n(db, 200, (C, R, 100), (B, R, 600))
yield from tx_n(db, 0, (C, X, 500), (D, X, 0))
yield from tx_n(db, 200, (A, R, 0), (D, R, 600))
yield from test_fini(db)
def test13(db):
'''One of the simplest possible serialization anomalies
T1 T2
read A
write A'
write B'
D-commit
read B
D-commit
'''
yield from tx_n(db, 0, (A, R, 0), (B, R, 300))
yield from tx_n(db, 0, (A, X, 100), (B, X, 0))
yield from test_fini(db)
def test14(db):
'''Scenario suggested by Alan Fekete:
T1 T2 T3
Write A'
Read B
Write B'
Write C'
Commit
Read A'
Commit
Write C''
Commit
T2 should abort.
'''
yield from tx_n(db, 0, (C, X, 400), (A, R, 200))
yield from tx_n(db, 0, (B, R, 100), (C, X, 600))
yield from tx_n(db, 200, (A, X, 0), (B, X, 200))
yield from test_fini(db)
def test15(db):
'''This scenario found during simulation run:
T1 T2 T3 T4
write A'
write B'
read B
D-commit
read C
write C'
D-commit
read D
write D'
read A'
D-commit
D-commit
Produces the cycle T1 -- T4 rw T3 rw T2 rw T1
'''
yield from tx_n(db, 200, (A, X, 0), (B, X, 0))
yield from tx_n(db, 0, (B, R, 200), (C, X, 300))
yield from tx_n(db, 200, (C, R, 400), (D, X, 400))
yield from tx_n(db, 200, (D, R, 700), (A, R, 200))
yield from test_fini(db)
def test16(db):
'''Scenario found during a simulation run. Requires Test #2 to avoid
wedging the DB, even under SI. T4 should abort.
T1 T2 T3 T4
read C
write D'
read A
write C'
D-commit
write A'
write B
D-commit
read B'
read D
D-commit
D-commit
T1 rw T2 -- T4 rw T3 rw T1
'''
yield from tx_n(db, 0, (A, R, 100), (C, X, 0))
yield from tx_n(db, 0, (A, X, 400), (B, X, 0))
yield from tx_n(db, 600, (C, R, 0), (D, X, 0))
yield from tx_n(db, 200, (B, R, 600), (D, R, 0))
yield from test_fini(db)
failures = 0
for test in (test1,test2,test3,test4,test5,test6,test7,test8,test9,
test10,test11,test12,test13,test14,test15,test16):
#for test in (testrc):
errlog('\n%s\nRunning test:\n\n\t%s\n', '='*72, test.__doc__)
stats = NamedTuple()
tracker = dependency_tracker(stats)
db = make_db(stats=stats, tracker=tracker,db_size=10, ww_blocks=True, verbose=True)
db.begin_tracking()
try:
simulator(test(db), log=log_svg_task())
except:
traceback.print_exc()
failures +=1
if failures:
exit(1)
if __name__ == '__main__':
try:
seed = sys.argv[1]
except:
seed = make_seed()
errlog('Using RNG seed: %s', seed)
random.seed(seed)
test_ssi3_db()
#simulator(make_benchmark(make_db=make_si_db, nclients=50, max_inflight=100, db_size=1000, duration=10000), log=log_svg_task())