Oz, Declarative Concurrency, and Active Objects VRH 4, 7.8 - PowerPoint PPT Presentation

About This Presentation
Title:

Oz, Declarative Concurrency, and Active Objects VRH 4, 7.8

Description:

Imagine that these threads really execute in parallel, each has its own ... Threads suspend on data unavailability in dataflow variables ... – PowerPoint PPT presentation

Number of Views:87
Avg rating:3.0/5.0
Slides: 83
Provided by: seifh
Learn more at: http://www.cs.rpi.edu
Category:

less

Transcript and Presenter's Notes

Title: Oz, Declarative Concurrency, and Active Objects VRH 4, 7.8


1
Oz, Declarative Concurrency, and Active Objects
(VRH 4, 7.8)
  • Carlos Varela
  • RPI
  • Adapted with permission from
  • Seif Haridi
  • KTH
  • Peter Van Roy
  • UCL

2
Overview of concurrent programming
  • Four basic approaches to programming are
  • Sequential programming (no concurrency)
  • Declarative concurrency (streams in a functional
    language, Oz)
  • Message passing with active objects (Erlang,
    SALSA)
  • Atomic actions on shared state (Java)
  • The atomic action approach is the most difficult,
    yet it is the one you will probably be most
    exposed to!
  • But, if you have the choice, which approach to
    use?
  • Use the simplest approach that does the job
    sequential if that is ok, else declarative
    concurrency if there is no observable
    nondeterminism, else message passing if you can
    get away with it.

3
Concurrency
  • Some programs are best written as a set of
    activities that run independently (concurrent
    programs)
  • Concurrency is essential for interaction with the
    external environment
  • Examples includes GUI (Graphical User
    Interfaces), operating systems, web services
  • Also programs that are written independently but
    interact only when needed (client-server,
    peer-to-peer applications)
  • First, we will cover declarative concurrency,
    programs with no observable nondeterminism, the
    result is a function
  • Independent procedures that execute at their own
    pace and may communicate through shared dataflow
    variables
  • Then, we will cover message passing, programs
    consisting of components with encapsulated state
    communicating asynchronously

4
Overview of declarative concurrency
  • Programming with threads
  • The model is augmented with threads
  • Programming techniques stream communication,
    order-determining concurrency, concurrent
    composition
  • Lazy execution
  • demand-driven computations, lazy streams
  • Soft real-time programming

5
The sequential model
Statements are executed sequentially from a
single semantic stack
Semantic Stack
w a z person(age y) x y 42 u
Single-assignment store
6
The concurrent model
Semantic Stack 1
Semantic Stack N
Multiple semantic stacks (threads)
w a z person(age y) x y 42 u
Single-assignment store
7
Concurrent declarative model
The following defines the syntax of a statement,
?s? denotes a statement
?s? skip
empty statement ?x? ?y?

variable-variable binding
?x?
?v? variable-value binding
?s1?
?s2? sequential composition local ?x?
in ?s1? end declaration proc ?x? ?y1?
?yn? ?s1? end procedure introduction if
?x? then ?s1? else ?s2? end conditional
?x? ?y1? ?yn? procedure
application case ?x? of ?pattern? then ?s1?
else ?s2? end pattern matching thread ?s1?
end thread creation
8
The concurrent model
ST thread ?s1? end,E
Top of Stack, Thread i
Single-assignment store
9
The concurrent model
ST
Top of Stack, Thread i
?s1?,E
Single-assignment store
10
Basic concepts
  • The model allows multiple statements to execute
    at the same time.
  • Imagine that these threads really execute in
    parallel, each has its own processor, but share
    the same memory
  • Reading and writing different variables can be
    done simultaneously by different threads, as well
    as reading the same variable
  • Writing the same variable is done sequentially
  • The above view is in fact equivalent to an
    interleaving execution a totally ordered
    sequence of computation steps, where threads take
    turn doing one or more steps in sequence

11
Causal order
  • In a sequential program all execution states are
    totally ordered
  • In a concurrent program all execution states of a
    given thread are totally ordered
  • The execution state of the concurrent program as
    a whole is partially ordered

12
Total order
  • In a sequential program all execution states are
    totally ordered

sequential execution
computation step
13
Causal order in the declarative model
  • In a concurrent program all execution states of a
    given thread are totally ordered
  • The execution state of the concurrent program is
    partially ordered

thread T3
thread T2
fork a thread
thread T1
computation step
14
Causal order in the declarative model
synchonize on a dataflow variable
bind a dataflow variable
thread T3
x
thread T2
fork a thread
y
thread T1
computation step
15
Nondeterminism
  • An execution is nondeterministic if there is a
    computation step in which there is a choice what
    to do next
  • Nondeterminism appears naturally when there is
    concurrent access to shared state

16
Example of nondeterminism
Thread 1
Thread 2
store
x
y 5
x 1
x 3
time
time
The thread that binds x first will continue, the
other thread will raise an exception
17
Nondeterminism
  • An execution is nondeterministic if there is a
    computation step in which there is a choice what
    to do next
  • Nondeterminism appears naturally when there is
    concurrent access to shared state
  • In the concurrent declarative model when there is
    only one binder for each dataflow variable, the
    nondeterminism is not observable on the store
    (i.e. the store develops to the same final
    results)
  • This means for correctness we can ignore the
    concurrency

18
Scheduling
  • The choice of which thread to execute next and
    for how long is done by a part of the system
    called the scheduler
  • A thread is runnable if its next statement to
    execute is not blocked on a dataflow variable,
    otherwise the thread is suspended
  • A scheduler is fair if it does not starve a
    runnable thread, i.e. all runnable threads
    eventually execute
  • Fair scheduling makes it easy to reason about
    programs and program composition
  • Otherwise some correct program (in isolation) may
    never get processing time when composed with
    other programs

19
The semantics
  • In the sequential model we had
  • (ST , ? )
  • ST is a stack of semantic statements
  • ? is the single assignment store
  • In the concurrent model we have
  • (MST , ? )
  • MST is a (multi)set of stacks of semantic
    statements
  • ? is the single assignment store

20
The initial execution state
statement
  • ( (?s?,?) , ?)

stack
store
multiset
21
Execution (the scheduler)
  • At each step, one runnable semantic stack is
    selected from MST (the multiset of stacks), call
    it ST, s.t. MST ST ? MST
  • Assume the current store is ?, one computation
    step is done that transforms ST to ST and ? to
    ?
  • The total computation state is transformed from
    (MST, ?) to (ST ? MST, ?)
  • Which stack is selected, and how many steps are
    taken is the task of the scheduler, a good
    scheduler should be fair, i.e., each runnable
    thread will eventually be selected
  • The computation stops when there are no runnable
    stacks

22
Example of runnable threads
  • proc Loop P N
  • if N gt 0 then
  • P Loop P N-1
  • else skip end
  • end
  • thread Loop proc Show 1 end
    1000
  • end
  • thread Loop
  • proc Show 2 end
  • 1000
  • end
  • This program will interleave the execution of two
    threads, one printing 1, and the other printing 2
  • We assume a fair scheduler

23
Dataflow computation
  • Threads suspend on data unavailability in
    dataflow variables
  • The Delay X primitive makes the thread suspends
    for X milliseconds, after that, the thread is
    runnable

declare X Browse X local Y in thread Delay
1000 Y 1010 end X Y 100100 end
24
Illustrating Dataflow computation
  • Enter incrementally the values of X0 to X3
  • When X0 is bound the thread will compute Y0X01,
    and will suspend again until X1 is bound

declare X0 X1 X2 X3 Browse X0 X1 X2
X3 thread Y0 Y1 Y2 Y3 in Browse Y0 Y1
Y2 Y3 Y0 X0 1 Y1 X1 Y0 Y2 X2
Y1 Y3 X3 Y2 Browse completed end
25
Concurrent Map
  • fun Map Xs F
  • case Xs
  • of nil then nil
  • XXr then thread F X endMap Xr F
  • end
  • end
  • This will fork a thread for each individual
    element in the input list
  • Each thread will run only if both the element X
    and the procedure F is known

26
Concurrent Map Function
  • fun Map Xs F case Xs of nil then nil
    XXr then thread F X end Map Xr F end
  • end
  • What this looks like in the kernel language
  • proc Map Xs F Rs case Xs of nil then Rs
    nil XXr then R Rr in Rs RRr
    thread R F X end Map Xr F Rr end
  • end

27
How does it work?
  • If we enter the following statementsdeclare F X
    Y ZBrowse thread Map X F end
  • A thread executing Map is created.
  • It will suspend immediately in the case-statement
    because X is unbound.
  • If we thereafter enter the following
    statementsX 12Yfun F X XX end
  • The main thread will traverse the list creating
    two threads for the first two arguments of the
    list,

28
How does it work?
  • The main thread will traverse the list creating
    two threads for the first two arguments of the
    list
  • thread F 1 end, and thread F 2 end, Y
    3ZZ nil
  • will complete the computation of the main thread
    and the newly created thread thread F 3 end,
    resulting in the final list 1 4 9.

29
Cheap concurrency and dataflow
  • Declarative programs can be easily made
    concurrent
  • Just use the thread statement where concurrency
    is needed
  • fun Fib X
  • if Xlt2 then 1
  • else
  • thread Fib X-1 end Fib X-2
  • end
  • end

30
Understanding why
  • fun Fib X
  • if Xlt2 then 1
  • else F1 F2 in
  • F1 thread Fib X-1 end F2 Fib
    X-2
  • F1 F2end
  • end

Dataflow dependency
31
Execution of Fib 6
F2
Fork a thread
F1
F3
F2
F4
Synchronize on result
F2
F5
F1
F3
F2
Running thread
F1
F3
F6
F4
F2
32
Fib
33
Streams
  • A stream is a sequence of messages
  • A stream is First-In First-Out (FIFO) channel
  • The producer augments the stream with new
    messages, and the consumer reads the messages,
    one by one.

x5 x4 x3 x2 x1
producer
consumer
34
Stream Communication I
  • The data-flow property of Oz easily enables
    writing threads that communicate through streams
    in a producer-consumer pattern.
  • A stream is a list that is created incrementally
    by one thread (the producer) and subsequently
    consumed by one or more threads (the consumers).
  • The consumers consume the same elements of the
    stream.

35
Stream Communication II
  • Producer, that produces incrementally the
    elements
  • Transducer(s), that transforms the elements of
    the stream
  • Consumer, that accumulates the results

thread 1
thread 2
thread 3
thread N
producer
transducer
transducer
consumer
36
Program patterns
  • The producer, transducers, and the consumer can,
    in general, be described by certain program
    patterns
  • We show the various patterns

37
Producer
  • fun Producer State
  • if More State then
  • X Produce State in
  • X Producer Transform State
  • else nil end
  • end
  • The definition of More, Produce, and Transform is
    problem dependent
  • State could be multiple arguments
  • The above definition is not a complete program!

38
Example Producer
  • fun Generate N Limit
  • if NltLimit then
  • N Generate N1 Limit
  • else nil end
  • end
  • The State is the two arguments N and Limit
  • The predicate More is the condition NltLimit
  • The Transform function (N,Limit) ? (N1,Limit)

fun Producer State if More State then
X Produce State in X Producer
Transform State else nil end end
39
Consumer Pattern
  • fun Consumer State InStream
  • case InStream
  • of nil then Final State
  • X RestInStream then
  • NextState Consume X State in
  • Consumer NextState RestInStream
  • end
  • end
  • Final and Consume are problem dependent

The consumer suspends until InStream is either a
cons or a nil
40
Example Consumer
fun Consumer State InStream case InStream
of nil then Final State X RestInStream
then NextState Consume X State in
Consumer NextState RestInStream end end
  • fun Sum A Xs
  • case Xs
  • of XXr then Sum AX Xr
  • nil then A
  • end
  • end
  • The State is A
  • Final is just the identity function on State
  • Consume takes X and State ? X State

41
Transducer Pattern 1
  • fun Transducer State Instream
  • case InStream
  • of nil then nil
  • X RestInStream then
  • NextStateTX Transform X State
  • TX Transducer NextState RestInStream
  • end
  • end
  • A transducer keeps its state in State, receives
    messages on InStream and sends messages on
    OutStream

42
Transducer Pattern 2
  • fun Transducer State Instream
  • case InStream
  • of nil then nil
  • X RestInStream then if Test XState
    then
  • NextStateTX Transform X State
  • TX Consumer NextState
    RestInStreamelse Consumer NextState
    RestInStream end
  • end
  • end
  • A transducer keeps its state in State, receives
    messages on InStream and sends messages on
    OutStream

43
Example Transducer
IsOdd
6 5 4 3 2 1
5 3 1
Generate
Filter
Filter is a transducer that takes an Instream and
incremently produces an Outstream that
satisfies the predicate F
  • fun Filter Xs F
  • case Xs
  • of nil then nil
  • XXr then
  • if F X then XFilter Xr F
  • else Filter Xr F end
  • end
  • end

local Xs Ys in thread Xs Generate 1 100
end thread Ys Filter Xs IsOdd end
thread Browse Ys end end
44
Larger ExampleThe sieve of Eratosthenes
  • Produces prime numbers
  • It takes a stream 2...N, peals off 2 from the
    rest of the stream
  • Delivers the rest to the next sieve

Sieve
X
Xs
XZs
Filter
Sieve
Zs
Xr
Ys
45
Sieve
  • fun Sieve Xs
  • case Xs
  • of nil then nil
  • XXr then Ys in
  • thread Ys Filter Xr fun Y Y mod X \
    0 end end
  • X Sieve Ys
  • end
  • end
  • The program forks a filter thread on each sieve
    call

46
Example Call
  • local Xs Ys in
  • thread Xs Generate 2 100000 end
  • thread Ys Sieve Xs end
  • thread for Y in Ys do Show Y end end
  • end

7 11 ...
Filter 3
Sieve
Filter 5
Filter 2
47
Limitation of eager stream processing
  • The producer might be much faster than the
    consumer
  • This will produce a large intermediate stream
    that requires potentially unbounded memory storage

x5 x4 x3 x2 x1
producer
consumer
48
Solutions
  • There are three alternatives
  • Play with the speed of the different threads,
    i.e. play with the scheduler to make the producer
    slower
  • Create a bounded buffer, say of size N, so that
    the producer waits automatically when the buffer
    is full
  • Use demand-driven approach, where the consumer
    activates the producer when it needs a new
    element (lazy evaluation)
  • The last two approaches introduce the notion of
    flow-control between concurrent activities (very
    common)

49
Time
  • In concurrent computation one would like to
    handle time
  • proc Time.delay T The running thread suspends
    for T milliseconds
  • proc Time.alarm T U Immediately creates its
    own thread, and binds U to unit after T
    milliseconds

50
Example
  • local
  • proc Ping N
  • for I in 1..N do
  • Delay 500 Browse ping
  • end
  • Browse 'ping terminate'
  • end
  • proc Pong N
  • for I in 1..N do
  • Delay 600 Browse pong
  • end
  • Browse 'pong terminate'
  • end
  • in .... end

local .... in Browse 'game started'
thread Ping 1000 end thread Pong 1000
end end
51
Thread Priority and Real Time
  • Try to run the program using the following
    statement
  • Consumer thread Producer 5000000 end
  • Switch on the panel and observe the memory
    behavior of the program.
  • You will quickly notice that this program does
    not behave well.
  • The reason has to do with the asynchronous
    message passing. If the producer sends messages
    i.e. create new elements in the stream, in a
    faster rate than the consumer can consume,
    increasingly more buffering will be needed until
    the system starts to break down.
  • One possible solution is to control
    experimentally the rate of thread execution so
    that the consumers get a larger time-slice than
    the producers do.

52
Priorities
  • There are three priority levels
  • high,
  • medium, and
  • low (the default)
  • A priority level determines how often a runnable
    thread is allocated a time slice.
  • In Oz, a high priority thread cannot starve a low
    priority one. Priority determines only how large
    piece of the processor-cake a thread can get.
  • Each thread has a unique name. To get the name of
    the current thread the procedure Thread.this/1 is
    called.
  • Having a reference to a thread, by using its
    name, enables operations on threads such as
  • terminating a thread, or
  • raising an exception in a thread.
  • Thread operations are defined the standard module
    Thread.

53
Thread priority and thread control
  • fun Thread.state T returns thread state
  • procThread.injectException T E exception E
    injected into thread
  • fun Thread.this returns 1st class
    reference to thread
  • procThread.setPriority T P P is high,
    medium or low
  • procThread.setThisPriority P as above on
    current thread
  • funProperty.get priorities get priority
    ratios
  • procProperty.put priorities(highH mediumM)

54
Thread Priorities
  • Oz has three priority levels. The system
    procedure
  • Property.put priorities p(mediumY highX)
  • Sets the processor-time ratio to X1 between
    high-priority threads and medium-priority thread.
  • It also sets the processor-time ratio to Y1
    between medium-priority threads and low-priority
    threads. X and Y are integers.
  • Example
  • Property.put priorities p(high10 medium10)
  • Now let us make our producer-consumer program
    work. We give the producer low priority, and the
    consumer high. We also set the priority ratios to
    101 and 101.

55
The program
  • local L in Property.put priorities p(high10
    medium10) thread Thread.setThisPriorit
    y low L Producer 5000000 end
    thread Thread.setThisPriority high
    Consumer L endend

56
Concurrent control abstraction
  • We have seen how threads are forked by thread
    ... end
  • A natural question to ask is how can we join
    threads?

fork
threads
join
57
Termination detection
  • This is a special case of detecting termination
    of multiple threads, and making another thread
    wait on that event.
  • The general scheme is quite easy because of
    dataflow variables
  • thread ?S1? X1 unit end thread ?S2?
    X2 X1 end ... thread ?Sn? Xn Xn-1
    end Wait Xn Continue main thread
  • When all threads terminate the variables X1 XN
    will be merged together labeling a single box
    that contains the value unit.
  • Wait XN suspends the main thread until XN is
    bound.

58
Concurrent Composition
  • conc S1 S2 Sn end
  • Conc proc S1 end proc S2
    end ... proc Sn end
  • Takes a single argument that is a list of nullary
    procedures.
  • When it is executed, the procedures are forked
    concurrently. The next statement is executed only
    when all procedures in the list terminate.

59
Conc
  • local proc Conc1 Ps I O case Ps of
    PPr then M in thread P M
    I end Conc1 Pr M O nil then O
    I end endin proc Conc Ps X
    in Conc1 Ps unit X Wait X
  • endend

This abstraction takes a list of
zero-argument procedures and terminate after all
these threads have terminated
60
Example
  • local
  • proc Ping N
  • for I in 1..N do
  • Delay 500 Browse ping
  • end
  • Browse 'ping terminate'
  • end
  • proc Pong N
  • for I in 1..N do
  • Delay 600 Browse pong
  • end
  • Browse 'pong terminate'
  • end
  • in .... end

local .... in Browse 'game started' Conc
proc Ping 1000 end proc Pong
1000 end Browse game terminated end
61
Futures
  • A future is a read-only capability of a
    single-assignment variable. For example to create
    a future of the variable X we perform the
    operation !! to create a future Y Y  !!X 
  • A thread trying to use the value of a future,
    e.g. using Y, will suspend until the variable of
    the future, e.g. X, gets bound.
  • One way to execute a procedure lazily, i.e. in a
    demand-driven manner, is to use the operation
    ByNeed P ?F.
  • ByNeed takes a zero-argument function P, and
    returns a future F. When a thread tries to access
    the value of F, the function P is called, and
    its result is bound to F.
  • This allows us to perform demand-driven
    computations in a straightforward manner.

62
Example
  • declare YByNeed fun  1 end YBrowse Y
  • we will observe that Y becomes a future, i.e. we
    will see YltFuturegt in the Browser.
  • If we try to access the value of Y, it will get
    bound to 1.
  • One way to access Y is by perform the operation
    Wait Y which triggers the producing procedure.

63
Why not always use declarative concurrency?
  • The concurrent declarative model is much simpler
  • Programs give the same results as if they were
    sequential, but they give the results
    incrementally (assuming a single binder per
    dataflow variable)
  • Why is this model so easy?
  • Because dataflow variables can be bound to only
    one value. A thread that shares a variable with
    another thread does not have to worry that the
    other thread will change the binding.
  • So why not stick with this model?
  • In many cases, we can stick with this model
  • But not always. For example, two clients that
    communicate with one server cannot be programmed
    in this model. Why not? Because there is an
    observable nondeterminism.
  • The concurrent declarative model is
    deterministic. If the program we write has an
    observable nondeterminism, then we cannot use the
    model.

64
Concurrent stateful model
?s? skip
empty statement ?x? ?y?

variable-variable binding
?x? ?v?
variable-value binding
?s1? ?s2?
sequential composition local ?x? in
?s1? end declaration proc ?x? ?y1?
?yn? ?s1? end procedure creation if ?x?
then ?s1? else ?s2? end conditional ?x?
?y1? ?yn? procedure application case
?x? of ?pattern? then ?s1? else ?s2? end
pattern matching NewName ?x? name
creation thread ?s? end thread
creation ByNeed ?x? ?y? trigger
creation try ?s1? catch ?x? then ?s2? end
exception context raise ?x? end
raise exception NewCell ?x? ?y?
cell creation Exchange ?x? ?y? ?z?
cell exchange
65
Concurrency and stateare tough when used together
  • Execution consists of multiple threads, all
    executing independently and all using shared
    cells
  • A threads execution is a sequence of Access and
    Assign operations (or Exchange operations)
  • Because of interleaving semantics, execution
    happens as if there was one global order of
    operations
  • Assume two threads and each thread does k
    operations. Then the total number of possible
    interleavings is This is exponential in
    k.
  • One can program by reasoning on all possible
    interleavings, but this is extremely hard. What
    do we do?

(
)
2k k
66
Programming withconcurrency and state
  • Programming with concurrency and state is largely
    a matter of reducing the number of interleavings,
    so that we can reason about programs in a simpler
    way. There are two basic approaches message
    passing and atomic actions.
  • Message passing with active objects Programs
    consist of threads that send asynchronous
    messages to each other. Each thread only
    receives a message when it is ready, which
    reduces the number of interleavings.
  • Atomic actions on shared state Programs consist
    of passive objects that are called by threads.
    We build large atomic actions (e.g., with locks,
    monitors, or transactions) to reduce the number
    of interleavings.

67
When to use each approach
  • Message passing useful for multi-agent
    applications, i.e., programs that consist of
    autonomous entities ( agents ,  actors or
     active objects ) that communicate with each
    other.
  • Atomic actions useful for data-centered
    applications, i.e., programs that consist of a
    large repository of data ( database  or
     shared state ) that is accessed and updated
    concurrently.
  • Both approaches can be used together in the same
    application, for different parts

68
Ports and cells
  • We have seen cells, the basic unit of
    encapsulated state, as a primitive concept
    underlying stateful and object-oriented
    programming. Cells are like variables in
    imperative languages.
  • Cells are the natural concept for programming
    with shared state
  • There is another way to add state to a language,
    which we call a port. A port is an asynchronous
    FIFO communication channel.
  • Ports are a natural concept for programming with
    active objects
  • Cells and ports are duals of each other
  • Each can be implemented with the other, so they
    are equal in expressiveness
  • Each is more natural in some circumstances
  • They are equivalent because each allows
    many-to-one communication (cell shared by
    threads, port shared by threads)

69
Ports
  • A port is an ADT with two operations
  • NewPort S P create a new port P with a new
    stream S. The stream is a list with unbound
    tail, used to model the FIFO nature of the
    communications channel.
  • Send P X send message X on port P. The
    message is appended to the stream S and can be
    read by threads reading S.
  • Example
  • declare P S inNewPort S PBrowse
    SthreadSend P 1end
  • threadSend P 2end

70
Building locks with cells
  • The basic way to program with shared state is by
    using locks
  • A lock is a region of the program that can only
    be occupied by one thread at a time. If a second
    thread attempts to enter, it will suspend until
    the first thread exits.
  • More sophisticated versions of locks are monitors
    and transactions
  • Monitors locks with a gating mechanism (e.g.,
    wait/notify in Java) to control which threads
    enter and exit and when. Monitors are the
    standard primitive for concurrent programming in
    Java.
  • Transactions locks that have two exits, a normal
    and abnormal exit. Upon abnormal exit (called
     abort ), all operations performed in the lock
    are undone, as if they were never done. Normal
    exit is called  commit  .
  • Locks can be built with cells. The idea is
    simple the cell contains a token. A thread
    attempting to enter the lock takes the token. A
    thread that finds no token will wait until the
    token is put back.

71
Building active objects with ports
  • Here is a simple active objectdeclare P
    inlocal Xs in NewPort Xs P thread ForAll Xs
    proc X Browse X end endendSend P
    foo(1)thread Send P bar(2) end

72
Defining ports with cells
  • A port is an unbundled stateful ADTproc
    NewPort S P CNewCell Sin PWrap
    Cendproc Send P X CUnwrap P Old
  • in Exchange C XOld Oldend

Anyone can do a send becauseanyone can do an
exchange
73
Active objects with classes
  • An active objects behavior can be defined by a
    class
  • The class is used to create a (passive) object,
    which is invoked by one thread that reads from a
    ports stream
  • Anyone can send a message to the object
    asynchronously, and the object will execute them
    one after the other, in sequential
    fashiondeclare ActObj inlocal Obj Xs P
    in ObjNew Class init NewPort Xs P thread
    ForAll Xs proc M Obj M end end proc
    ActObj M Send P M endendActObj msg(1)
  • Note that Obj M is synchronous and ActObj M
    is asynchronous!

74
Creating active objectswith NewActive
  • We can create a function NewActive that behaves
    like New except that it creates an active
    objectfun NewActive Class Init Obj Xs
    Pin ObjNew Class Init NewPort Xs
    P thread ForAll Xs proc M Obj M end
    end proc M Send P M endendActObj
    NewActive Class init

75
Making active objectssynchronous
  • We can make an active object synchronous by using
    a dataflow variable to store a result, and
    waiting for the result before continuingfun
    NewSynchronousActive Class Init Obj Xs
    Pin ObjNew Class Init NewPort Xs
    P thread ForAll Xs proc msg(M X) Obj M
    Xunit end end proc M X in Send P msg(M
    X) Wait X endend
  • This can be modified to handle when the active
    object raises an exception, to pass the exception
    back to the caller

76
Playing catch
ball
  • class Bounce attr other count0 meth
    init(Other) otherOther end meth ball
    count_at_count1 _at_other ball end meth
    get(X) X_at_count endend

B1
B2
ball
declare B1 B2 inB1NewActive Bounce
init(B2)B2NewActive Bounce init(B1) Get
the ball bouncingB1 ball Follow the
bouncesBrowse B1 get()
77
An area server
  • class AreaServer
  • meth init skip end meth square(X A)
    AXX end meth circle(R A)
    A3.14RR endend

declare S inSNewActive AreaServer init
Query the serverdeclare A inS square(10 A)
Browse Adeclare A in S circle(20 A)
Browse A
78
Event manager with active objects
  • An event manager contains a set of event handlers
  • Each handler is a triple IdFS where Id
    identifies it, F is the state update function,
    and S is the state
  • Reception of an event causes all triples to be
    replaced by IdFF E S (transition from F to F
    E S)
  • The manager EM is an active object with four
    methods
  • EM init initializes the event manager
  • EM event(E) posts event E at the manager
  • EM add(F S Id) adds new handler with F, S, and
    returns Id
  • EM delete(Id S) removed handler Id, returns
    state
  • This example taken from real use in Erlang

79
Defining the event manager
  • Mix of functional and object-oriented style

class EventManager attr handlers meth init
handlersnil end meth event(E)
handlers Map _at_handlers fun IdFS
IdFF E S end end meth add(F S Id)
IdNewName handlersIdFS_at_handlers
end meth delete(DId DS)
handlersList.partition _at_handlers fun
IdFS DIdId end __DS end end
State transition done using functional programming
80
Using the event manager
  • Simple memory-based handler keeps list of events

declare EM MemH Id in EMNewActive EventManager
init MemHfun E Buf EBuf end EM add(MemH
nil Id) EM event(a1) EM event(a2) ...
  • An event handler is purely functional, yet when
    put in the event manager, the latter is a
    concurrent imperative program. This is an
    example of separation of concerns by using
    multiple paradigms.

81
Exercises
  • VRH Exercise 4.11.3 (page 339)
  • Compare the sequential vs concurrent execution
    performance of equivalent SALSA programs.
  • VRH Exercise 4.11.5 (page 339)
  • SALSA asynchronous message passing enables to tag
    messages with properties priority, delay, and
    waitfor. Compare these mechanisms with Oz thread
    priorities, time delays and alarms, and futures.
  • How do SALSA tokens relate to Oz dataflow
    variables and futures?
  • What is the difference between multiple thread
    termination detection in Oz and join blocks in
    SALSA?

82
Exercises
  • Do Python, Java and C provide a linguistic
    abstraction for active objects? If so, which? If
    not, how would you go about implementing this
    abstraction?
  • Exercise VRH 7.9.6(a) (pg 568)
  • Write a Producer-Consumer program in SALSA (see
    VRH Section 4.3.1. for specification.)
Write a Comment
User Comments (0)
About PowerShow.com