Page Menu
Home
Phabricator
Search
Configure Global Search
Log In
Files
F7170765
session.py
No One
Temporary
Actions
Download File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Award Token
Flag For Later
Size
17 KB
Referenced Files
None
Subscribers
None
session.py
View Options
# Copyright (C) 2008-2009 AG Projects. See LICENSE for details
from
__future__
import
with_statement
import
sys
import
random
import
traceback
from
time
import
time
from
twisted.internet.error
import
ConnectionClosed
,
ConnectionDone
from
twisted.python.failure
import
Failure
from
gnutls.errors
import
GNUTLSError
from
eventlet
import
api
,
coros
,
proc
from
eventlet.twistedutil.protocol
import
ValueQueue
from
msrplib
import
protocol
,
MSRPError
from
msrplib.transport
import
make_response
,
MSRPTransactionError
from
msrplib.protocol
import
StatusHeader
,
ContentTypeHeader
ConnectionClosedErrors
=
(
ConnectionClosed
,
GNUTLSError
)
class
MSRPSessionError
(
MSRPError
):
pass
class
MSRPBadContentType
(
MSRPTransactionError
):
code
=
415
comment
=
'Unsupported media type'
class
LocalResponse
(
MSRPTransactionError
):
def
__repr__
(
self
):
return
'<LocalResponse
%s
%s
>'
%
(
self
.
code
,
self
.
comment
)
Response200OK
=
LocalResponse
(
"OK"
,
200
)
Response408Timeout
=
LocalResponse
(
"Timed out while waiting for transaction response"
,
408
)
def
contains_mime_type
(
mimetypelist
,
mimetype
):
"""Return True if mimetypelist contains mimetype.
mimietypelist either contains the complete mime types, such as 'text/plain',
or simple patterns, like 'text/*', or simply '*'.
"""
mimetype
=
mimetype
.
lower
()
for
pattern
in
mimetypelist
:
pattern
=
pattern
.
lower
()
if
pattern
==
'*'
:
return
True
if
pattern
==
mimetype
:
return
True
if
pattern
.
endswith
(
'/*'
)
and
mimetype
.
startswith
(
pattern
[:
-
1
]):
return
True
return
False
class
OutgoingFile
(
object
):
# will not implement support for CPIM here, because it's already complicated
# instead, use a file wrapper that will concatenate together CPIM header and the actual file
def
__init__
(
self
,
fileobj
,
size
,
content_type
=
None
,
position
=
0
,
message_id
=
None
):
self
.
fileobj
=
fileobj
self
.
size
=
size
self
.
position
=
position
if
message_id
is
None
:
message_id
=
'
%x
'
%
random
.
getrandbits
(
64
)
self
.
message_id
=
message_id
self
.
headers
=
{}
if
content_type
is
not
None
:
self
.
headers
[
ContentTypeHeader
.
name
]
=
ContentTypeHeader
(
content_type
)
def
on_transaction_response
(
self
,
chunk
):
# do something meaningful in case of an error?
pass
# TODO: how to cancel/pause file transfer? def cancel(self): ...
GOT_NEW_FILE
=
object
()
class
MSRPSession
(
object
):
# if incoming chunk is bigger than this, split it (for the reporting purposes)
INCOMING_CHUNK_SIZE
=
1024
*
16
RESPONSE_TIMEOUT
=
30
SHUTDOWN_TIMEOUT
=
1
# when sending a file, a chunk of this size will be read and sent uninterruptibly
FILE_PIECE_SIZE
=
1024
*
4
def
__init__
(
self
,
msrptransport
,
accept_types
=
[
'*'
],
on_incoming_cb
=
None
):
self
.
msrp
=
msrptransport
self
.
accept_types
=
accept_types
if
on_incoming_cb
is
not
None
:
self
.
_on_incoming_cb
=
on_incoming_cb
self
.
expected_responses
=
{}
self
.
outgoing
=
coros
.
queue
()
self
.
outgoing_files
=
coros
.
queue
()
self
.
reader_job
=
proc
.
spawn
(
self
.
_reader
)
self
.
writer_job
=
proc
.
spawn
(
self
.
_writer
)
self
.
state
=
'CONNECTED'
# -> 'FLUSHING' -> 'CLOSING' -> 'DONE'
# in FLUSHING writer sends only while there's something in the outgoing queue
# then it exits and sets state to 'CLOSING' which makes reader only pay attention
# to responses and success reports. (XXX it could now discard incoming data chunks
# with direct write() since writer is dead)
self
.
reader_job
.
link
(
self
.
writer_job
)
self
.
last_expected_response
=
0
if
not
callable
(
self
.
_on_incoming_cb
):
raise
TypeError
(
'on_incoming_cb must be callable:
%r
'
%
(
self
.
_on_incoming_cb
,
))
def
_get_logger
(
self
):
return
self
.
msrp
.
logger
def
_set_logger
(
self
,
logger
):
self
.
msrp
.
logger
=
logger
logger
=
property
(
_get_logger
,
_set_logger
)
def
set_state
(
self
,
state
):
self
.
logger
.
debug
(
'
%s
(was
%s
)'
%
(
state
,
self
.
state
))
self
.
state
=
state
@property
def
connected
(
self
):
return
self
.
state
==
'CONNECTED'
def
shutdown
(
self
,
wait
=
True
):
"""Send the messages already in queue then close the connection"""
self
.
set_state
(
'FLUSHING'
)
self
.
outgoing
.
send
(
None
)
if
wait
:
self
.
writer_job
.
wait
(
None
,
None
)
self
.
reader_job
.
wait
(
None
,
None
)
def
_handle_incoming_response
(
self
,
chunk
):
try
:
response_cb
,
timer
=
self
.
expected_responses
.
pop
(
chunk
.
transaction_id
)
except
KeyError
:
pass
else
:
if
timer
is
not
None
:
timer
.
cancel
()
response_cb
(
chunk
)
def
_check_incoming_SEND
(
self
,
chunk
):
if
chunk
.
method
==
'SEND'
and
chunk
.
segment
is
None
:
error
=
self
.
msrp
.
check_incoming_SEND_chunk
(
chunk
)
if
error
is
not
None
:
return
error
if
chunk
.
headers
.
get
(
'Content-Type'
)
is
None
:
return
MSRPBadContentType
(
'Content-type header missing'
)
if
not
contains_mime_type
(
self
.
accept_types
,
chunk
.
headers
[
'Content-Type'
]
.
decoded
):
return
MSRPBadContentType
def
_handle_incoming_SEND
(
self
,
chunk
):
error
=
self
.
_check_incoming_SEND
(
chunk
)
if
error
is
None
:
code
,
comment
=
200
,
'OK'
else
:
code
,
comment
=
error
.
code
,
error
.
comment
if
chunk
.
final
:
response
=
make_response
(
chunk
,
code
,
comment
)
if
response
is
not
None
:
self
.
outgoing
.
send
((
response
,
None
))
if
code
==
200
:
self
.
_on_incoming_cb
(
chunk
)
if
chunk
.
final
and
chunk
.
success_report
==
'yes'
:
report
=
self
.
msrp
.
make_chunk
(
method
=
'REPORT'
,
message_id
=
chunk
.
message_id
)
report
.
add_header
(
StatusHeader
(
'000 200 OK'
))
byterange
=
chunk
.
headers
.
get
(
'Byte-Range'
)
if
byterange
is
not
None
:
report
.
add_header
(
byterange
)
self
.
outgoing
.
send
((
report
,
None
))
def
_reader
(
self
):
"""Wait forever for new chunks. Notify the user about the good ones through self._on_incoming_cb.
If a response to a previously sent chunk is received, pop the corresponding
response_cb from self.expected_responses and send the response there.
"""
error
=
Failure
(
ConnectionDone
())
try
:
self
.
writer_job
.
link
(
self
.
reader_job
)
try
:
while
self
.
state
in
[
'CONNECTED'
,
'FLUSHING'
]:
chunk
=
self
.
msrp
.
read_chunk
(
self
.
INCOMING_CHUNK_SIZE
)
if
chunk
.
method
is
None
:
# response
self
.
_handle_incoming_response
(
chunk
)
elif
chunk
.
method
==
'SEND'
:
self
.
_handle_incoming_SEND
(
chunk
)
elif
chunk
.
method
==
'REPORT'
:
self
.
_on_incoming_cb
(
chunk
)
else
:
response
=
make_response
(
chunk
,
'501'
,
'Method unknown'
)
self
.
outgoing
.
send
((
response
,
None
))
except
proc
.
LinkedExited
:
# writer has exited
pass
finally
:
self
.
writer_job
.
unlink
(
self
.
reader_job
)
self
.
writer_job
.
kill
()
self
.
logger
.
debug
(
'reader: expecting responses only'
)
delay
=
time
()
-
self
.
last_expected_response
if
delay
>=
0
and
self
.
expected_responses
:
# continue read the responses until the last timeout expires
with
api
.
timeout
(
delay
,
None
):
while
self
.
expected_responses
:
chunk
=
self
.
msrp
.
read_chunk
()
if
chunk
.
method
is
None
:
self
.
_handle_incoming_response
(
chunk
)
else
:
self
.
logger
.
debug
(
'dropping incoming
%r
'
%
chunk
)
# read whatever left in the queue
with
api
.
timeout
(
0
,
None
):
while
self
.
msrp
.
_queue
:
chunk
=
self
.
msrp
.
read_chunk
()
if
chunk
.
method
is
None
:
self
.
_handle_incoming_response
(
chunk
)
else
:
self
.
logger
.
debug
(
'dropping incoming
%r
'
%
chunk
)
self
.
logger
.
debug
(
'reader: done'
)
except
ConnectionClosedErrors
,
ex
:
self
.
logger
.
debug
(
'reader: exiting because of
%r
'
%
ex
)
error
=
Failure
(
ex
)
except
:
self
.
logger
.
debug
(
'reader: losing connection because of
%r
'
%
(
sys
.
exc_info
(),
))
error
=
Failure
()
raise
finally
:
self
.
_on_incoming_cb
(
error
=
error
)
self
.
msrp
.
loseConnection
(
wait
=
False
)
self
.
set_state
(
'DONE'
)
def
_write_file
(
self
):
outgoing_file
=
self
.
outgoing_files
.
items
[
0
][
0
]
if
outgoing_file
.
size
is
None
or
outgoing_file
.
position
<
outgoing_file
.
size
:
try
:
piece
=
outgoing_file
.
fileobj
.
read
(
self
.
FILE_PIECE_SIZE
)
except
Exception
:
self
.
logger
.
err
(
'error while reading file
%r
\n
%s
'
%
(
outgoing_file
,
traceback
.
format_exc
()))
return
self
.
logger
.
debug
(
'_write_file: read
%s
bytes from
%s
'
%
(
len
(
piece
),
outgoing_file
))
chunk
=
self
.
msrp
.
make_chunk
(
contflag
=
'X'
,
start
=
outgoing_file
.
position
+
1
,
end
=
'*'
,
length
=
outgoing_file
.
size
or
'*'
,
message_id
=
outgoing_file
.
message_id
)
chunk
.
headers
.
update
(
outgoing_file
.
headers
)
id
=
chunk
.
transaction_id
assert
id
not
in
self
.
expected_responses
,
"MSRP transaction
%r
is already in progress"
%
id
cb_and_timer
=
[
outgoing_file
.
on_transaction_response
,
None
]
self
.
expected_responses
[
id
]
=
cb_and_timer
try
:
self
.
msrp
.
write
(
chunk
.
encode_start
())
self
.
logger
.
debug
(
'_write_file: wrote header
%r
'
%
chunk
)
try
:
while
piece
:
self
.
msrp
.
write
(
piece
)
outgoing_file
.
position
+=
len
(
piece
)
#self.logger.debug('_write_file: wrote %s bytes' % len(piece))
for
x
in
self
.
outgoing
.
items
:
# stop sending the file if there is something else in the queue other than files
if
x
is
not
GOT_NEW_FILE
:
return
if
outgoing_file
.
size
is
None
or
outgoing_file
.
position
<
outgoing_file
.
size
:
try
:
piece
=
outgoing_file
.
fileobj
.
read
(
self
.
FILE_PIECE_SIZE
)
except
Exception
:
self
.
logger
.
err
(
'error while reading file
%r
\n
%s
'
%
(
outgoing_file
,
traceback
.
format_exc
()))
return
else
:
break
finally
:
if
not
piece
or
outgoing_file
.
position
==
outgoing_file
.
size
:
contflag
=
'$'
if
self
.
outgoing_files
:
self
.
outgoing_files
.
wait
()
else
:
contflag
=
'+'
self
.
msrp
.
write
(
chunk
.
encode_end
(
contflag
))
self
.
logger
.
debug
(
'_write_file: wrote chunk end
%s
'
%
contflag
)
except
:
self
.
expected_responses
.
pop
(
id
,
None
)
raise
else
:
timeout_error
=
Response408Timeout
if
chunk
.
failure_report
==
'yes'
else
Response200OK
timer
=
api
.
get_hub
()
.
schedule_call_global
(
self
.
RESPONSE_TIMEOUT
,
self
.
_response_timeout
,
id
,
timeout_error
)
self
.
last_expected_response
=
time
()
+
self
.
RESPONSE_TIMEOUT
cb_and_timer
[
1
]
=
timer
def
_writer
(
self
):
try
:
while
self
.
state
==
'CONNECTED'
or
(
self
.
state
==
'FLUSHING'
and
self
.
outgoing
and
self
.
outgoing_files
):
if
not
self
.
outgoing
and
self
.
outgoing_files
:
send_params
=
GOT_NEW_FILE
else
:
send_params
=
self
.
outgoing
.
wait
()
if
send_params
is
None
:
break
elif
send_params
is
GOT_NEW_FILE
and
self
.
outgoing_files
:
self
.
_write_file
()
else
:
self
.
_write_chunk
(
*
send_params
)
except
ConnectionClosedErrors
+
(
proc
.
LinkedExited
,
proc
.
ProcExit
),
ex
:
self
.
logger
.
debug
(
'writer: exiting because of
%r
'
%
(
ex
,
))
except
:
self
.
logger
.
debug
(
'writer: losing connection because of
%r
'
%
(
sys
.
exc_info
(),
))
raise
finally
:
self
.
msrp
.
loseConnection
(
wait
=
False
)
self
.
set_state
(
'CLOSING'
)
def
send_chunk
(
self
,
chunk
,
response_cb
=
None
):
"""Send `chunk'. Report the result via `response_cb'.
When `response_cb' argument is present, it will be used to report
the transaction response to the caller. When a response is received
or generated locally, `response_cb' is called with one argument. The function
must do something quickly and must not block, because otherwise it would
the reader greenlet.
If no response was received after RESPONSE_TIMEOUT seconds,
* 408 response is generated if Failure-Report was 'yes' or absent
* 200 response is generated if Failure-Report was 'partial' or 'no'
Note that it's rather wasteful to provide `response_cb' argument other than None
for chunks with Failure-Report='no' since it will always fire 30 seconds later
with 200 result (unless the other party is broken and ignores Failure-Report header)
If sending is impossible raise MSRPSessionError.
"""
assert
id
not
in
self
.
expected_responses
,
"MSRP transaction
%r
is already in progress"
%
id
if
response_cb
is
not
None
and
not
callable
(
response_cb
):
raise
TypeError
(
'response_cb must be callable:
%r
'
%
(
response_cb
,
))
if
self
.
state
!=
'CONNECTED'
:
raise
MSRPSessionError
(
'Cannot send chunk because MSRPSession is
%s
'
%
self
.
state
)
if
self
.
msrp
.
_disconnected_event
.
ready
():
raise
MSRPSessionError
(
str
(
self
.
msrp
.
_disconnected_event
.
wait
()))
self
.
outgoing
.
send
((
chunk
,
response_cb
))
def
send_file
(
self
,
outgoing_file
):
if
self
.
state
!=
'CONNECTED'
:
raise
MSRPSessionError
(
'Cannot send chunk because MSRPSession is
%s
'
%
self
.
state
)
if
self
.
msrp
.
_disconnected_event
.
ready
():
raise
MSRPSessionError
(
str
(
self
.
msrp
.
_disconnected_event
.
wait
()))
self
.
outgoing_files
.
send
(
outgoing_file
)
self
.
outgoing
.
send
(
GOT_NEW_FILE
)
# just to wake up the writer greenlet
def
_write_chunk
(
self
,
chunk
,
response_cb
=
None
):
id
=
chunk
.
transaction_id
assert
id
not
in
self
.
expected_responses
,
"MSRP transaction
%r
is already in progress"
%
id
if
response_cb
is
not
None
:
# since reader is in another greenlet and write() will switch the greenlets,
# let's setup response_cb and timer before write() call, just in case
cb_and_timer
=
[
response_cb
,
None
]
self
.
expected_responses
[
id
]
=
cb_and_timer
try
:
self
.
msrp
.
write
(
chunk
.
encode
())
except
:
if
response_cb
is
not
None
:
self
.
expected_responses
.
pop
(
id
,
None
)
raise
else
:
if
response_cb
is
not
None
:
timeout_error
=
Response408Timeout
if
chunk
.
failure_report
==
'yes'
else
Response200OK
timer
=
api
.
get_hub
()
.
schedule_call_global
(
self
.
RESPONSE_TIMEOUT
,
self
.
_response_timeout
,
id
,
timeout_error
)
self
.
last_expected_response
=
time
()
+
self
.
RESPONSE_TIMEOUT
cb_and_timer
[
1
]
=
timer
def
_response_timeout
(
self
,
id
,
timeout_error
):
response_cb
,
timer
=
self
.
expected_responses
.
pop
(
id
,
(
None
,
None
))
if
response_cb
is
not
None
:
response_cb
(
timeout_error
)
if
timer
is
not
None
:
timer
.
cancel
()
def
deliver_chunk
(
self
,
chunk
,
event
=
None
):
"""Send chunk, wait for the transaction response (if Failure-Report header is not 'no').
Return the transaction response if it's a success, raise MSRPTransactionError if it's not.
If chunk's Failure-Report is 'no', return None immediately.
"""
if
chunk
.
failure_report
!=
'no'
and
event
is
None
:
event
=
coros
.
event
()
self
.
send_chunk
(
chunk
,
event
.
send
)
if
event
is
not
None
:
response
=
event
.
wait
()
if
isinstance
(
response
,
Exception
):
raise
response
elif
200
<=
response
.
code
<=
299
:
return
response
raise
MSRPTransactionError
(
comment
=
response
.
comment
,
code
=
response
.
code
)
def
make_message
(
self
,
msg
,
content_type
,
message_id
=
None
):
chunk
=
self
.
msrp
.
make_chunk
(
data
=
msg
,
message_id
=
message_id
)
chunk
.
add_header
(
protocol
.
ContentTypeHeader
(
content_type
))
return
chunk
def
send_message
(
self
,
msg
,
content_type
):
chunk
=
self
.
make_message
(
msg
,
content_type
)
self
.
send_chunk
(
chunk
)
return
chunk
def
deliver_message
(
self
,
msg
,
content_type
):
chunk
=
self
.
make_message
(
msg
,
content_type
)
self
.
deliver_chunk
(
chunk
)
return
chunk
class
GreenMSRPSession
(
MSRPSession
):
def
__init__
(
self
,
msrptransport
,
accept_types
=
[
'*'
]):
MSRPSession
.
__init__
(
self
,
msrptransport
,
accept_types
)
self
.
incoming
=
ValueQueue
()
def
receive_chunk
(
self
):
return
self
.
incoming
.
wait
()
def
_on_incoming_cb
(
self
,
value
=
None
,
error
=
None
):
if
error
is
not
None
:
self
.
incoming
.
send_exception
(
error
.
type
,
error
.
value
,
error
.
tb
)
else
:
self
.
incoming
.
send
(
value
)
# TODO:
# 413 - requires special action both in reader and in writer
# continuation: #
# All MSRP endpoints MUST be able to receive the multipart/mixed [15] and multipart/alternative [15] media-types.
File Metadata
Details
Attached
Mime Type
text/x-python
Expires
Tue, Nov 26, 5:18 AM (1 d, 8 h)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
3414065
Default Alt Text
session.py (17 KB)
Attached To
Mode
rPYMSRPLIB python3-msrplib
Attached
Detach File
Event Timeline
Log In to Comment