QTcpServer with multiple persistent connections
-
Speaking of the code you referenced: in
ChatClient::onReadyRead()
, an infinite loop is started (viafor(;;) { ... }
). This will block the event loop until all data is transmitted, won't it?I read about the TCP communication via a QDataStream, that one can't expect all data to arrive at once. I'm not sure that I understood what's happening there correctly.
The fortune cookie example handles
readyRead()
without an infinite loop. What does happen there? Calling the function again won't reuse the stack variables created there, because they are deleted as soon as the function finishes (which won't happen in the inifinte loop approach). So will the data heap inside the socket, and each time a chunk arrives, thereadyRead
signal will be emitted, andstream >> variable
takes what it has and one can check if it's all?And what if the transmisson slows down or aborts, will the inifinite loop wait forever for the missing data?
-
Speaking of the code you referenced: in
ChatClient::onReadyRead()
, an infinite loop is started (viafor(;;) { ... }
). This will block the event loop until all data is transmitted, won't it?I read about the TCP communication via a QDataStream, that one can't expect all data to arrive at once. I'm not sure that I understood what's happening there correctly.
The fortune cookie example handles
readyRead()
without an infinite loop. What does happen there? Calling the function again won't reuse the stack variables created there, because they are deleted as soon as the function finishes (which won't happen in the inifinte loop approach). So will the data heap inside the socket, and each time a chunk arrives, thereadyRead
signal will be emitted, andstream >> variable
takes what it has and one can check if it's all?And what if the transmisson slows down or aborts, will the inifinite loop wait forever for the missing data?
Speaking of the code you referenced: in ChatClient::onReadyRead(), an infinite loop is started (via
for(;;) { ... })
. This will block the event loop until all data is transmitted, won't it?First, I'm an amateur, so I cannot promise this is right! I've looked at that code and it does seem strange at first understanding. I don't think it's working the way you think. I think the
for
loop does exit, and the wholereadyRead()
consequently, when there's insufficient data.} else { // the read failed, the socket goes automatically back to the state it was in before the transaction started // we just exit the loop and wait for more data to become available break; }
The transaction-y calls then make it so the socket behaves like the data read so far has not been read yet. You then re-enter the
readyRead()
when more data arrives. Then the new call to the slot doesn't maintain state about what was left over from last time, it simply gets to re-read the data from last time.Which then means all your questions about behaviour in multi-client are answered. The server does not block on one client.
-
Speaking of the code you referenced: in ChatClient::onReadyRead(), an infinite loop is started (via
for(;;) { ... })
. This will block the event loop until all data is transmitted, won't it?First, I'm an amateur, so I cannot promise this is right! I've looked at that code and it does seem strange at first understanding. I don't think it's working the way you think. I think the
for
loop does exit, and the wholereadyRead()
consequently, when there's insufficient data.} else { // the read failed, the socket goes automatically back to the state it was in before the transaction started // we just exit the loop and wait for more data to become available break; }
The transaction-y calls then make it so the socket behaves like the data read so far has not been read yet. You then re-enter the
readyRead()
when more data arrives. Then the new call to the slot doesn't maintain state about what was left over from last time, it simply gets to re-read the data from last time.Which then means all your questions about behaviour in multi-client are answered. The server does not block on one client.
Okay, after re-reading the code, I think the infinite loop is there in case multiple JSON "packets" are sent, and it's more than one available in one "readyRead" call.
Man, that TCP stuff is quite complicated ;-) But it's surely a huge help that this wiki entry exists. This really should go to the official docs as an example for persistent connections. That fortune cookie example is probably a good basic example for a single-use one-way connection, but there's so many things one has to think about if it's a two-way persistent connection, and esp. if there are multiple ones …
-
Speaking of the code you referenced: in
ChatClient::onReadyRead()
, an infinite loop is started (viafor(;;) { ... }
). This will block the event loop until all data is transmitted, won't it?I read about the TCP communication via a QDataStream, that one can't expect all data to arrive at once. I'm not sure that I understood what's happening there correctly.
The fortune cookie example handles
readyRead()
without an infinite loop. What does happen there? Calling the function again won't reuse the stack variables created there, because they are deleted as soon as the function finishes (which won't happen in the inifinte loop approach). So will the data heap inside the socket, and each time a chunk arrives, thereadyRead
signal will be emitted, andstream >> variable
takes what it has and one can check if it's all?And what if the transmisson slows down or aborts, will the inifinite loop wait forever for the missing data?
@l3u_ said in QTcpServer with multiple persistent connections:
This will block the event loop until all data is transmitted, won't it?
Nope, it has a different purpose. Also note that the socket is buffered itself (Qt-side) meaning the data is already read (or to be written) from (to) the actual device, but it sits inside Qt's buffer, in memory.
And what if the transmisson slows down or aborts, will the inifinite loop wait forever for the missing data?
No. The loop is there for another reason. You have to realize when working with the network (specific to TCP, not Qt), the data may arrive in chunks, granted, as you read already but there are 2 distinct cases:
- The data you sent may be split into pieces.
You have to handle partially reading a data block - this is where theQDataStream
transactions come in play. You start the transaction and try reading the data you had sent from the other side. Without going into too much details about the internals, if the data stream couldn't read the whole blockQDataStream::commitTransaction
will fail and is going to "restore" the internal buffer of the socket as if no reading had occurred at all. That's why it's enough in this case to do nothing - just wait for more data. - The data you sent may be appended into one bigger piece.
Now this means once you read a message successfully, i.e.QDataStream::commitTransaction
succeeds, and you process it you have still more in the socket's buffer. As it happens it may be yet a whole another message waiting there, but you're going to get only onereadyRead
emission, thus you need to make sure you loop around to read that second message as well. This is where the infinite loop comes in play. It just ensures all completely received messages that are sitting in the socket's buffer are read. If you have only one message, then the code in the loop does one iteration, otherwise it loops around until it extracts all completely received messages.
Okay, after re-reading the code, I think the infinite loop is there in case multiple JSON "packets" are sent, and it's more than one available in one "readyRead" call.
Yes. That is correct.
But it's surely a huge help that this wiki entry exists.
I'm sure @VRonin is going to be happy to read that (as well as I am).
@JonB said in QTcpServer with multiple persistent connections:
I don't think it's working the way you think. I think the for loop does exit, and the whole readyRead() consequently, when there's insufficient data.
Yes, you think correctly. ;)
The transaction-y calls then make it so the socket behaves like the data read so far has not been read yet. You then re-enter the readyRead() when more data arrives. Then the new call to the slot doesn't maintain state about what was left over from last time, it simply gets to re-read the data from last time.
Correct again.
- The data you sent may be split into pieces.
-
@l3u_ said in QTcpServer with multiple persistent connections:
This will block the event loop until all data is transmitted, won't it?
Nope, it has a different purpose. Also note that the socket is buffered itself (Qt-side) meaning the data is already read (or to be written) from (to) the actual device, but it sits inside Qt's buffer, in memory.
And what if the transmisson slows down or aborts, will the inifinite loop wait forever for the missing data?
No. The loop is there for another reason. You have to realize when working with the network (specific to TCP, not Qt), the data may arrive in chunks, granted, as you read already but there are 2 distinct cases:
- The data you sent may be split into pieces.
You have to handle partially reading a data block - this is where theQDataStream
transactions come in play. You start the transaction and try reading the data you had sent from the other side. Without going into too much details about the internals, if the data stream couldn't read the whole blockQDataStream::commitTransaction
will fail and is going to "restore" the internal buffer of the socket as if no reading had occurred at all. That's why it's enough in this case to do nothing - just wait for more data. - The data you sent may be appended into one bigger piece.
Now this means once you read a message successfully, i.e.QDataStream::commitTransaction
succeeds, and you process it you have still more in the socket's buffer. As it happens it may be yet a whole another message waiting there, but you're going to get only onereadyRead
emission, thus you need to make sure you loop around to read that second message as well. This is where the infinite loop comes in play. It just ensures all completely received messages that are sitting in the socket's buffer are read. If you have only one message, then the code in the loop does one iteration, otherwise it loops around until it extracts all completely received messages.
Okay, after re-reading the code, I think the infinite loop is there in case multiple JSON "packets" are sent, and it's more than one available in one "readyRead" call.
Yes. That is correct.
But it's surely a huge help that this wiki entry exists.
I'm sure @VRonin is going to be happy to read that (as well as I am).
@JonB said in QTcpServer with multiple persistent connections:
I don't think it's working the way you think. I think the for loop does exit, and the whole readyRead() consequently, when there's insufficient data.
Yes, you think correctly. ;)
The transaction-y calls then make it so the socket behaves like the data read so far has not been read yet. You then re-enter the readyRead() when more data arrives. Then the new call to the slot doesn't maintain state about what was left over from last time, it simply gets to re-read the data from last time.
Correct again.
Without going into too much details about the internals, if the data stream couldn't read the whole block QDataStream::commitTransaction will fail
That's the bit I wasn't sure about and had to guess! So, "Without going into too much details", how does it decide that (and where is it documented for us plebs)? Like, does it implement a timeout of some nature?
- The data you sent may be split into pieces.
-
Without going into too much details about the internals, if the data stream couldn't read the whole block QDataStream::commitTransaction will fail
That's the bit I wasn't sure about and had to guess! So, "Without going into too much details", how does it decide that (and where is it documented for us plebs)? Like, does it implement a timeout of some nature?
@JonB said in QTcpServer with multiple persistent connections:
how does it decide that (and where is it documented for us plebs)?
The method docs, I guess? Too lazy to check to be honest.
The decision is made based on theQDataStream
status. Whenever you're writing/reading from the stream, e.g. when you provide serialization for the objects (think<<
and>>
operators) you can (and should) set the status if the (de)serialization failed. This is already done for the Qt types anyway. Here, simply reading a byte array, the validation is rather straight-forward (internal to Qt):- Try to read the buffer size (i.e. an integer). If there's not enough data on the
QIODevice
to read an integer, the operation fails. - Try to read the actual buffer contents (you already know the size from 1). If there's not enough data on the device, the operation fails.
Like, does it implement a timeout of some nature?
No, it's simpler than that. The data is buffered in the socket object. So you just try to read it from that memory buffer. If reading fails, the transaction fails. If reading succeeds the buffer is "shortened" with the data you read. The buffer is filled asynchronously by Qt whenever the data comes through the underlying system socket.
- Try to read the buffer size (i.e. an integer). If there's not enough data on the
-
@JonB said in QTcpServer with multiple persistent connections:
how does it decide that (and where is it documented for us plebs)?
The method docs, I guess? Too lazy to check to be honest.
The decision is made based on theQDataStream
status. Whenever you're writing/reading from the stream, e.g. when you provide serialization for the objects (think<<
and>>
operators) you can (and should) set the status if the (de)serialization failed. This is already done for the Qt types anyway. Here, simply reading a byte array, the validation is rather straight-forward (internal to Qt):- Try to read the buffer size (i.e. an integer). If there's not enough data on the
QIODevice
to read an integer, the operation fails. - Try to read the actual buffer contents (you already know the size from 1). If there's not enough data on the device, the operation fails.
Like, does it implement a timeout of some nature?
No, it's simpler than that. The data is buffered in the socket object. So you just try to read it from that memory buffer. If reading fails, the transaction fails. If reading succeeds the buffer is "shortened" with the data you read. The buffer is filled asynchronously by Qt whenever the data comes through the underlying system socket.
@kshegunov
OK, so in a word,QDataStream::commitTransaction()
fails/rolls back if there is not enough data already there at the instant it is called, period.BTW: so if I decide to transfer a 1GB (or more) file content as one
QByteArray
, and it arrives in chunks, your transaction approach will have to buffer up to 1GB of previously-received bytes, and that's presumably in memory? Hmmmm. Nobody finds that a problem? - Try to read the buffer size (i.e. an integer). If there's not enough data on the
-
@kshegunov
OK, so in a word,QDataStream::commitTransaction()
fails/rolls back if there is not enough data already there at the instant it is called, period.BTW: so if I decide to transfer a 1GB (or more) file content as one
QByteArray
, and it arrives in chunks, your transaction approach will have to buffer up to 1GB of previously-received bytes, and that's presumably in memory? Hmmmm. Nobody finds that a problem?@JonB said in QTcpServer with multiple persistent connections:
OK, so in a word, QDataStream::commitTransaction() fails/rolls back if there is not enough data already there at the instant it is called, period.
Correct
so if I decide to transfer a 1GB (or more) file content as one QByteArray, and it arrives in chunks, your transaction approach will have to buffer up to 1GB of previously-received bytes, and that's presumably in memory? Hmmmm. Nobody finds that a problem?
In this case you are free to build your own buffer (i'll use the chat example again):
- add a private
QTemporaryFile m_dataBuffer;
toChatClient
; - in the constructor of
ChatClient
callm_dataBuffer.open();
- modify
ChatClient::onReadyRead()
as below
void ChatClient::onReadyRead() { const qint64 oldPos = m_dataBuffer.pos(); // save the position in the file m_dataBuffer.seek(m_dataBuffer.size()); // go to the end of the file m_dataBuffer.write(m_clientSocket.readAll()); // append all the data available on the socket into the file m_dataBuffer.seek(oldPos); // go back to the old position QDataStream socketStream(m_dataBuffer); // set the stream to read from the file //////////////////////////////////////////////////////////////////// // identical to the original version QByteArray jsonData; socketStream.setVersion(QDataStream::Qt_5_7); for (;;) { socketStream.startTransaction(); socketStream >> jsonData; if (!socketStream.commitTransaction()) break; QJsonParseError parseError; const QJsonDocument jsonDoc = QJsonDocument::fromJson(jsonData, &parseError); if (parseError.error == QJsonParseError::NoError) { if (jsonDoc.isObject()) jsonReceived(jsonDoc.object()); } } //////////////////////////////////////////////////////////////////// if(m_dataBuffer.atEnd()) // we read all the data there was on the file m_dataBuffer.resize(0); // clear everything from the file }
- add a private
-
@kshegunov
OK, so in a word,QDataStream::commitTransaction()
fails/rolls back if there is not enough data already there at the instant it is called, period.BTW: so if I decide to transfer a 1GB (or more) file content as one
QByteArray
, and it arrives in chunks, your transaction approach will have to buffer up to 1GB of previously-received bytes, and that's presumably in memory? Hmmmm. Nobody finds that a problem?@JonB said in QTcpServer with multiple persistent connections:
so if I decide to transfer a 1GB (or more) file content as one QByteArray, and it arrives in chunks, your transaction approach will have to buffer up to 1GB of previously-received bytes, and that's presumably in memory? Hmmmm. Nobody finds that a problem?
Not really, as you already committed a heinous crime against the system already: you read 1GB off the hard disk and send it directly to the socket. ;)
Joking aside, what @VRonin said could work. Or you could simply split it up at the peer before sending. -
@JonB said in QTcpServer with multiple persistent connections:
so if I decide to transfer a 1GB (or more) file content as one QByteArray, and it arrives in chunks, your transaction approach will have to buffer up to 1GB of previously-received bytes, and that's presumably in memory? Hmmmm. Nobody finds that a problem?
Not really, as you already committed a heinous crime against the system already: you read 1GB off the hard disk and send it directly to the socket. ;)
Joking aside, what @VRonin said could work. Or you could simply split it up at the peer before sending.@kshegunov
I realise @VRonin's reply is the way to go, and have already upvoted that.However, your
as you already committed a heinous crime against the system already: you read 1GB off the hard disk and send it directly to the socket. ;)
? My "TCP clients" in this case were the ones who read data off the disk and sent them to my server. How they did that is up to them, I have no knowledge, where's the crime? What's that got to do with your approach of having the server receive the 1GB into a memory buffer? And at least the clients, if they did read into memory, only did 1GB each. The server services 100 clients, they're all in the process of sending 1GB each, my server needs 100GB memory.... :(
-
@kshegunov
I realise @VRonin's reply is the way to go, and have already upvoted that.However, your
as you already committed a heinous crime against the system already: you read 1GB off the hard disk and send it directly to the socket. ;)
? My "TCP clients" in this case were the ones who read data off the disk and sent them to my server. How they did that is up to them, I have no knowledge, where's the crime? What's that got to do with your approach of having the server receive the 1GB into a memory buffer? And at least the clients, if they did read into memory, only did 1GB each. The server services 100 clients, they're all in the process of sending 1GB each, my server needs 100GB memory.... :(
-
@kshegunov
I realise @VRonin's reply is the way to go, and have already upvoted that.However, your
as you already committed a heinous crime against the system already: you read 1GB off the hard disk and send it directly to the socket. ;)
? My "TCP clients" in this case were the ones who read data off the disk and sent them to my server. How they did that is up to them, I have no knowledge, where's the crime? What's that got to do with your approach of having the server receive the 1GB into a memory buffer? And at least the clients, if they did read into memory, only did 1GB each. The server services 100 clients, they're all in the process of sending 1GB each, my server needs 100GB memory.... :(
@JonB I think what he wanted to say is that your protocol shouldn't allow sending huge packets of data in one big piece at all, and that – if you want to handle such amounts – you will have to implement it in the proper way so that everything is fine. Because if you send 1 GB to a socket using the default behavior (which will probably result in the reveiver's socket buffer filling up to 1 GB until you can read it out), you will have to read it into memory before. And you shouldn't have done that in the first place.
-
@VRonin @kshegunov My stuff is far from finished, but at this point, I can say that this forum post and esp. the chat example helped like 100 times more than the official docs.
I think the chat example really should be part of the docs, as a more sophisticated example, and one for persistent connections. The fortune cookie example is nice when it comes to a very simple, one-time-use one-way connection. But somebody like me not having programmed TCP stuff at all until now simply can't figure out how to implement a bidirectional communication with multiple clients from it.
I'm a big step closer to being able to implement what I need thanks to your help and work, and I'm pretty sure I can learn quite a lot about TCP conenctions and handling them here :-)
-
@JonB I think what he wanted to say is that your protocol shouldn't allow sending huge packets of data in one big piece at all, and that – if you want to handle such amounts – you will have to implement it in the proper way so that everything is fine. Because if you send 1 GB to a socket using the default behavior (which will probably result in the reveiver's socket buffer filling up to 1 GB until you can read it out), you will have to read it into memory before. And you shouldn't have done that in the first place.
@l3u_
I have written TCP client/server in previous products to allow transfer of files. They seem to have worked fine for years. Neither the client nor the server read the entire file into memory to send or receive. That's all I meant.Because if you send 1 GB to a socket using the default behavior (which will probably result in the reveiver's socket buffer filling up to 1 GB until you can read it out),
Umm, will it?? I don't think so... Pipes wouldn't do well...
-
@l3u_
I have written TCP client/server in previous products to allow transfer of files. They seem to have worked fine for years. Neither the client nor the server read the entire file into memory to send or receive. That's all I meant.Because if you send 1 GB to a socket using the default behavior (which will probably result in the reveiver's socket buffer filling up to 1 GB until you can read it out),
Umm, will it?? I don't think so... Pipes wouldn't do well...
@JonB I'm totally fine with everything and everybody here – I'm the one that surely has the least experience with all this here ;-) And I won't be transferring stuff that is bigger than a few hundred characters of JSON data, so the problem is surely interesting, but I don't think I have to worry about it in my case.
-
@l3u_
I have written TCP client/server in previous products to allow transfer of files. They seem to have worked fine for years. Neither the client nor the server read the entire file into memory to send or receive. That's all I meant.Because if you send 1 GB to a socket using the default behavior (which will probably result in the reveiver's socket buffer filling up to 1 GB until you can read it out),
Umm, will it?? I don't think so... Pipes wouldn't do well...
@JonB said in QTcpServer with multiple persistent connections:
Umm, will it?? I don't think so... Pipes wouldn't do well...
Probably, yes. Qt has its own buffering on top of the socket layer, so unless you change that, by default the socket's buffer will accommodate it provided you have enough memory.
@l3u_ said in QTcpServer with multiple persistent connections:
I'm the one that surely has the least experience with all this here
Don't worry, give it a decade or two and you'd be the one giving the advice. ;P
-
@JonB said in QTcpServer with multiple persistent connections:
Umm, will it?? I don't think so... Pipes wouldn't do well...
Probably, yes. Qt has its own buffering on top of the socket layer, so unless you change that, by default the socket's buffer will accommodate it provided you have enough memory.
@l3u_ said in QTcpServer with multiple persistent connections:
I'm the one that surely has the least experience with all this here
Don't worry, give it a decade or two and you'd be the one giving the advice. ;P
@kshegunov said in QTcpServer with multiple persistent connections:
Qt has its own buffering on top of the socket layer
So if you then use socket "transactions" on top of the socket, I read that as doing the buffering in memory. Do you mean the transactions have to add another layer of buffering, or do they use whatever the
QTcpSocket
is already using? -
@kshegunov said in QTcpServer with multiple persistent connections:
Qt has its own buffering on top of the socket layer
So if you then use socket "transactions" on top of the socket, I read that as doing the buffering in memory. Do you mean the transactions have to add another layer of buffering, or do they use whatever the
QTcpSocket
is already using?@JonB said in QTcpServer with multiple persistent connections:
So if you then use socket "transactions" on top of the socket, I read that as doing the buffering in memory. Do you mean the transactions have to add another layer of buffering, or do they use whatever the QTcpSocket is already using?
They use the QAbstractSocket's buffer directly. Think of the transaction as only adding atomicity to the reading from the Qt's buffer. They do stuff similar to what @VRonin wrote about splitting your big data chunk: try to read it in the known format, if that fails, rewind the buffer's pointer. If it succeeds cut the internal buffer with what you read.
-
Sorry to jump in late here, but I've been grooving on WebSockets lately
(a stupid name for a good protocol). Very easy to setup and use in Qt 5. Connect a couple of signals and slots and you can reliably send and receive text and/or binary data. You can ignore the grind of TCP/IP and just mostly deal with your text. -
Sorry to jump in late here, but I've been grooving on WebSockets lately
(a stupid name for a good protocol). Very easy to setup and use in Qt 5. Connect a couple of signals and slots and you can reliably send and receive text and/or binary data. You can ignore the grind of TCP/IP and just mostly deal with your text.That is true. There are 2 conditions to do so, however:
- You want to commit to the websockets protocol, which isn't always the case.
- You are willing to pull a module outside of qtbase, which is fine most of the time.
To expand on 1:
In the project I'm currently working on I have to support multiple protocols, so websockets does me no good. You may be surprised, but that case is more common than not.