mirror of
https://github.com/badaix/snapcast.git
synced 2025-05-27 16:06:15 +02:00
add clang-format file
reformat code
This commit is contained in:
parent
b733f646ea
commit
b20add3815
105 changed files with 7773 additions and 7723 deletions
|
@ -1,6 +1,6 @@
|
|||
/***
|
||||
This file is part of snapcast
|
||||
Copyright (C) 2014-2018 Johannes Pohl
|
||||
Copyright (C) 2014-2019 Johannes Pohl
|
||||
|
||||
This program is free software: you can redistribute it and/or modify
|
||||
it under the terms of the GNU General Public License as published by
|
||||
|
@ -16,221 +16,222 @@
|
|||
along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
***/
|
||||
|
||||
#include "clientConnection.h"
|
||||
#include "aixlog.hpp"
|
||||
#include "common/snapException.h"
|
||||
#include "common/strCompat.h"
|
||||
#include "message/hello.h"
|
||||
#include <iostream>
|
||||
#include <mutex>
|
||||
#include "clientConnection.h"
|
||||
#include "common/strCompat.h"
|
||||
#include "common/snapException.h"
|
||||
#include "message/hello.h"
|
||||
#include "aixlog.hpp"
|
||||
|
||||
|
||||
using namespace std;
|
||||
|
||||
|
||||
ClientConnection::ClientConnection(MessageReceiver* receiver, const std::string& host, size_t port) : socket_(nullptr), active_(false), connected_(false), messageReceiver_(receiver), reqId_(1), host_(host), port_(port), readerThread_(NULL), sumTimeout_(chronos::msec(0))
|
||||
ClientConnection::ClientConnection(MessageReceiver* receiver, const std::string& host, size_t port)
|
||||
: socket_(nullptr), active_(false), connected_(false), messageReceiver_(receiver), reqId_(1), host_(host), port_(port), readerThread_(NULL),
|
||||
sumTimeout_(chronos::msec(0))
|
||||
{
|
||||
}
|
||||
|
||||
|
||||
ClientConnection::~ClientConnection()
|
||||
{
|
||||
stop();
|
||||
stop();
|
||||
}
|
||||
|
||||
|
||||
|
||||
void ClientConnection::socketRead(void* _to, size_t _bytes)
|
||||
{
|
||||
size_t toRead = _bytes;
|
||||
size_t len = 0;
|
||||
do
|
||||
{
|
||||
len += socket_->read_some(asio::buffer((char*)_to + len, toRead));
|
||||
//cout << "len: " << len << ", error: " << error << endl;
|
||||
toRead = _bytes - len;
|
||||
}
|
||||
while (toRead > 0);
|
||||
size_t toRead = _bytes;
|
||||
size_t len = 0;
|
||||
do
|
||||
{
|
||||
len += socket_->read_some(asio::buffer((char*)_to + len, toRead));
|
||||
// cout << "len: " << len << ", error: " << error << endl;
|
||||
toRead = _bytes - len;
|
||||
} while (toRead > 0);
|
||||
}
|
||||
|
||||
|
||||
std::string ClientConnection::getMacAddress() const
|
||||
{
|
||||
if (socket_ == nullptr)
|
||||
throw SnapException("socket not connected");
|
||||
if (socket_ == nullptr)
|
||||
throw SnapException("socket not connected");
|
||||
|
||||
std::string mac = ::getMacAddress(socket_->native_handle());
|
||||
if (mac.empty())
|
||||
mac = "00:00:00:00:00:00";
|
||||
LOG(INFO) << "My MAC: \"" << mac << "\", socket: " << socket_->native_handle() << "\n";
|
||||
return mac;
|
||||
std::string mac = ::getMacAddress(socket_->native_handle());
|
||||
if (mac.empty())
|
||||
mac = "00:00:00:00:00:00";
|
||||
LOG(INFO) << "My MAC: \"" << mac << "\", socket: " << socket_->native_handle() << "\n";
|
||||
return mac;
|
||||
}
|
||||
|
||||
|
||||
void ClientConnection::start()
|
||||
{
|
||||
tcp::resolver resolver(io_service_);
|
||||
tcp::resolver::query query(host_, cpt::to_string(port_), asio::ip::resolver_query_base::numeric_service);
|
||||
auto iterator = resolver.resolve(query);
|
||||
LOG(DEBUG) << "Connecting\n";
|
||||
socket_.reset(new tcp::socket(io_service_));
|
||||
// struct timeval tv;
|
||||
// tv.tv_sec = 5;
|
||||
// tv.tv_usec = 0;
|
||||
// cout << "socket: " << socket->native_handle() << "\n";
|
||||
// setsockopt(socket->native_handle(), SOL_SOCKET, SO_RCVTIMEO, &tv, sizeof(tv));
|
||||
// setsockopt(socket->native_handle(), SOL_SOCKET, SO_SNDTIMEO, &tv, sizeof(tv));
|
||||
socket_->connect(*iterator);
|
||||
connected_ = true;
|
||||
SLOG(NOTICE) << "Connected to " << socket_->remote_endpoint().address().to_string() << endl;
|
||||
active_ = true;
|
||||
sumTimeout_ = chronos::msec(0);
|
||||
readerThread_ = new thread(&ClientConnection::reader, this);
|
||||
tcp::resolver resolver(io_service_);
|
||||
tcp::resolver::query query(host_, cpt::to_string(port_), asio::ip::resolver_query_base::numeric_service);
|
||||
auto iterator = resolver.resolve(query);
|
||||
LOG(DEBUG) << "Connecting\n";
|
||||
socket_.reset(new tcp::socket(io_service_));
|
||||
// struct timeval tv;
|
||||
// tv.tv_sec = 5;
|
||||
// tv.tv_usec = 0;
|
||||
// cout << "socket: " << socket->native_handle() << "\n";
|
||||
// setsockopt(socket->native_handle(), SOL_SOCKET, SO_RCVTIMEO, &tv, sizeof(tv));
|
||||
// setsockopt(socket->native_handle(), SOL_SOCKET, SO_SNDTIMEO, &tv, sizeof(tv));
|
||||
socket_->connect(*iterator);
|
||||
connected_ = true;
|
||||
SLOG(NOTICE) << "Connected to " << socket_->remote_endpoint().address().to_string() << endl;
|
||||
active_ = true;
|
||||
sumTimeout_ = chronos::msec(0);
|
||||
readerThread_ = new thread(&ClientConnection::reader, this);
|
||||
}
|
||||
|
||||
|
||||
void ClientConnection::stop()
|
||||
{
|
||||
connected_ = false;
|
||||
active_ = false;
|
||||
try
|
||||
{
|
||||
std::error_code ec;
|
||||
if (socket_)
|
||||
{
|
||||
socket_->shutdown(asio::ip::tcp::socket::shutdown_both, ec);
|
||||
if (ec) LOG(ERROR) << "Error in socket shutdown: " << ec.message() << endl;
|
||||
socket_->close(ec);
|
||||
if (ec) LOG(ERROR) << "Error in socket close: " << ec.message() << endl;
|
||||
}
|
||||
if (readerThread_)
|
||||
{
|
||||
LOG(DEBUG) << "joining readerThread\n";
|
||||
readerThread_->join();
|
||||
delete readerThread_;
|
||||
}
|
||||
}
|
||||
catch(...)
|
||||
{
|
||||
}
|
||||
readerThread_ = NULL;
|
||||
socket_.reset();
|
||||
LOG(DEBUG) << "readerThread terminated\n";
|
||||
connected_ = false;
|
||||
active_ = false;
|
||||
try
|
||||
{
|
||||
std::error_code ec;
|
||||
if (socket_)
|
||||
{
|
||||
socket_->shutdown(asio::ip::tcp::socket::shutdown_both, ec);
|
||||
if (ec)
|
||||
LOG(ERROR) << "Error in socket shutdown: " << ec.message() << endl;
|
||||
socket_->close(ec);
|
||||
if (ec)
|
||||
LOG(ERROR) << "Error in socket close: " << ec.message() << endl;
|
||||
}
|
||||
if (readerThread_)
|
||||
{
|
||||
LOG(DEBUG) << "joining readerThread\n";
|
||||
readerThread_->join();
|
||||
delete readerThread_;
|
||||
}
|
||||
}
|
||||
catch (...)
|
||||
{
|
||||
}
|
||||
readerThread_ = NULL;
|
||||
socket_.reset();
|
||||
LOG(DEBUG) << "readerThread terminated\n";
|
||||
}
|
||||
|
||||
|
||||
bool ClientConnection::send(const msg::BaseMessage* message) const
|
||||
{
|
||||
// std::unique_lock<std::mutex> mlock(mutex_);
|
||||
//LOG(DEBUG) << "send: " << message->type << ", size: " << message->getSize() << "\n";
|
||||
std::lock_guard<std::mutex> socketLock(socketMutex_);
|
||||
if (!connected())
|
||||
return false;
|
||||
//LOG(DEBUG) << "send: " << message->type << ", size: " << message->getSize() << "\n";
|
||||
asio::streambuf streambuf;
|
||||
std::ostream stream(&streambuf);
|
||||
tv t;
|
||||
message->sent = t;
|
||||
message->serialize(stream);
|
||||
asio::write(*socket_.get(), streambuf);
|
||||
return true;
|
||||
// std::unique_lock<std::mutex> mlock(mutex_);
|
||||
// LOG(DEBUG) << "send: " << message->type << ", size: " << message->getSize() << "\n";
|
||||
std::lock_guard<std::mutex> socketLock(socketMutex_);
|
||||
if (!connected())
|
||||
return false;
|
||||
// LOG(DEBUG) << "send: " << message->type << ", size: " << message->getSize() << "\n";
|
||||
asio::streambuf streambuf;
|
||||
std::ostream stream(&streambuf);
|
||||
tv t;
|
||||
message->sent = t;
|
||||
message->serialize(stream);
|
||||
asio::write(*socket_.get(), streambuf);
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
shared_ptr<msg::SerializedMessage> ClientConnection::sendRequest(const msg::BaseMessage* message, const chronos::msec& timeout)
|
||||
{
|
||||
shared_ptr<msg::SerializedMessage> response(NULL);
|
||||
if (++reqId_ >= 10000)
|
||||
reqId_ = 1;
|
||||
message->id = reqId_;
|
||||
// LOG(INFO) << "Req: " << message->id << "\n";
|
||||
shared_ptr<PendingRequest> pendingRequest(new PendingRequest(reqId_));
|
||||
shared_ptr<msg::SerializedMessage> response(NULL);
|
||||
if (++reqId_ >= 10000)
|
||||
reqId_ = 1;
|
||||
message->id = reqId_;
|
||||
// LOG(INFO) << "Req: " << message->id << "\n";
|
||||
shared_ptr<PendingRequest> pendingRequest(new PendingRequest(reqId_));
|
||||
|
||||
std::unique_lock<std::mutex> lock(pendingRequestsMutex_);
|
||||
pendingRequests_.insert(pendingRequest);
|
||||
send(message);
|
||||
if (pendingRequest->cv.wait_for(lock, std::chrono::milliseconds(timeout)) == std::cv_status::no_timeout)
|
||||
{
|
||||
response = pendingRequest->response;
|
||||
sumTimeout_ = chronos::msec(0);
|
||||
// LOG(INFO) << "Resp: " << pendingRequest->id << "\n";
|
||||
}
|
||||
else
|
||||
{
|
||||
sumTimeout_ += timeout;
|
||||
LOG(WARNING) << "timeout while waiting for response to: " << reqId_ << ", timeout " << sumTimeout_.count() << "\n";
|
||||
if (sumTimeout_ > chronos::sec(10))
|
||||
throw SnapException("sum timeout exceeded 10s");
|
||||
}
|
||||
pendingRequests_.erase(pendingRequest);
|
||||
return response;
|
||||
std::unique_lock<std::mutex> lock(pendingRequestsMutex_);
|
||||
pendingRequests_.insert(pendingRequest);
|
||||
send(message);
|
||||
if (pendingRequest->cv.wait_for(lock, std::chrono::milliseconds(timeout)) == std::cv_status::no_timeout)
|
||||
{
|
||||
response = pendingRequest->response;
|
||||
sumTimeout_ = chronos::msec(0);
|
||||
// LOG(INFO) << "Resp: " << pendingRequest->id << "\n";
|
||||
}
|
||||
else
|
||||
{
|
||||
sumTimeout_ += timeout;
|
||||
LOG(WARNING) << "timeout while waiting for response to: " << reqId_ << ", timeout " << sumTimeout_.count() << "\n";
|
||||
if (sumTimeout_ > chronos::sec(10))
|
||||
throw SnapException("sum timeout exceeded 10s");
|
||||
}
|
||||
pendingRequests_.erase(pendingRequest);
|
||||
return response;
|
||||
}
|
||||
|
||||
|
||||
void ClientConnection::getNextMessage()
|
||||
{
|
||||
msg::BaseMessage baseMessage;
|
||||
size_t baseMsgSize = baseMessage.getSize();
|
||||
vector<char> buffer(baseMsgSize);
|
||||
socketRead(&buffer[0], baseMsgSize);
|
||||
baseMessage.deserialize(&buffer[0]);
|
||||
// LOG(DEBUG) << "getNextMessage: " << baseMessage.type << ", size: " << baseMessage.size << ", id: " << baseMessage.id << ", refers: " << baseMessage.refersTo << "\n";
|
||||
if (baseMessage.size > buffer.size())
|
||||
buffer.resize(baseMessage.size);
|
||||
// {
|
||||
// std::lock_guard<std::mutex> socketLock(socketMutex_);
|
||||
socketRead(&buffer[0], baseMessage.size);
|
||||
// }
|
||||
tv t;
|
||||
baseMessage.received = t;
|
||||
msg::BaseMessage baseMessage;
|
||||
size_t baseMsgSize = baseMessage.getSize();
|
||||
vector<char> buffer(baseMsgSize);
|
||||
socketRead(&buffer[0], baseMsgSize);
|
||||
baseMessage.deserialize(&buffer[0]);
|
||||
// LOG(DEBUG) << "getNextMessage: " << baseMessage.type << ", size: " << baseMessage.size << ", id: " << baseMessage.id << ", refers: " <<
|
||||
//baseMessage.refersTo << "\n";
|
||||
if (baseMessage.size > buffer.size())
|
||||
buffer.resize(baseMessage.size);
|
||||
// {
|
||||
// std::lock_guard<std::mutex> socketLock(socketMutex_);
|
||||
socketRead(&buffer[0], baseMessage.size);
|
||||
// }
|
||||
tv t;
|
||||
baseMessage.received = t;
|
||||
|
||||
{
|
||||
std::unique_lock<std::mutex> lock(pendingRequestsMutex_);
|
||||
// LOG(DEBUG) << "got lock - getNextMessage: " << baseMessage.type << ", size: " << baseMessage.size << ", id: " << baseMessage.id << ", refers: " << baseMessage.refersTo << "\n";
|
||||
{
|
||||
for (auto req: pendingRequests_)
|
||||
{
|
||||
if (req->id == baseMessage.refersTo)
|
||||
{
|
||||
req->response.reset(new msg::SerializedMessage());
|
||||
req->response->message = baseMessage;
|
||||
req->response->buffer = (char*)malloc(baseMessage.size);
|
||||
memcpy(req->response->buffer, &buffer[0], baseMessage.size);
|
||||
lock.unlock();
|
||||
req->cv.notify_one();
|
||||
return;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
{
|
||||
std::unique_lock<std::mutex> lock(pendingRequestsMutex_);
|
||||
// LOG(DEBUG) << "got lock - getNextMessage: " << baseMessage.type << ", size: " << baseMessage.size << ", id: " << baseMessage.id << ",
|
||||
//refers: " << baseMessage.refersTo << "\n";
|
||||
{
|
||||
for (auto req : pendingRequests_)
|
||||
{
|
||||
if (req->id == baseMessage.refersTo)
|
||||
{
|
||||
req->response.reset(new msg::SerializedMessage());
|
||||
req->response->message = baseMessage;
|
||||
req->response->buffer = (char*)malloc(baseMessage.size);
|
||||
memcpy(req->response->buffer, &buffer[0], baseMessage.size);
|
||||
lock.unlock();
|
||||
req->cv.notify_one();
|
||||
return;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (messageReceiver_ != NULL)
|
||||
messageReceiver_->onMessageReceived(this, baseMessage, &buffer[0]);
|
||||
if (messageReceiver_ != NULL)
|
||||
messageReceiver_->onMessageReceived(this, baseMessage, &buffer[0]);
|
||||
}
|
||||
|
||||
|
||||
|
||||
void ClientConnection::reader()
|
||||
{
|
||||
try
|
||||
{
|
||||
while(active_)
|
||||
{
|
||||
getNextMessage();
|
||||
}
|
||||
}
|
||||
catch (const std::exception& e)
|
||||
{
|
||||
if (messageReceiver_ != NULL)
|
||||
messageReceiver_->onException(this, make_shared<SnapException>(e.what()));
|
||||
}
|
||||
catch (...)
|
||||
{
|
||||
}
|
||||
connected_ = false;
|
||||
active_ = false;
|
||||
try
|
||||
{
|
||||
while (active_)
|
||||
{
|
||||
getNextMessage();
|
||||
}
|
||||
}
|
||||
catch (const std::exception& e)
|
||||
{
|
||||
if (messageReceiver_ != NULL)
|
||||
messageReceiver_->onException(this, make_shared<SnapException>(e.what()));
|
||||
}
|
||||
catch (...)
|
||||
{
|
||||
}
|
||||
connected_ = false;
|
||||
active_ = false;
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue