Fixes some other third person camera specific attachments. Implements a single new flag for entities to be forced visible in first person mode. Old mods do not need to be updated to use the new flag and are fully backwards compatible.
912 lines
23 KiB
C++
912 lines
23 KiB
C++
/*
|
|
Minetest
|
|
Copyright (C) 2010-2014 celeron55, Perttu Ahola <celeron55@gmail.com>
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU Lesser General Public License as published by
|
|
the Free Software Foundation; either version 2.1 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public License along
|
|
with this program; if not, write to the Free Software Foundation, Inc.,
|
|
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
|
|
#include <sstream>
|
|
#include "clientiface.h"
|
|
#include "network/connection.h"
|
|
#include "network/serveropcodes.h"
|
|
#include "remoteplayer.h"
|
|
#include "settings.h"
|
|
#include "mapblock.h"
|
|
#include "serverenvironment.h"
|
|
#include "map.h"
|
|
#include "emerge.h"
|
|
#include "server/luaentity_sao.h"
|
|
#include "server/player_sao.h"
|
|
#include "log.h"
|
|
#include "util/srp.h"
|
|
#include "face_position_cache.h"
|
|
|
|
const char *ClientInterface::statenames[] = {
|
|
"Invalid",
|
|
"Disconnecting",
|
|
"Denied",
|
|
"Created",
|
|
"AwaitingInit2",
|
|
"HelloSent",
|
|
"InitDone",
|
|
"DefinitionsSent",
|
|
"Active",
|
|
"SudoMode",
|
|
};
|
|
|
|
|
|
|
|
std::string ClientInterface::state2Name(ClientState state)
|
|
{
|
|
return statenames[state];
|
|
}
|
|
|
|
RemoteClient::RemoteClient() :
|
|
m_max_simul_sends(g_settings->getU16("max_simultaneous_block_sends_per_client")),
|
|
m_min_time_from_building(
|
|
g_settings->getFloat("full_block_send_enable_min_time_from_building")),
|
|
m_max_send_distance(g_settings->getS16("max_block_send_distance")),
|
|
m_block_optimize_distance(g_settings->getS16("block_send_optimize_distance")),
|
|
m_max_gen_distance(g_settings->getS16("max_block_generate_distance")),
|
|
m_occ_cull(g_settings->getBool("server_side_occlusion_culling"))
|
|
{
|
|
}
|
|
|
|
void RemoteClient::ResendBlockIfOnWire(v3s16 p)
|
|
{
|
|
// if this block is on wire, mark it for sending again as soon as possible
|
|
if (m_blocks_sending.find(p) != m_blocks_sending.end()) {
|
|
SetBlockNotSent(p);
|
|
}
|
|
}
|
|
|
|
LuaEntitySAO *getAttachedObject(PlayerSAO *sao, ServerEnvironment *env)
|
|
{
|
|
if (!sao->isAttached())
|
|
return nullptr;
|
|
|
|
int id;
|
|
std::string bone;
|
|
v3f dummy;
|
|
bool force_visible;
|
|
sao->getAttachment(&id, &bone, &dummy, &dummy, &force_visible);
|
|
ServerActiveObject *ao = env->getActiveObject(id);
|
|
while (id && ao) {
|
|
ao->getAttachment(&id, &bone, &dummy, &dummy, &force_visible);
|
|
if (id)
|
|
ao = env->getActiveObject(id);
|
|
}
|
|
return dynamic_cast<LuaEntitySAO *>(ao);
|
|
}
|
|
|
|
void RemoteClient::GetNextBlocks (
|
|
ServerEnvironment *env,
|
|
EmergeManager * emerge,
|
|
float dtime,
|
|
std::vector<PrioritySortedBlockTransfer> &dest)
|
|
{
|
|
// Increment timers
|
|
m_nothing_to_send_pause_timer -= dtime;
|
|
m_nearest_unsent_reset_timer += dtime;
|
|
|
|
if (m_nothing_to_send_pause_timer >= 0)
|
|
return;
|
|
|
|
RemotePlayer *player = env->getPlayer(peer_id);
|
|
// This can happen sometimes; clients and players are not in perfect sync.
|
|
if (!player)
|
|
return;
|
|
|
|
PlayerSAO *sao = player->getPlayerSAO();
|
|
if (!sao)
|
|
return;
|
|
|
|
// Won't send anything if already sending
|
|
if (m_blocks_sending.size() >= m_max_simul_sends) {
|
|
//infostream<<"Not sending any blocks, Queue full."<<std::endl;
|
|
return;
|
|
}
|
|
|
|
v3f playerpos = sao->getBasePosition();
|
|
// if the player is attached, get the velocity from the attached object
|
|
LuaEntitySAO *lsao = getAttachedObject(sao, env);
|
|
const v3f &playerspeed = lsao? lsao->getVelocity() : player->getSpeed();
|
|
v3f playerspeeddir(0,0,0);
|
|
if (playerspeed.getLength() > 1.0f * BS)
|
|
playerspeeddir = playerspeed / playerspeed.getLength();
|
|
// Predict to next block
|
|
v3f playerpos_predicted = playerpos + playerspeeddir * (MAP_BLOCKSIZE * BS);
|
|
|
|
v3s16 center_nodepos = floatToInt(playerpos_predicted, BS);
|
|
|
|
v3s16 center = getNodeBlockPos(center_nodepos);
|
|
|
|
// Camera position and direction
|
|
v3f camera_pos = sao->getEyePosition();
|
|
v3f camera_dir = v3f(0,0,1);
|
|
camera_dir.rotateYZBy(sao->getLookPitch());
|
|
camera_dir.rotateXZBy(sao->getRotation().Y);
|
|
|
|
/*infostream<<"camera_dir=("<<camera_dir.X<<","<<camera_dir.Y<<","
|
|
<<camera_dir.Z<<")"<<std::endl;*/
|
|
|
|
/*
|
|
Get the starting value of the block finder radius.
|
|
*/
|
|
|
|
if (m_last_center != center) {
|
|
m_nearest_unsent_d = 0;
|
|
m_last_center = center;
|
|
}
|
|
|
|
/*infostream<<"m_nearest_unsent_reset_timer="
|
|
<<m_nearest_unsent_reset_timer<<std::endl;*/
|
|
|
|
// Reset periodically to workaround for some bugs or stuff
|
|
if (m_nearest_unsent_reset_timer > 20.0f) {
|
|
m_nearest_unsent_reset_timer = 0.0f;
|
|
m_nearest_unsent_d = 0;
|
|
//infostream<<"Resetting m_nearest_unsent_d for "
|
|
// <<server->getPlayerName(peer_id)<<std::endl;
|
|
}
|
|
|
|
//s16 last_nearest_unsent_d = m_nearest_unsent_d;
|
|
s16 d_start = m_nearest_unsent_d;
|
|
|
|
//infostream<<"d_start="<<d_start<<std::endl;
|
|
|
|
u16 max_simul_sends_usually = m_max_simul_sends;
|
|
|
|
/*
|
|
Check the time from last addNode/removeNode.
|
|
|
|
Decrease send rate if player is building stuff.
|
|
*/
|
|
m_time_from_building += dtime;
|
|
if (m_time_from_building < m_min_time_from_building) {
|
|
max_simul_sends_usually
|
|
= LIMITED_MAX_SIMULTANEOUS_BLOCK_SENDS;
|
|
}
|
|
|
|
/*
|
|
Number of blocks sending + number of blocks selected for sending
|
|
*/
|
|
u32 num_blocks_selected = m_blocks_sending.size();
|
|
|
|
/*
|
|
next time d will be continued from the d from which the nearest
|
|
unsent block was found this time.
|
|
|
|
This is because not necessarily any of the blocks found this
|
|
time are actually sent.
|
|
*/
|
|
s32 new_nearest_unsent_d = -1;
|
|
|
|
// Get view range and camera fov (radians) from the client
|
|
s16 wanted_range = sao->getWantedRange() + 1;
|
|
float camera_fov = sao->getFov();
|
|
|
|
// Distrust client-sent FOV and get server-set player object property
|
|
// zoom FOV (degrees) as a check to avoid hacked clients using FOV to load
|
|
// distant world.
|
|
// (zoom is disabled by value 0)
|
|
float prop_zoom_fov = sao->getZoomFOV() < 0.001f ?
|
|
0.0f :
|
|
std::max(camera_fov, sao->getZoomFOV() * core::DEGTORAD);
|
|
|
|
const s16 full_d_max = std::min(adjustDist(m_max_send_distance, prop_zoom_fov),
|
|
wanted_range);
|
|
const s16 d_opt = std::min(adjustDist(m_block_optimize_distance, prop_zoom_fov),
|
|
wanted_range);
|
|
const s16 d_blocks_in_sight = full_d_max * BS * MAP_BLOCKSIZE;
|
|
|
|
s16 d_max = full_d_max;
|
|
s16 d_max_gen = std::min(adjustDist(m_max_gen_distance, prop_zoom_fov),
|
|
wanted_range);
|
|
|
|
// Don't loop very much at a time, adjust with distance,
|
|
// do more work per RTT with greater distances.
|
|
s16 max_d_increment_at_time = full_d_max / 9 + 1;
|
|
if (d_max > d_start + max_d_increment_at_time)
|
|
d_max = d_start + max_d_increment_at_time;
|
|
|
|
// cos(angle between velocity and camera) * |velocity|
|
|
// Limit to 0.0f in case player moves backwards.
|
|
f32 dot = rangelim(camera_dir.dotProduct(playerspeed), 0.0f, 300.0f);
|
|
|
|
// Reduce the field of view when a player moves and looks forward.
|
|
// limit max fov effect to 50%, 60% at 20n/s fly speed
|
|
camera_fov = camera_fov / (1 + dot / 300.0f);
|
|
|
|
s32 nearest_emerged_d = -1;
|
|
s32 nearest_emergefull_d = -1;
|
|
s32 nearest_sent_d = -1;
|
|
//bool queue_is_full = false;
|
|
|
|
const v3s16 cam_pos_nodes = floatToInt(camera_pos, BS);
|
|
|
|
s16 d;
|
|
for (d = d_start; d <= d_max; d++) {
|
|
/*
|
|
Get the border/face dot coordinates of a "d-radiused"
|
|
box
|
|
*/
|
|
std::vector<v3s16> list = FacePositionCache::getFacePositions(d);
|
|
|
|
std::vector<v3s16>::iterator li;
|
|
for (li = list.begin(); li != list.end(); ++li) {
|
|
v3s16 p = *li + center;
|
|
|
|
/*
|
|
Send throttling
|
|
- Don't allow too many simultaneous transfers
|
|
- EXCEPT when the blocks are very close
|
|
|
|
Also, don't send blocks that are already flying.
|
|
*/
|
|
|
|
// Start with the usual maximum
|
|
u16 max_simul_dynamic = max_simul_sends_usually;
|
|
|
|
// If block is very close, allow full maximum
|
|
if (d <= BLOCK_SEND_DISABLE_LIMITS_MAX_D)
|
|
max_simul_dynamic = m_max_simul_sends;
|
|
|
|
// Don't select too many blocks for sending
|
|
if (num_blocks_selected >= max_simul_dynamic) {
|
|
//queue_is_full = true;
|
|
goto queue_full_break;
|
|
}
|
|
|
|
// Don't send blocks that are currently being transferred
|
|
if (m_blocks_sending.find(p) != m_blocks_sending.end())
|
|
continue;
|
|
|
|
/*
|
|
Do not go over max mapgen limit
|
|
*/
|
|
if (blockpos_over_max_limit(p))
|
|
continue;
|
|
|
|
// If this is true, inexistent block will be made from scratch
|
|
bool generate = d <= d_max_gen;
|
|
|
|
/*
|
|
Don't generate or send if not in sight
|
|
FIXME This only works if the client uses a small enough
|
|
FOV setting. The default of 72 degrees is fine.
|
|
Also retrieve a smaller view cone in the direction of the player's
|
|
movement.
|
|
(0.1 is about 4 degrees)
|
|
*/
|
|
f32 dist;
|
|
if (!(isBlockInSight(p, camera_pos, camera_dir, camera_fov,
|
|
d_blocks_in_sight, &dist) ||
|
|
(playerspeed.getLength() > 1.0f * BS &&
|
|
isBlockInSight(p, camera_pos, playerspeeddir, 0.1f,
|
|
d_blocks_in_sight)))) {
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
Don't send already sent blocks
|
|
*/
|
|
if (m_blocks_sent.find(p) != m_blocks_sent.end())
|
|
continue;
|
|
|
|
/*
|
|
Check if map has this block
|
|
*/
|
|
MapBlock *block = env->getMap().getBlockNoCreateNoEx(p);
|
|
|
|
bool surely_not_found_on_disk = false;
|
|
bool block_is_invalid = false;
|
|
if (block) {
|
|
// Reset usage timer, this block will be of use in the future.
|
|
block->resetUsageTimer();
|
|
|
|
// Block is dummy if data doesn't exist.
|
|
// It means it has been not found from disk and not generated
|
|
if (block->isDummy()) {
|
|
surely_not_found_on_disk = true;
|
|
}
|
|
|
|
if (!block->isGenerated())
|
|
block_is_invalid = true;
|
|
|
|
/*
|
|
If block is not close, don't send it unless it is near
|
|
ground level.
|
|
|
|
Block is near ground level if night-time mesh
|
|
differs from day-time mesh.
|
|
*/
|
|
if (d >= d_opt) {
|
|
if (!block->getIsUnderground() && !block->getDayNightDiff())
|
|
continue;
|
|
}
|
|
|
|
if (m_occ_cull && !block_is_invalid &&
|
|
env->getMap().isBlockOccluded(block, cam_pos_nodes)) {
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/*
|
|
If block has been marked to not exist on disk (dummy) or is
|
|
not generated and generating new ones is not wanted, skip block.
|
|
*/
|
|
if (!generate && (surely_not_found_on_disk || block_is_invalid)) {
|
|
// get next one.
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
Add inexistent block to emerge queue.
|
|
*/
|
|
if (block == NULL || surely_not_found_on_disk || block_is_invalid) {
|
|
if (emerge->enqueueBlockEmerge(peer_id, p, generate)) {
|
|
if (nearest_emerged_d == -1)
|
|
nearest_emerged_d = d;
|
|
} else {
|
|
if (nearest_emergefull_d == -1)
|
|
nearest_emergefull_d = d;
|
|
goto queue_full_break;
|
|
}
|
|
|
|
// get next one.
|
|
continue;
|
|
}
|
|
|
|
if (nearest_sent_d == -1)
|
|
nearest_sent_d = d;
|
|
|
|
/*
|
|
Add block to send queue
|
|
*/
|
|
PrioritySortedBlockTransfer q((float)dist, p, peer_id);
|
|
|
|
dest.push_back(q);
|
|
|
|
num_blocks_selected += 1;
|
|
}
|
|
}
|
|
queue_full_break:
|
|
|
|
// If nothing was found for sending and nothing was queued for
|
|
// emerging, continue next time browsing from here
|
|
if (nearest_emerged_d != -1) {
|
|
new_nearest_unsent_d = nearest_emerged_d;
|
|
} else if (nearest_emergefull_d != -1) {
|
|
new_nearest_unsent_d = nearest_emergefull_d;
|
|
} else {
|
|
if (d > full_d_max) {
|
|
new_nearest_unsent_d = 0;
|
|
m_nothing_to_send_pause_timer = 2.0f;
|
|
} else {
|
|
if (nearest_sent_d != -1)
|
|
new_nearest_unsent_d = nearest_sent_d;
|
|
else
|
|
new_nearest_unsent_d = d;
|
|
}
|
|
}
|
|
|
|
if (new_nearest_unsent_d != -1)
|
|
m_nearest_unsent_d = new_nearest_unsent_d;
|
|
}
|
|
|
|
void RemoteClient::GotBlock(v3s16 p)
|
|
{
|
|
if (m_blocks_modified.find(p) == m_blocks_modified.end()) {
|
|
if (m_blocks_sending.find(p) != m_blocks_sending.end())
|
|
m_blocks_sending.erase(p);
|
|
else
|
|
m_excess_gotblocks++;
|
|
|
|
m_blocks_sent.insert(p);
|
|
}
|
|
}
|
|
|
|
void RemoteClient::SentBlock(v3s16 p)
|
|
{
|
|
if (m_blocks_modified.find(p) != m_blocks_modified.end())
|
|
m_blocks_modified.erase(p);
|
|
|
|
if (m_blocks_sending.find(p) == m_blocks_sending.end())
|
|
m_blocks_sending[p] = 0.0f;
|
|
else
|
|
infostream<<"RemoteClient::SentBlock(): Sent block"
|
|
" already in m_blocks_sending"<<std::endl;
|
|
}
|
|
|
|
void RemoteClient::SetBlockNotSent(v3s16 p)
|
|
{
|
|
m_nearest_unsent_d = 0;
|
|
m_nothing_to_send_pause_timer = 0;
|
|
|
|
if (m_blocks_sending.find(p) != m_blocks_sending.end())
|
|
m_blocks_sending.erase(p);
|
|
if (m_blocks_sent.find(p) != m_blocks_sent.end())
|
|
m_blocks_sent.erase(p);
|
|
m_blocks_modified.insert(p);
|
|
}
|
|
|
|
void RemoteClient::SetBlocksNotSent(std::map<v3s16, MapBlock*> &blocks)
|
|
{
|
|
m_nearest_unsent_d = 0;
|
|
m_nothing_to_send_pause_timer = 0;
|
|
|
|
for (auto &block : blocks) {
|
|
v3s16 p = block.first;
|
|
m_blocks_modified.insert(p);
|
|
|
|
if (m_blocks_sending.find(p) != m_blocks_sending.end())
|
|
m_blocks_sending.erase(p);
|
|
if (m_blocks_sent.find(p) != m_blocks_sent.end())
|
|
m_blocks_sent.erase(p);
|
|
}
|
|
}
|
|
|
|
void RemoteClient::notifyEvent(ClientStateEvent event)
|
|
{
|
|
std::ostringstream myerror;
|
|
switch (m_state)
|
|
{
|
|
case CS_Invalid:
|
|
//intentionally do nothing
|
|
break;
|
|
case CS_Created:
|
|
switch (event) {
|
|
case CSE_Hello:
|
|
m_state = CS_HelloSent;
|
|
break;
|
|
case CSE_InitLegacy:
|
|
m_state = CS_AwaitingInit2;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
break;
|
|
/* GotInit2 SetDefinitionsSent SetMediaSent */
|
|
default:
|
|
myerror << "Created: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
}
|
|
break;
|
|
case CS_Denied:
|
|
/* don't do anything if in denied state */
|
|
break;
|
|
case CS_HelloSent:
|
|
switch(event)
|
|
{
|
|
case CSE_AuthAccept:
|
|
m_state = CS_AwaitingInit2;
|
|
if (chosen_mech == AUTH_MECHANISM_SRP ||
|
|
chosen_mech == AUTH_MECHANISM_LEGACY_PASSWORD)
|
|
srp_verifier_delete((SRPVerifier *) auth_data);
|
|
chosen_mech = AUTH_MECHANISM_NONE;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
if (chosen_mech == AUTH_MECHANISM_SRP ||
|
|
chosen_mech == AUTH_MECHANISM_LEGACY_PASSWORD)
|
|
srp_verifier_delete((SRPVerifier *) auth_data);
|
|
chosen_mech = AUTH_MECHANISM_NONE;
|
|
break;
|
|
default:
|
|
myerror << "HelloSent: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
}
|
|
break;
|
|
case CS_AwaitingInit2:
|
|
switch(event)
|
|
{
|
|
case CSE_GotInit2:
|
|
confirmSerializationVersion();
|
|
m_state = CS_InitDone;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
break;
|
|
|
|
/* Init SetDefinitionsSent SetMediaSent */
|
|
default:
|
|
myerror << "InitSent: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
}
|
|
break;
|
|
|
|
case CS_InitDone:
|
|
switch(event)
|
|
{
|
|
case CSE_SetDefinitionsSent:
|
|
m_state = CS_DefinitionsSent;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
break;
|
|
|
|
/* Init GotInit2 SetMediaSent */
|
|
default:
|
|
myerror << "InitDone: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
}
|
|
break;
|
|
case CS_DefinitionsSent:
|
|
switch(event)
|
|
{
|
|
case CSE_SetClientReady:
|
|
m_state = CS_Active;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
break;
|
|
/* Init GotInit2 SetDefinitionsSent */
|
|
default:
|
|
myerror << "DefinitionsSent: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
}
|
|
break;
|
|
case CS_Active:
|
|
switch(event)
|
|
{
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SudoSuccess:
|
|
m_state = CS_SudoMode;
|
|
if (chosen_mech == AUTH_MECHANISM_SRP)
|
|
srp_verifier_delete((SRPVerifier *) auth_data);
|
|
chosen_mech = AUTH_MECHANISM_NONE;
|
|
break;
|
|
/* Init GotInit2 SetDefinitionsSent SetMediaSent SetDenied */
|
|
default:
|
|
myerror << "Active: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
break;
|
|
}
|
|
break;
|
|
case CS_SudoMode:
|
|
switch(event)
|
|
{
|
|
case CSE_SetDenied:
|
|
m_state = CS_Denied;
|
|
break;
|
|
case CSE_Disconnect:
|
|
m_state = CS_Disconnecting;
|
|
break;
|
|
case CSE_SudoLeave:
|
|
m_state = CS_Active;
|
|
break;
|
|
default:
|
|
myerror << "Active: Invalid client state transition! " << event;
|
|
throw ClientStateError(myerror.str());
|
|
break;
|
|
}
|
|
break;
|
|
case CS_Disconnecting:
|
|
/* we are already disconnecting */
|
|
break;
|
|
}
|
|
}
|
|
|
|
u64 RemoteClient::uptime() const
|
|
{
|
|
return porting::getTimeS() - m_connection_time;
|
|
}
|
|
|
|
ClientInterface::ClientInterface(const std::shared_ptr<con::Connection> & con)
|
|
:
|
|
m_con(con),
|
|
m_env(NULL),
|
|
m_print_info_timer(0.0f)
|
|
{
|
|
|
|
}
|
|
ClientInterface::~ClientInterface()
|
|
{
|
|
/*
|
|
Delete clients
|
|
*/
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
|
|
for (auto &client_it : m_clients) {
|
|
// Delete client
|
|
delete client_it.second;
|
|
}
|
|
}
|
|
}
|
|
|
|
std::vector<session_t> ClientInterface::getClientIDs(ClientState min_state)
|
|
{
|
|
std::vector<session_t> reply;
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
|
|
for (const auto &m_client : m_clients) {
|
|
if (m_client.second->getState() >= min_state)
|
|
reply.push_back(m_client.second->peer_id);
|
|
}
|
|
|
|
return reply;
|
|
}
|
|
|
|
void ClientInterface::markBlockposAsNotSent(const v3s16 &pos)
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
for (const auto &client : m_clients) {
|
|
if (client.second->getState() >= CS_Active)
|
|
client.second->SetBlockNotSent(pos);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Verify if user limit was reached.
|
|
* User limit count all clients from HelloSent state (MT protocol user) to Active state
|
|
* @return true if user limit was reached
|
|
*/
|
|
bool ClientInterface::isUserLimitReached()
|
|
{
|
|
return getClientIDs(CS_HelloSent).size() >= g_settings->getU16("max_users");
|
|
}
|
|
|
|
void ClientInterface::step(float dtime)
|
|
{
|
|
m_print_info_timer += dtime;
|
|
if (m_print_info_timer >= 30.0f) {
|
|
m_print_info_timer = 0.0f;
|
|
UpdatePlayerList();
|
|
}
|
|
}
|
|
|
|
void ClientInterface::UpdatePlayerList()
|
|
{
|
|
if (m_env) {
|
|
std::vector<session_t> clients = getClientIDs();
|
|
m_clients_names.clear();
|
|
|
|
|
|
if (!clients.empty())
|
|
infostream<<"Players:"<<std::endl;
|
|
|
|
for (session_t i : clients) {
|
|
RemotePlayer *player = m_env->getPlayer(i);
|
|
|
|
if (player == NULL)
|
|
continue;
|
|
|
|
infostream << "* " << player->getName() << "\t";
|
|
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
RemoteClient* client = lockedGetClientNoEx(i);
|
|
if (client)
|
|
client->PrintInfo(infostream);
|
|
}
|
|
|
|
m_clients_names.emplace_back(player->getName());
|
|
}
|
|
}
|
|
}
|
|
|
|
void ClientInterface::send(session_t peer_id, u8 channelnum,
|
|
NetworkPacket *pkt, bool reliable)
|
|
{
|
|
m_con->Send(peer_id, channelnum, pkt, reliable);
|
|
}
|
|
|
|
void ClientInterface::sendToAll(NetworkPacket *pkt)
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
for (auto &client_it : m_clients) {
|
|
RemoteClient *client = client_it.second;
|
|
|
|
if (client->net_proto_version != 0) {
|
|
m_con->Send(client->peer_id,
|
|
clientCommandFactoryTable[pkt->getCommand()].channel, pkt,
|
|
clientCommandFactoryTable[pkt->getCommand()].reliable);
|
|
}
|
|
}
|
|
}
|
|
|
|
void ClientInterface::sendToAllCompat(NetworkPacket *pkt, NetworkPacket *legacypkt,
|
|
u16 min_proto_ver)
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
for (auto &client_it : m_clients) {
|
|
RemoteClient *client = client_it.second;
|
|
NetworkPacket *pkt_to_send = nullptr;
|
|
|
|
if (client->net_proto_version >= min_proto_ver) {
|
|
pkt_to_send = pkt;
|
|
} else if (client->net_proto_version != 0) {
|
|
pkt_to_send = legacypkt;
|
|
} else {
|
|
warningstream << "Client with unhandled version to handle: '"
|
|
<< client->net_proto_version << "'";
|
|
continue;
|
|
}
|
|
|
|
m_con->Send(client->peer_id,
|
|
clientCommandFactoryTable[pkt_to_send->getCommand()].channel,
|
|
pkt_to_send,
|
|
clientCommandFactoryTable[pkt_to_send->getCommand()].reliable);
|
|
}
|
|
}
|
|
|
|
RemoteClient* ClientInterface::getClientNoEx(session_t peer_id, ClientState state_min)
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
RemoteClientMap::const_iterator n = m_clients.find(peer_id);
|
|
// The client may not exist; clients are immediately removed if their
|
|
// access is denied, and this event occurs later then.
|
|
if (n == m_clients.end())
|
|
return NULL;
|
|
|
|
if (n->second->getState() >= state_min)
|
|
return n->second;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
RemoteClient* ClientInterface::lockedGetClientNoEx(session_t peer_id, ClientState state_min)
|
|
{
|
|
RemoteClientMap::const_iterator n = m_clients.find(peer_id);
|
|
// The client may not exist; clients are immediately removed if their
|
|
// access is denied, and this event occurs later then.
|
|
if (n == m_clients.end())
|
|
return NULL;
|
|
|
|
if (n->second->getState() >= state_min)
|
|
return n->second;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
ClientState ClientInterface::getClientState(session_t peer_id)
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
RemoteClientMap::const_iterator n = m_clients.find(peer_id);
|
|
// The client may not exist; clients are immediately removed if their
|
|
// access is denied, and this event occurs later then.
|
|
if (n == m_clients.end())
|
|
return CS_Invalid;
|
|
|
|
return n->second->getState();
|
|
}
|
|
|
|
void ClientInterface::setPlayerName(session_t peer_id, const std::string &name)
|
|
{
|
|
RecursiveMutexAutoLock clientslock(m_clients_mutex);
|
|
RemoteClientMap::iterator n = m_clients.find(peer_id);
|
|
// The client may not exist; clients are immediately removed if their
|
|
// access is denied, and this event occurs later then.
|
|
if (n != m_clients.end())
|
|
n->second->setName(name);
|
|
}
|
|
|
|
void ClientInterface::DeleteClient(session_t peer_id)
|
|
{
|
|
RecursiveMutexAutoLock conlock(m_clients_mutex);
|
|
|
|
// Error check
|
|
RemoteClientMap::iterator n = m_clients.find(peer_id);
|
|
// The client may not exist; clients are immediately removed if their
|
|
// access is denied, and this event occurs later then.
|
|
if (n == m_clients.end())
|
|
return;
|
|
|
|
/*
|
|
Mark objects to be not known by the client
|
|
*/
|
|
//TODO this should be done by client destructor!!!
|
|
RemoteClient *client = n->second;
|
|
// Handle objects
|
|
for (u16 id : client->m_known_objects) {
|
|
// Get object
|
|
ServerActiveObject* obj = m_env->getActiveObject(id);
|
|
|
|
if(obj && obj->m_known_by_count > 0)
|
|
obj->m_known_by_count--;
|
|
}
|
|
|
|
// Delete client
|
|
delete m_clients[peer_id];
|
|
m_clients.erase(peer_id);
|
|
}
|
|
|
|
void ClientInterface::CreateClient(session_t peer_id)
|
|
{
|
|
RecursiveMutexAutoLock conlock(m_clients_mutex);
|
|
|
|
// Error check
|
|
RemoteClientMap::iterator n = m_clients.find(peer_id);
|
|
// The client shouldn't already exist
|
|
if (n != m_clients.end()) return;
|
|
|
|
// Create client
|
|
RemoteClient *client = new RemoteClient();
|
|
client->peer_id = peer_id;
|
|
m_clients[client->peer_id] = client;
|
|
}
|
|
|
|
void ClientInterface::event(session_t peer_id, ClientStateEvent event)
|
|
{
|
|
{
|
|
RecursiveMutexAutoLock clientlock(m_clients_mutex);
|
|
|
|
// Error check
|
|
RemoteClientMap::iterator n = m_clients.find(peer_id);
|
|
|
|
// No client to deliver event
|
|
if (n == m_clients.end())
|
|
return;
|
|
n->second->notifyEvent(event);
|
|
}
|
|
|
|
if ((event == CSE_SetClientReady) ||
|
|
(event == CSE_Disconnect) ||
|
|
(event == CSE_SetDenied))
|
|
{
|
|
UpdatePlayerList();
|
|
}
|
|
}
|
|
|
|
u16 ClientInterface::getProtocolVersion(session_t peer_id)
|
|
{
|
|
RecursiveMutexAutoLock conlock(m_clients_mutex);
|
|
|
|
// Error check
|
|
RemoteClientMap::iterator n = m_clients.find(peer_id);
|
|
|
|
// No client to get version
|
|
if (n == m_clients.end())
|
|
return 0;
|
|
|
|
return n->second->net_proto_version;
|
|
}
|
|
|
|
void ClientInterface::setClientVersion(session_t peer_id, u8 major, u8 minor, u8 patch,
|
|
const std::string &full)
|
|
{
|
|
RecursiveMutexAutoLock conlock(m_clients_mutex);
|
|
|
|
// Error check
|
|
RemoteClientMap::iterator n = m_clients.find(peer_id);
|
|
|
|
// No client to set versions
|
|
if (n == m_clients.end())
|
|
return;
|
|
|
|
n->second->setVersionInfo(major, minor, patch, full);
|
|
}
|