Linux premium155.web-hosting.com 4.18.0-513.11.1.lve.el8.x86_64 #1 SMP Thu Jan 18 16:21:02 UTC 2024 x86_64
LiteSpeed
: 162.0.235.200 | : 3.149.237.231
Cant Read [ /etc/named.conf ]
7.4.33
varifktc
www.github.com/MadExploits
Terminal
AUTO ROOT
Adminer
Backdoor Destroyer
Linux Exploit
Lock Shell
Lock File
Create User
CREATE RDP
PHP Mailer
BACKCONNECT
UNLOCK SHELL
HASH IDENTIFIER
CPANEL RESET
CREATE WP USER
README
+ Create Folder
+ Create File
/
opt /
alt /
php83 /
usr /
include /
php /
ext /
swoole /
include /
[ HOME SHELL ]
Name
Size
Permission
Action
swoole.h
22.89
KB
-rw-r--r--
swoole_api.h
3.39
KB
-rw-r--r--
swoole_asm_context.h
1.76
KB
-rw-r--r--
swoole_async.h
2.93
KB
-rw-r--r--
swoole_atomic.h
2.04
KB
-rw-r--r--
swoole_base64.h
1.29
KB
-rw-r--r--
swoole_buffer.h
2.28
KB
-rw-r--r--
swoole_c_api.h
2.16
KB
-rw-r--r--
swoole_channel.h
2.34
KB
-rw-r--r--
swoole_client.h
8.41
KB
-rw-r--r--
swoole_config.h
8.8
KB
-rw-r--r--
swoole_coroutine.h
8.91
KB
-rw-r--r--
swoole_coroutine_c_api.h
4.6
KB
-rw-r--r--
swoole_coroutine_channel.h
4.18
KB
-rw-r--r--
swoole_coroutine_context.h
2.59
KB
-rw-r--r--
swoole_coroutine_socket.h
18.79
KB
-rw-r--r--
swoole_coroutine_system.h
3.48
KB
-rw-r--r--
swoole_dtls.h
2.41
KB
-rw-r--r--
swoole_error.h
5.81
KB
-rw-r--r--
swoole_file.h
4.09
KB
-rw-r--r--
swoole_file_hook.h
2.58
KB
-rw-r--r--
swoole_hash.h
9.69
KB
-rw-r--r--
swoole_heap.h
1.79
KB
-rw-r--r--
swoole_http.h
6.94
KB
-rw-r--r--
swoole_http2.h
8.38
KB
-rw-r--r--
swoole_lock.h
2.64
KB
-rw-r--r--
swoole_log.h
15.72
KB
-rw-r--r--
swoole_lru_cache.h
3.17
KB
-rw-r--r--
swoole_memory.h
2.75
KB
-rw-r--r--
swoole_message_bus.h
5.53
KB
-rw-r--r--
swoole_mime_type.h
1.44
KB
-rw-r--r--
swoole_mqtt.h
2.18
KB
-rw-r--r--
swoole_msg_queue.h
1.92
KB
-rw-r--r--
swoole_pipe.h
2.64
KB
-rw-r--r--
swoole_process_pool.h
7.82
KB
-rw-r--r--
swoole_protocol.h
5.32
KB
-rw-r--r--
swoole_proxy.h
2.14
KB
-rw-r--r--
swoole_reactor.h
11.22
KB
-rw-r--r--
swoole_redis.h
2.09
KB
-rw-r--r--
swoole_server.h
39.1
KB
-rw-r--r--
swoole_signal.h
1.68
KB
-rw-r--r--
swoole_socket.h
17.82
KB
-rw-r--r--
swoole_socket_hook.h
2.23
KB
-rw-r--r--
swoole_ssl.h
4.94
KB
-rw-r--r--
swoole_static_handler.h
3.9
KB
-rw-r--r--
swoole_string.h
8.06
KB
-rw-r--r--
swoole_table.h
7.42
KB
-rw-r--r--
swoole_timer.h
4.05
KB
-rw-r--r--
swoole_util.h
5.58
KB
-rw-r--r--
swoole_version.h
2.25
KB
-rw-r--r--
swoole_websocket.h
4.43
KB
-rw-r--r--
swoole_wheel_timer.h
2.48
KB
-rw-r--r--
Delete
Unzip
Zip
${this.title}
Close
Code Editor : swoole_server.h
/* +----------------------------------------------------------------------+ | Swoole | +----------------------------------------------------------------------+ | This source file is subject to version 2.0 of the Apache license, | | that is bundled with this package in the file LICENSE, and is | | available through the world-wide-web at the following url: | | http://www.apache.org/licenses/LICENSE-2.0.html | | If you did not receive a copy of the Apache2.0 license and are unable| | to obtain it through the world-wide-web, please send a note to | | license@php.net so we can mail you a copy immediately. | +----------------------------------------------------------------------+ | Author: Tianfeng Han <rango@swoole.com> | +----------------------------------------------------------------------+ */ #pragma once #include "swoole_api.h" #include "swoole_string.h" #include "swoole_socket.h" #include "swoole_timer.h" #include "swoole_reactor.h" #include "swoole_signal.h" #include "swoole_protocol.h" #include "swoole_process_pool.h" #include "swoole_pipe.h" #include "swoole_channel.h" #include "swoole_message_bus.h" #ifdef SW_USE_OPENSSL #include "swoole_dtls.h" #endif #ifdef __MACH__ #include <sys/syslimits.h> #endif #include <string> #include <queue> #include <thread> #include <mutex> #include <atomic> #include <unordered_map> #include <unordered_set> //------------------------------------Server------------------------------------------- namespace swoole { namespace http_server { struct Request; } class Server; struct Manager; struct Session { SessionId id; int fd; uint32_t reactor_id : 8; uint32_t reserve_ : 24; }; struct Connection { /** * It must be in the header. When set to 0, it means that connection does not exist. * One-write and multiple-read operation is thread-safe * system fd must be 0. en: signalfd, listen socket */ uint8_t active; SocketType socket_type; int fd; int worker_id; SessionId session_id; //-------------------------------------------------------------- #ifdef SW_USE_OPENSSL uint8_t ssl; uint8_t ssl_ready; #endif uint8_t overflow; uint8_t high_watermark; uint8_t http_upgrade; uint8_t http2_stream; #ifdef SW_HAVE_ZLIB uint8_t websocket_compression; #endif // If it is equal to 1, it means server actively closed the connection uint8_t close_actively; uint8_t closed; uint8_t close_queued; uint8_t closing; uint8_t close_reset; uint8_t peer_closed; // protected connection, do not close connection when receiving/sending timeout uint8_t protect; uint8_t close_notify; uint8_t close_force; ReactorId reactor_id; uint16_t close_errno; int server_fd; sw_atomic_t recv_queued_bytes; uint32_t send_queued_bytes; uint16_t waiting_time; TimerNode *timer; /** * socket address */ network::Address info; /** * link any thing, for kernel, do not use with application. */ void *object; /** * socket, only operated in the main process */ network::Socket *socket; /** * connect/recv/send/close time */ double connect_time; double last_recv_time; double last_send_time; double last_dispatch_time; /** * bind uid */ uint32_t uid; /** * upgarde websocket */ uint8_t websocket_status; /** * unfinished data frame */ String *websocket_buffer; #ifdef SW_USE_OPENSSL String *ssl_client_cert; pid_t ssl_client_cert_pid; #endif sw_atomic_t lock; }; //------------------------------------ReactorThread------------------------------------------- struct ReactorThread { int id; std::thread thread; network::Socket *notify_pipe = nullptr; uint32_t pipe_num = 0; uint64_t dispatch_count = 0; network::Socket *pipe_sockets = nullptr; network::Socket *pipe_command = nullptr; MessageBus message_bus; int init(Server *serv, Reactor *reactor, uint16_t reactor_id); }; struct ServerPortGS { sw_atomic_t connection_num; sw_atomic_t *connection_nums = nullptr; sw_atomic_long_t abort_count; sw_atomic_long_t accept_count; sw_atomic_long_t close_count; sw_atomic_long_t dispatch_count; sw_atomic_long_t request_count; sw_atomic_long_t response_count; sw_atomic_long_t total_recv_bytes; sw_atomic_long_t total_send_bytes; }; struct ListenPort { /** * tcp socket listen backlog */ uint16_t backlog = SW_BACKLOG; bool listening = false; /** * open tcp_defer_accept option */ int tcp_defer_accept = 0; /** * TCP_FASTOPEN */ int tcp_fastopen = 0; /** * TCP KeepAlive */ int tcp_keepidle = SW_TCP_KEEPIDLE; int tcp_keepinterval = SW_TCP_KEEPINTERVAL; int tcp_keepcount = SW_TCP_KEEPCOUNT; int tcp_user_timeout = 0; double max_idle_time = 0; int socket_buffer_size = network::Socket::default_buffer_size; uint32_t buffer_high_watermark = 0; uint32_t buffer_low_watermark = 0; SocketType type = SW_SOCK_TCP; uint8_t ssl = 0; std::string host; int port = 0; network::Socket *socket = nullptr; pthread_t thread_id = 0; uint16_t heartbeat_idle_time = 0; /** * check data eof */ bool open_eof_check = false; /** * built-in http protocol */ bool open_http_protocol = false; /** * built-in http2.0 protocol */ bool open_http2_protocol = false; /** * built-in websocket protocol */ bool open_websocket_protocol = false; /** * open websocket close frame */ bool open_websocket_close_frame = false; /** * open websocket ping frame */ bool open_websocket_ping_frame = false; /** * open websocket pong frame */ bool open_websocket_pong_frame = false; /** * one package: length check */ bool open_length_check = false; /** * for mqtt protocol */ bool open_mqtt_protocol = false; /** * redis protocol */ bool open_redis_protocol = false; /** * open tcp nodelay option */ bool open_tcp_nodelay = false; /** * open tcp nopush option(for sendfile) */ bool open_tcp_nopush = true; /** * open tcp keepalive */ bool open_tcp_keepalive = false; /** * Sec-WebSocket-Protocol */ std::string websocket_subprotocol; /** * set socket option */ int kernel_socket_recv_buffer_size = 0; int kernel_socket_send_buffer_size = 0; #ifdef SW_USE_OPENSSL SSLContext *ssl_context = nullptr; std::unordered_map<std::string, std::shared_ptr<SSLContext>> sni_contexts; #ifdef SW_SUPPORT_DTLS std::unordered_map<int, dtls::Session *> *dtls_sessions = nullptr; bool is_dtls() { return ssl_context && (ssl_context->protocols & SW_SSL_DTLS); } #endif #endif ServerPortGS *gs = nullptr; Protocol protocol = {}; void *ptr = nullptr; int (*onRead)(Reactor *reactor, ListenPort *port, Event *event) = nullptr; bool is_dgram() { return network::Socket::is_dgram(type); } bool is_stream() { return network::Socket::is_stream(type); } void set_eof_protocol(const std::string &eof, bool find_from_right = false) { open_eof_check = true; protocol.split_by_eof = !find_from_right; protocol.package_eof_len = std::min(eof.length(), sizeof(protocol.package_eof)); memcpy(protocol.package_eof, eof.c_str(), protocol.package_eof_len); } void set_length_protocol(uint32_t length_offset, char length_type, uint32_t body_offset) { open_length_check = true; protocol.package_length_type = length_type; protocol.package_length_size = swoole_type_size(length_type); protocol.package_length_offset = length_offset; protocol.package_body_offset = body_offset; } void set_package_max_length(uint32_t max_length) { protocol.package_max_length = max_length; } ListenPort(); ~ListenPort() = default; int listen(); void close(); bool import(int sock); const char *get_protocols(); #ifdef SW_USE_OPENSSL bool ssl_create_context(SSLContext *context); bool ssl_create(Connection *conn, network::Socket *sock); bool ssl_add_sni_cert(const std::string &name, SSLContext *context); bool ssl_init(); void ssl_set_key_file(const std::string &file) { ssl_context->key_file = file; } void ssl_set_cert_file(const std::string &file) { ssl_context->cert_file = file; } #endif void clear_protocol(); network::Socket *get_socket() { return socket; } int get_port() { return port; } const char *get_host() { return host.c_str(); } SocketType get_type() { return type; } int get_fd() { return socket ? socket->fd : -1; } size_t get_connection_num(); }; struct ServerGS { pid_t master_pid; pid_t manager_pid; SessionId session_round; sw_atomic_t start; sw_atomic_t shutdown; int max_fd; int min_fd; bool called_onStart; time_t start_time; sw_atomic_t connection_num; sw_atomic_t *connection_nums = nullptr; sw_atomic_t tasking_num; uint32_t max_concurrency; sw_atomic_t concurrency; sw_atomic_long_t abort_count; sw_atomic_long_t accept_count; sw_atomic_long_t close_count; sw_atomic_long_t dispatch_count; sw_atomic_long_t request_count; sw_atomic_long_t response_count; sw_atomic_long_t total_recv_bytes; sw_atomic_long_t total_send_bytes; sw_atomic_long_t pipe_packet_msg_id; sw_atomic_t spinlock; #ifdef HAVE_PTHREAD_BARRIER pthread_barrier_t manager_barrier; pthread_barrierattr_t manager_barrier_attr; #endif ProcessPool task_workers; ProcessPool event_workers; }; class Factory { protected: Server *server_; public: Factory(Server *_server) { server_ = _server; } virtual ~Factory() {} virtual bool start() = 0; virtual bool shutdown() = 0; virtual bool dispatch(SendData *) = 0; virtual bool finish(SendData *) = 0; virtual bool notify(DataHead *) = 0; virtual bool end(SessionId sesion_id, int flags) = 0; }; class BaseFactory : public Factory { public: BaseFactory(Server *server) : Factory(server) {} ~BaseFactory(); bool start() override; bool shutdown() override; bool dispatch(SendData *) override; bool finish(SendData *) override; bool notify(DataHead *) override; bool end(SessionId sesion_id, int flags) override; }; class ProcessFactory : public Factory { private: std::vector<std::shared_ptr<UnixSocket>> pipes; public: ProcessFactory(Server *server); ~ProcessFactory(); bool start() override; bool shutdown() override; bool dispatch(SendData *) override; bool finish(SendData *) override; bool notify(DataHead *) override; bool end(SessionId sesion_id, int flags) override; }; enum ServerEventType { // recv data payload SW_SERVER_EVENT_RECV_DATA, SW_SERVER_EVENT_RECV_DGRAM, // send data SW_SERVER_EVENT_SEND_DATA, SW_SERVER_EVENT_SEND_FILE, // connection event SW_SERVER_EVENT_CLOSE, SW_SERVER_EVENT_CONNECT, SW_SERVER_EVENT_CLOSE_FORCE, // task SW_SERVER_EVENT_TASK, SW_SERVER_EVENT_FINISH, // pipe SW_SERVER_EVENT_PIPE_MESSAGE, // event operate SW_SERVER_EVENT_PAUSE_RECV, SW_SERVER_EVENT_RESUME_RECV, // buffer event SW_SERVER_EVENT_BUFFER_FULL, SW_SERVER_EVENT_BUFFER_EMPTY, // process message SW_SERVER_EVENT_INCOMING, SW_SERVER_EVENT_SHUTDOWN, SW_SERVER_EVENT_COMMAND_REQUEST, SW_SERVER_EVENT_COMMAND_RESPONSE, }; class Server { public: typedef int (*DispatchFunction)(Server *, Connection *, SendData *); struct Command { typedef std::function<void(Server *, const std::string &msg)> Callback; typedef std::function<std::string(Server *, const std::string &msg)> Handler; enum ProcessType { MASTER = 1u << 1, REACTOR_THREAD = 1u << 2, EVENT_WORKER = 1u << 3, TASK_WORKER = 1u << 4, MANAGER = 1u << 5, ALL_PROCESS = MASTER | REACTOR_THREAD | EVENT_WORKER | TASK_WORKER | MANAGER, }; int id; int accepted_process_types; std::string name; }; enum Mode { MODE_BASE = 1, MODE_PROCESS = 2, }; enum TaskIpcMode { TASK_IPC_UNIXSOCK = 1, TASK_IPC_MSGQUEUE = 2, TASK_IPC_PREEMPTIVE = 3, TASK_IPC_STREAM = 4, }; enum ThreadType { THREAD_MASTER = 1, THREAD_REACTOR = 2, THREAD_HEARTBEAT = 3, }; enum DispatchMode { DISPATCH_ROUND = 1, DISPATCH_FDMOD = 2, DISPATCH_IDLE_WORKER = 3, DISPATCH_IPMOD = 4, DISPATCH_UIDMOD = 5, DISPATCH_USERFUNC = 6, DISPATCH_CO_CONN_LB = 8, DISPATCH_CO_REQ_LB = 9, DISPATCH_CONCURRENT_LB = 10, }; enum FactoryDispatchResult { DISPATCH_RESULT_DISCARD_PACKET = -1, DISPATCH_RESULT_CLOSE_CONNECTION = -2, DISPATCH_RESULT_USERFUNC_FALLBACK = -3, }; enum HookType { HOOK_MASTER_START, HOOK_MASTER_TIMER, HOOK_REACTOR_START, HOOK_WORKER_START, HOOK_TASK_WORKER_START, HOOK_MASTER_CONNECT, HOOK_REACTOR_CONNECT, HOOK_WORKER_CONNECT, HOOK_REACTOR_RECEIVE, HOOK_WORKER_RECEIVE, HOOK_REACTOR_CLOSE, HOOK_WORKER_CLOSE, HOOK_MANAGER_START, HOOK_MANAGER_TIMER, HOOK_PROCESS_TIMER, HOOK_END = SW_MAX_HOOK_TYPE - 1, }; enum CloseFlag { CLOSE_RESET = 1u << 1, CLOSE_ACTIVELY = 1u << 2, }; /** * reactor thread/process num */ uint16_t reactor_num = 0; /** * worker process num */ uint32_t worker_num = 0; uint8_t dgram_port_num = 0; /** * package dispatch mode */ uint8_t dispatch_mode = DISPATCH_FDMOD; /** * No idle work process is available. */ bool scheduler_warning = false; int worker_uid = 0; int worker_groupid = 0; /** * worker process max request */ uint32_t max_request = 0; uint32_t max_request_grace = 0; network::Socket *udp_socket_ipv4 = nullptr; network::Socket *udp_socket_ipv6 = nullptr; network::Socket *dgram_socket = nullptr; int null_fd = -1; uint32_t max_wait_time = SW_WORKER_MAX_WAIT_TIME; uint32_t worker_max_concurrency = UINT_MAX; /*----------------------------Reactor schedule--------------------------------*/ sw_atomic_t worker_round_id = 0; /** * worker(worker and task_worker) process chroot / user / group */ std::string chroot_; std::string user_; std::string group_; /** * run as a daemon process */ bool daemonize = false; /** * have dgram socket */ bool have_dgram_sock = false; /** * have stream socket */ bool have_stream_sock = false; /** * open cpu affinity setting */ bool open_cpu_affinity = false; /** * disable notice when use SW_DISPATCH_ROUND and SW_DISPATCH_QUEUE */ bool disable_notify = false; /** * discard the timeout request */ bool discard_timeout_request = false; /** * parse cookie header */ bool http_parse_cookie = true; /** * parse x-www-form-urlencoded data */ bool http_parse_post = true; /** * parse multipart/form-data files to match $_FILES */ bool http_parse_files = false; /** * http content compression */ bool http_compression = false; /** * RFC-7692 */ bool websocket_compression = false; /** * handle static files */ bool enable_static_handler = false; /** * show file list in the current directory */ bool http_autoindex = false; /** * enable onConnect/onClose event when use dispatch_mode=1/3 */ bool enable_unsafe_event = false; /** * waiting for worker onConnect callback function to return */ bool enable_delay_receive = false; /** * reuse port */ bool enable_reuse_port = false; /** * asynchronous reloading */ bool reload_async = true; /** * use event object */ bool event_object = false; /** * use task object */ bool task_object = false; /** * enable coroutine in task worker */ bool task_enable_coroutine = false; /** * yield coroutine when the output buffer is full */ bool send_yield = true; /** * enable coroutine */ bool enable_coroutine = true; /** * disable multi-threads */ bool single_thread = false; /** * server status */ bool running = true; int *cpu_affinity_available = 0; int cpu_affinity_available_num = 0; UnixSocket *pipe_command = nullptr; MessageBus message_bus; double send_timeout = 0; uint16_t heartbeat_check_interval = 0; time_t reload_time = 0; time_t warning_time = 0; long timezone_ = 0; TimerNode *master_timer = nullptr; TimerNode *heartbeat_timer = nullptr; /* buffer output/input setting*/ uint32_t output_buffer_size = UINT_MAX; uint32_t input_buffer_size = SW_INPUT_BUFFER_SIZE; uint32_t max_queued_bytes = 0; /** * the master process and worker process communicate using unix socket dgram. * ipc_max_size represents the maximum size of each datagram, * which is obtained from the kernel send buffer of unix socket in swServer_set_ipc_max_size function. */ uint32_t ipc_max_size = SW_IPC_MAX_SIZE; void *private_data_1 = nullptr; void *private_data_2 = nullptr; void *private_data_3 = nullptr; Factory *factory = nullptr; Manager *manager = nullptr; std::vector<ListenPort *> ports; ListenPort *get_primary_port() { return ports.front(); } ListenPort *get_port(int _port) { for (auto port : ports) { if (port->port == _port || _port == 0) { return port; } } return nullptr; } ListenPort *get_port_by_server_fd(int server_fd) { return (ListenPort *) connection_list[server_fd].object; } ListenPort *get_port_by_fd(int fd) { return get_port_by_server_fd(connection_list[fd].server_fd); } ListenPort *get_port_by_session_id(SessionId session_id) { Connection *conn = get_connection_by_session_id(session_id); if (!conn) { return nullptr; } return get_port_by_fd(conn->fd); } network::Socket *get_server_socket(int fd) { return connection_list[fd].socket; } /** * [ReactorThread] */ network::Socket *get_worker_pipe_socket(Worker *worker) { return &get_thread(SwooleTG.id)->pipe_sockets[worker->pipe_master->fd]; } network::Socket *get_command_reply_socket() { return is_base_mode() ? get_worker(0)->pipe_master : pipe_command->get_socket(false); } /** * [Worker|Master] */ network::Socket *get_reactor_pipe_socket(SessionId session_id, int reactor_id) { int pipe_index = session_id % reactor_pipe_num; /** * pipe_worker_id: The pipe in which worker. */ int pipe_worker_id = reactor_id + (pipe_index * reactor_num); Worker *worker = get_worker(pipe_worker_id); return worker->pipe_worker; } /** * task process */ uint32_t task_worker_num = 0; uint8_t task_ipc_mode = TASK_IPC_UNIXSOCK; uint32_t task_max_request = 0; uint32_t task_max_request_grace = 0; std::vector<std::shared_ptr<Pipe>> task_notify_pipes; EventData *task_result = nullptr; /** * user process */ std::vector<Worker *> user_worker_list; std::unordered_map<pid_t, Worker *> user_worker_map; Worker *user_workers = nullptr; std::unordered_map<std::string, Command> commands; std::unordered_map<int, Command::Handler> command_handlers; std::unordered_map<int64_t, Command::Callback> command_callbacks; int command_current_id = 1; int64_t command_current_request_id = 1; Worker *workers = nullptr; ServerGS *gs = nullptr; std::shared_ptr<std::unordered_set<std::string>> locations = nullptr; std::shared_ptr<std::vector<std::string>> http_index_files = nullptr; std::shared_ptr<std::unordered_set<std::string>> http_compression_types = nullptr; #ifdef HAVE_PTHREAD_BARRIER pthread_barrier_t reactor_thread_barrier = {}; #endif /** * temporary directory for HTTP uploaded file. */ std::string upload_tmp_dir = "/tmp"; /** * Write the uploaded file in form-data to disk file */ size_t upload_max_filesize = 0; /** * http compression level for gzip/br */ #ifdef SW_HAVE_COMPRESSION uint8_t http_compression_level = 0; uint32_t compression_min_length; #endif /** * master process pid */ std::string pid_file; EventData *last_task = nullptr; std::queue<String *> *buffer_pool = nullptr; const Allocator *recv_buffer_allocator = &SwooleG.std_allocator; size_t recv_buffer_size = SW_BUFFER_SIZE_BIG; int manager_alarm = 0; /** * message queue key */ uint64_t message_queue_key = 0; void *hooks[SW_MAX_HOOK_TYPE] = {}; /*----------------------------Event Callback--------------------------------*/ /** * Master Process */ std::function<void(Server *)> onStart; std::function<void(Server *)> onBeforeShutdown; std::function<void(Server *)> onShutdown; /** * Manager Process */ std::function<void(Server *)> onManagerStart; std::function<void(Server *)> onManagerStop; std::function<void(Server *, Worker *, const ExitStatus &)> onWorkerError; std::function<void(Server *)> onBeforeReload; std::function<void(Server *)> onAfterReload; /** * Worker Process */ std::function<void(Server *, EventData *)> onPipeMessage; std::function<void(Server *, Worker *)> onWorkerStart; std::function<void(Server *, Worker *)> onWorkerStop; std::function<void(Server *, Worker *)> onWorkerExit; std::function<void(Server *, Worker *)> onUserWorkerStart; /** * Connection */ std::function<int(Server *, RecvData *)> onReceive; std::function<int(Server *, RecvData *)> onPacket; std::function<void(Server *, DataHead *)> onClose; std::function<void(Server *, DataHead *)> onConnect; std::function<void(Server *, DataHead *)> onBufferFull; std::function<void(Server *, DataHead *)> onBufferEmpty; /** * Task Worker */ std::function<int(Server *, EventData *)> onTask; std::function<int(Server *, EventData *)> onFinish; /** * Hook */ int (*dispatch_func)(Server *, Connection *, SendData *) = nullptr; public: Server(enum Mode _mode = MODE_BASE); ~Server(); bool set_document_root(const std::string &path) { if (path.length() > PATH_MAX) { swoole_warning("The length of document_root must be less than %d", PATH_MAX); return false; } char _realpath[PATH_MAX]; if (!realpath(path.c_str(), _realpath)) { swoole_warning("document_root[%s] does not exist", path.c_str()); return false; } document_root = std::string(_realpath); return true; } void add_static_handler_location(const std::string &); void add_static_handler_index_files(const std::string &); bool select_static_handler(http_server::Request *request, Connection *conn); void add_http_compression_type(const std::string &type); int create(); int start(); bool reload(bool reload_all_workers); bool shutdown(); int add_worker(Worker *worker); ListenPort *add_port(SocketType type, const char *host, int port); int add_systemd_socket(); int add_hook(enum HookType type, const Callback &func, int push_back); bool add_command(const std::string &command, int accepted_process_types, const Command::Handler &func); Connection *add_connection(ListenPort *ls, network::Socket *_socket, int server_fd); void abort_connection(Reactor *reactor, ListenPort *ls, network::Socket *_socket); int connection_incoming(Reactor *reactor, Connection *conn); int get_idle_worker_num(); int get_idle_task_worker_num(); int get_task_count(); int get_minfd() { return gs->min_fd; } int get_maxfd() { return gs->max_fd; } void set_maxfd(int maxfd) { gs->max_fd = maxfd; } void set_minfd(int minfd) { gs->min_fd = minfd; } pid_t get_master_pid() { return gs->master_pid; } pid_t get_manager_pid() { return gs->manager_pid; } void store_listen_socket(); void store_pipe_fd(UnixSocket *p); const std::string &get_document_root() { return document_root; } String *get_recv_buffer(network::Socket *_socket) { String *buffer = _socket->recv_buffer; if (buffer == nullptr) { buffer = swoole::make_string(SW_BUFFER_SIZE_BIG, recv_buffer_allocator); if (!buffer) { return nullptr; } _socket->recv_buffer = buffer; } return buffer; } uint32_t get_worker_buffer_num() { return is_base_mode() ? 1 : reactor_num + dgram_port_num; } bool is_support_unsafe_events() { if (is_hash_dispatch_mode()) { return true; } else { return enable_unsafe_event; } } bool is_process_mode() { return mode_ == MODE_PROCESS; } bool is_base_mode() { return mode_ == MODE_BASE; } bool is_enable_coroutine() { if (is_task_worker()) { return task_enable_coroutine; } else if (is_manager()) { return false; } else { return enable_coroutine; } } bool is_hash_dispatch_mode() { return dispatch_mode == DISPATCH_FDMOD || dispatch_mode == DISPATCH_IPMOD || dispatch_mode == DISPATCH_CO_CONN_LB; } bool is_support_send_yield() { return is_hash_dispatch_mode(); } bool if_require_packet_callback(ListenPort *port, bool isset) { #ifdef SW_USE_OPENSSL return (port->is_dgram() && !port->ssl && !isset); #else return (port->is_dgram() && !isset); #endif } bool if_require_receive_callback(ListenPort *port, bool isset) { #ifdef SW_USE_OPENSSL return (((port->is_dgram() && port->ssl) || port->is_stream()) && !isset); #else return (port->is_stream() && !isset); #endif } Worker *get_worker(uint16_t worker_id) { // Event Worker if (worker_id < worker_num) { return &(gs->event_workers.workers[worker_id]); } // Task Worker uint32_t task_worker_max = task_worker_num + worker_num; if (worker_id < task_worker_max) { return &(gs->task_workers.workers[worker_id - worker_num]); } // User Worker uint32_t user_worker_max = task_worker_max + user_worker_list.size(); if (worker_id < user_worker_max) { return &(user_workers[worker_id - task_worker_max]); } return nullptr; } void stop_async_worker(Worker *worker); Pipe *get_pipe_object(int pipe_fd) { return (Pipe *) connection_list[pipe_fd].object; } size_t get_all_worker_num() { return worker_num + task_worker_num + get_user_worker_num(); } size_t get_user_worker_num() { return user_worker_list.size(); } ReactorThread *get_thread(int reactor_id) { return &reactor_threads[reactor_id]; } bool is_started() { return gs->start; } bool is_created() { return factory != nullptr; } bool is_master() { return SwooleG.process_type == SW_PROCESS_MASTER; } bool is_worker() { return SwooleG.process_type == SW_PROCESS_WORKER; } bool is_task_worker() { return SwooleG.process_type == SW_PROCESS_TASKWORKER; } bool is_manager() { return SwooleG.process_type == SW_PROCESS_MANAGER; } bool is_user_worker() { return SwooleG.process_type == SW_PROCESS_USERWORKER; } bool is_reactor_thread() { return SwooleG.process_type == SW_PROCESS_MASTER && SwooleTG.type == Server::THREAD_REACTOR; } bool isset_hook(enum HookType type) { assert(type <= HOOK_END); return hooks[type]; } bool is_sync_process() { if (is_manager()) { return true; } if (is_task_worker() && !task_enable_coroutine) { return true; } return false; } bool is_shutdown() { return gs->shutdown; } // can only be used in the main process bool is_valid_connection(Connection *conn) { return (conn && conn->socket && conn->active && conn->socket->fd_type == SW_FD_SESSION); } bool is_healthy_connection(double now, Connection *conn); static int is_dgram_event(uint8_t type) { switch (type) { case SW_SERVER_EVENT_RECV_DGRAM: return true; default: return false; } } static int is_stream_event(uint8_t type) { switch (type) { case SW_SERVER_EVENT_RECV_DATA: case SW_SERVER_EVENT_SEND_DATA: case SW_SERVER_EVENT_SEND_FILE: case SW_SERVER_EVENT_CONNECT: case SW_SERVER_EVENT_CLOSE: case SW_SERVER_EVENT_PAUSE_RECV: case SW_SERVER_EVENT_RESUME_RECV: case SW_SERVER_EVENT_BUFFER_FULL: case SW_SERVER_EVENT_BUFFER_EMPTY: return true; default: return false; } } int get_connection_fd(SessionId session_id) { return session_list[session_id % SW_SESSION_LIST_SIZE].fd; } Connection *get_connection_verify_no_ssl(SessionId session_id) { Session *session = get_session(session_id); int fd = session->fd; Connection *conn = get_connection(fd); if (!conn || conn->active == 0) { return nullptr; } if (session->id != session_id || conn->session_id != session_id) { return nullptr; } return conn; } Connection *get_connection_verify(SessionId session_id) { Connection *conn = get_connection_verify_no_ssl(session_id); #ifdef SW_USE_OPENSSL if (conn && conn->ssl && !conn->ssl_ready) { return nullptr; } #endif return conn; } Connection *get_connection(int fd) { if ((uint32_t) fd > max_connection) { return nullptr; } return &connection_list[fd]; } Connection *get_connection_for_iterator(int fd) { Connection *conn = get_connection(fd); if (conn && conn->active && !conn->closed) { #ifdef SW_USE_OPENSSL if (conn->ssl && !conn->ssl_ready) { return nullptr; } #endif return conn; } return nullptr; } Connection *get_connection_by_session_id(SessionId session_id) { return get_connection(get_connection_fd(session_id)); } Session *get_session(SessionId session_id) { return &session_list[session_id % SW_SESSION_LIST_SIZE]; } void lock() { lock_.lock(); } void unlock() { lock_.unlock(); } void clear_timer(); static void timer_callback(Timer *timer, TimerNode *tnode); int create_task_workers(); int create_user_workers(); int start_manager_process(); void call_hook(enum HookType type, void *arg); void call_worker_start_callback(Worker *worker); ResultCode call_command_handler(MessageBus &mb, uint16_t worker_id, network::Socket *sock); std::string call_command_handler_in_master(int command_id, const std::string &msg); void call_command_callback(int64_t request_id, const std::string &result); void foreach_connection(const std::function<void(Connection *)> &callback); static int accept_connection(Reactor *reactor, Event *event); #ifdef SW_SUPPORT_DTLS dtls::Session *accept_dtls_connection(ListenPort *ls, network::Address *sa); #endif static int accept_command_result(Reactor *reactor, Event *event); static int close_connection(Reactor *reactor, network::Socket *_socket); static int dispatch_task(const Protocol *proto, network::Socket *_socket, const RecvData *rdata); int send_to_connection(SendData *); ssize_t send_to_worker_from_worker(Worker *dst_worker, const void *buf, size_t len, int flags); ssize_t send_to_worker_from_worker(WorkerId id, EventData *data, int flags) { return send_to_worker_from_worker(get_worker(id), data, sizeof(data->info) + data->info.len, flags); } ssize_t send_to_reactor_thread(const EventData *ev_data, size_t sendn, SessionId session_id); int reply_task_result(const char *data, size_t data_len, int flags, EventData *current_task); bool send(SessionId session_id, const void *data, uint32_t length); bool sendfile(SessionId session_id, const char *file, uint32_t l_file, off_t offset, size_t length); bool sendwait(SessionId session_id, const void *data, uint32_t length); bool close(SessionId session_id, bool reset); bool notify(Connection *conn, enum ServerEventType event); bool feedback(Connection *conn, enum ServerEventType event); bool command(WorkerId process_id, Command::ProcessType process_type, const std::string &name, const std::string &msg, const Command::Callback &fn); void init_reactor(Reactor *reactor); void init_worker(Worker *worker); void init_task_workers(); void init_port_protocol(ListenPort *port); void init_signal_handler(); void init_ipc_max_size(); void set_max_connection(uint32_t _max_connection); void set_max_concurrency(uint32_t _max_concurrency) { if (_max_concurrency == 0) { _max_concurrency = UINT_MAX; } gs->max_concurrency = _max_concurrency; } void set_worker_max_concurrency(uint32_t _max_concurrency) { if (_max_concurrency == 0) { _max_concurrency = UINT_MAX; } worker_max_concurrency = _max_concurrency; } uint32_t get_max_connection() { return max_connection; } uint32_t get_max_concurrency() { return gs->max_concurrency; } uint32_t get_concurrency() { return gs->concurrency; } bool is_unavailable() { return get_concurrency() >= get_max_concurrency(); } uint32_t get_worker_max_concurrency() { return worker_max_concurrency; } void set_start_session_id(SessionId value) { if (value > UINT_MAX) { value = UINT_MAX; } gs->session_round = value; } int create_pipe_buffers(); void release_pipe_buffers(); void create_worker(Worker *worker); void destroy_worker(Worker *worker); void disable_accept(); void destroy_http_request(Connection *conn); int schedule_worker(int fd, SendData *data); size_t get_connection_num() { if (gs->connection_nums) { size_t num = 0; for (uint32_t i = 0; i < worker_num; i++) { num += gs->connection_nums[i]; } return num; } else { return gs->connection_num; } } /** * [Manager] */ pid_t spawn_event_worker(Worker *worker); pid_t spawn_user_worker(Worker *worker); pid_t spawn_task_worker(Worker *worker); void kill_user_workers(); void kill_event_workers(); void kill_task_workers(); static int wait_other_worker(ProcessPool *pool, const ExitStatus &exit_status); static void read_worker_message(ProcessPool *pool, EventData *msg); void drain_worker_pipe(); void check_worker_exit_status(Worker *worker, const ExitStatus &exit_status); /** * [Worker] */ void worker_start_callback(Worker *worker); void worker_stop_callback(Worker *worker); void worker_accept_event(DataHead *info); static void worker_signal_handler(int signo); static void worker_signal_init(void); static bool task_pack(EventData *task, const void *data, size_t data_len); static bool task_unpack(EventData *task, String *buffer, PacketPtr *packet); private: enum Mode mode_; Connection *connection_list = nullptr; Session *session_list = nullptr; ServerPortGS *port_gs_list = nullptr; /** * http static file directory */ std::string document_root; std::mutex lock_; uint32_t max_connection = 0; TimerNode *enable_accept_timer = nullptr; std::thread heartbeat_thread; /** * The number of pipe per reactor maintenance */ uint16_t reactor_pipe_num = 0; ReactorThread *reactor_threads = nullptr; int start_check(); void check_port_type(ListenPort *ls); void destroy(); void destroy_reactor_threads(); void destroy_reactor_processes(); int create_reactor_processes(); int create_reactor_threads(); int start_reactor_threads(); int start_reactor_processes(); int start_master_thread(); int start_event_worker(Worker *worker); void start_heartbeat_thread(); void join_reactor_thread(); TimerCallback get_timeout_callback(ListenPort *port, Reactor *reactor, Connection *conn); int get_lowest_load_worker_id() { uint32_t lowest_load_worker_id = 0; size_t min_coroutine = workers[0].coroutine_num; for (uint32_t i = 1; i < worker_num; i++) { if (workers[i].coroutine_num < min_coroutine) { min_coroutine = workers[i].coroutine_num; lowest_load_worker_id = i; continue; } } return lowest_load_worker_id; } int get_lowest_concurrent_worker_id() { uint32_t lowest_concurrent_worker_id = 0; size_t min_concurrency = workers[0].concurrency; for (uint32_t i = 1; i < worker_num; i++) { if (workers[i].concurrency < min_concurrency) { min_concurrency = workers[i].concurrency; lowest_concurrent_worker_id = i; continue; } } return lowest_concurrent_worker_id; } int get_idle_worker_id() { bool found = false; uint32_t key = 0; SW_LOOP_N(worker_num + 1) { key = sw_atomic_fetch_add(&worker_round_id, 1) % worker_num; if (workers[key].status == SW_WORKER_IDLE) { found = true; break; } } if (sw_unlikely(!found)) { scheduler_warning = true; } swoole_trace_log(SW_TRACE_SERVER, "schedule=%d, round=%d", key, worker_round_id); return key; } }; } // namespace swoole typedef swoole::Server swServer; typedef swoole::ListenPort swListenPort; typedef swoole::RecvData swRecvData; extern swoole::Server *g_server_instance; static inline swoole::Server *sw_server() { return g_server_instance; }
Close