Skip to content

rtomac/dotnetcore-socket-server

Repository files navigation

Introduction

This is an exercise to experiment with raw TCP socket comms, using .NET Core, striving to optimize throughput. Just for fun.

In some areas of the implementation there are CLR constructs (async/await) and BCL classes that could make this simpler, but I'm striving to do all the thread management, parallelization, and synchronization of resources manually. Also just for fun.

The project contains a server app and a client app, which behave as follows:

  • Server accepts input from up to five concurrent clients (port 4000 by default).
  • Server expects lines of input containing exactly nine decimal digits (zero-padded if necessary) and terminated with a server-native newline sequence.
  • Server writes any unique numbers it receives to a log file.
  • Server writes a status report to the console every 10s.
  • If client sends a line of input with a "terminate" command, server will close all connections and shut itself down.

Building

Get .NET Core SDK with .NET 1.1 runtime

On Windows:

choco install -y dotnetcore-sdk

Clone repo

git clone https://github.com/rtomac/dotnetcore-socket-server.git

Build

cd dotnetcore-socket-server
build.bat

build.bat simply calls dotnet restore, dotnet build, and dotnet test.

Running

There are two command-line applications here:

Server

Fire it up, and it will start listening for connections on port 4000.

run-server.bat

run-server.bat simply calls dotnet run from the server app/project dir.

For command-line usage of the server app, run:

cd Server
dotnet run -- -?

Client

When started, it will connect to the server on port 4000 and start blasting data at it.

run-client.bat

run-client.bat simply calls dotnet run from the client app/project dir.

Again, for command-line usage, run:

cd Client
dotnet run -- -?

The code

Here is a brief overview of the code/class design (in the server app):

  • Program: Entry point for the server app. Handles command-line parsing, etc., but delegates the real work to the Application class.
  • Application: Class that orchestrates and controls the lifetime of the different components in the application, including the socket listener, the log file writer, and the status reporter.
  • LocalhostSocketListener: Binds to localhost:4000 (or specified port), listens for incoming connections on a background thread, and dispatches new threads to handle each of them.
  • SocketStreamReader: Reads data from a network stream over a socket connection, parses and processes it, and hands the good stuff off to other components for further processing (i.e. writing to log file).
  • QueueingLogWriter: Writes de-duped values transmitted to the server into a log file. Does this by managing an in-memory queue of values to be written, and processing that queue on a background worker thread, to isolate the file I/O and prevent it from blocking the other threads that are processing data from their network connections.
  • StatusReporter: Keeps track of aggregate statistics and periodically (on a background thread) writes a status report to stdout.

The source code has a bit more documentation on how all of these classes are doing their job. Check it out.

Key optimizations

  • Multithreading and parallelization: All distinct components in the application are running on separate threads. On multi-core systems, that means true parallelization.
  • Non-blocking interactions: Interactions between components on different threads are made to be as lightweight as possible to ensure that work in one component on one thread can never block work in another component on another thread. In some cases that means asynchronous interaction (e.g. in-memory queue) and in others it just means minimizing the amount of time that access to shared resources are being synchronzied. In no cases is work on one thread held up by I/O on another thread.
  • Binary data processing: All data processing is on raw binary data. No reliance on string conversion and/or parsing. (Both a performance and memory optimization.)

Further optimizations to try

A note on some other things to try to improve throughput and resiliency:

  • Read bigger blocks of data in from network stream at a single time: Right now, we're reading in 10- or 11-byte blocks of data at a time (known chunk size of a single value). There may be overhead associated with each read--it may be faster to read in 1k of data at a time, for instance, and send that through processing.
  • Parallelize reading of data from network stream and data processing/validation: It would be theoretically possible to read raw data from the network stream and shove it on a queue for a processing worker to pick up and analyze/process. Could improve throughput. (Creates a new challenge, which is messaging back to I/O thread to terminate connection when invalid data is detected.)

About

An experiment with .NET Core and raw TCP socket comms.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published