#ifndef MBED_EXTENDED_TESTS #error [NOT_SUPPORTED] Parallel pressure tests are not supported by default #endif #include "mbed.h" #include MBED_CONF_APP_HEADER_FILE #include "TCPSocket.h" #include "greentea-client/test_env.h" #include "unity/unity.h" #include "utest.h" using namespace utest::v1; #ifndef MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MIN #define MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MIN 64 #endif #ifndef MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MAX #define MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MAX 0x80000 #endif #ifndef MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_SEED #define MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_SEED 0x6d626564 #endif #ifndef MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS #define MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS 3 #endif #ifndef MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_DEBUG #define MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_DEBUG false #endif #define STRINGIZE(x) STRINGIZE2(x) #define STRINGIZE2(x) #x // Simple xorshift pseudorandom number generator class RandSeq { private: uint32_t x; uint32_t y; static const int A = 15; static const int B = 18; static const int C = 11; public: RandSeq(uint32_t seed=MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_SEED) : x(seed), y(seed) {} uint32_t next(void) { x ^= x << A; x ^= x >> B; x ^= y ^ (y >> C); return x + y; } void skip(size_t size) { for (size_t i = 0; i < size; i++) { next(); } } void buffer(uint8_t *buffer, size_t size) { RandSeq lookahead = *this; for (size_t i = 0; i < size; i++) { buffer[i] = lookahead.next() & 0xff; } } int cmp(uint8_t *buffer, size_t size) { RandSeq lookahead = *this; for (size_t i = 0; i < size; i++) { int diff = buffer[i] - (lookahead.next() & 0xff); if (diff != 0) { return diff; } } return 0; } }; // Tries to get the biggest buffer possible on the device. Exponentially // grows a buffer until heap runs out of space, and uses half to leave // space for the rest of the program void generate_buffer(uint8_t **buffer, size_t *size, size_t min, size_t max) { size_t i = min; while (i < max) { void *b = malloc(i); if (!b) { i /= 4; if (i < min) { i = min; } break; } free(b); i *= 2; } *buffer = (uint8_t *)malloc(i); *size = i; TEST_ASSERT(buffer); } // Global variables shared between pressure tests NetworkInterface* net; SocketAddress tcp_addr; Timer timer; Mutex iomutex; // Single instance of a pressure test class PressureTest { private: uint8_t *buffer; size_t buffer_size; TCPSocket sock; Thread thread; public: PressureTest(uint8_t *buffer, size_t buffer_size) : buffer(buffer), buffer_size(buffer_size) { } void start() { osStatus status = thread.start(callback(this, &PressureTest::run)); TEST_ASSERT_EQUAL(osOK, status); } void join() { osStatus status = thread.join(); TEST_ASSERT_EQUAL(osOK, status); } void run() { // Tests exponentially growing sequences for (size_t size = MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MIN; size < MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MAX; size *= 2) { int err = sock.open(net); TEST_ASSERT_EQUAL(0, err); err = sock.connect(tcp_addr); TEST_ASSERT_EQUAL(0, err); sock.recv(buffer, sizeof(MBED_CONF_APP_TCP_ECHO_PREFIX)); iomutex.lock(); printf("TCP: %s:%d streaming %d bytes\r\n", tcp_addr.get_ip_address(), tcp_addr.get_port(), size); iomutex.unlock(); sock.set_blocking(false); // Loop to send/recv all data RandSeq tx_seq; RandSeq rx_seq; size_t rx_count = 0; size_t tx_count = 0; size_t window = buffer_size; while (tx_count < size || rx_count < size) { // Send out data if (tx_count < size) { size_t chunk_size = size - tx_count; if (chunk_size > window) { chunk_size = window; } tx_seq.buffer(buffer, chunk_size); int td = sock.send(buffer, chunk_size); if (td > 0) { if (MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_DEBUG) { iomutex.lock(); printf("TCP: tx -> %d\r\n", td); iomutex.unlock(); } tx_seq.skip(td); tx_count += td; } else if (td != NSAPI_ERROR_WOULD_BLOCK) { // We may fail to send because of buffering issues, // cut buffer in half if (window > MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MIN) { window /= 2; } if (MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_DEBUG) { iomutex.lock(); printf("TCP: Not sent (%d), window = %d\r\n", td, window); iomutex.unlock(); } } } // Verify recieved data while (rx_count < size) { int rd = sock.recv(buffer, buffer_size); TEST_ASSERT(rd > 0 || rd == NSAPI_ERROR_WOULD_BLOCK); if (rd > 0) { if (MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_DEBUG) { iomutex.lock(); printf("TCP: rx <- %d\r\n", rd); iomutex.unlock(); } int diff = rx_seq.cmp(buffer, rd); TEST_ASSERT_EQUAL(0, diff); rx_seq.skip(rd); rx_count += rd; } else if (rd == NSAPI_ERROR_WOULD_BLOCK) { break; } } } err = sock.close(); TEST_ASSERT_EQUAL(0, err); } } }; PressureTest *pressure_tests[MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS]; void test_tcp_packet_pressure_parallel() { uint8_t *buffer; size_t buffer_size; generate_buffer(&buffer, &buffer_size, MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MIN, MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MAX); size_t buffer_subsize = buffer_size / MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS; printf("MBED: Generated buffer %d\r\n", buffer_size); printf("MBED: Split into %d buffers %d\r\n", MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS, buffer_subsize); net = MBED_CONF_APP_OBJECT_CONSTRUCTION; int err = MBED_CONF_APP_CONNECT_STATEMENT; TEST_ASSERT_EQUAL(0, err); printf("MBED: TCPClient IP address is '%s'\n", net->get_ip_address()); tcp_addr.set_ip_address(MBED_CONF_APP_ECHO_SERVER_ADDR); tcp_addr.set_port(MBED_CONF_APP_ECHO_SERVER_PORT); timer.start(); // Startup pressure tests in parallel for (int i = 0; i < MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS; i++) { pressure_tests[i] = new PressureTest(&buffer[i*buffer_subsize], buffer_subsize); pressure_tests[i]->start(); } for (int i = 0; i < MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS; i++) { pressure_tests[i]->join(); delete pressure_tests[i]; } timer.stop(); printf("MBED: Time taken: %fs\r\n", timer.read()); printf("MBED: Speed: %.3fkb/s\r\n", MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_THREADS* 8*(2*MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MAX - MBED_CFG_TCP_CLIENT_PACKET_PRESSURE_MIN) / (1000*timer.read())); net->disconnect(); } // Test setup utest::v1::status_t test_setup(const size_t number_of_cases) { GREENTEA_SETUP(120, "tcp_echo"); return verbose_test_setup_handler(number_of_cases); } Case cases[] = { Case("TCP packet pressure parallel", test_tcp_packet_pressure_parallel), }; Specification specification(test_setup, cases); int main() { return !Harness::run(specification); }