aio: Regularize IO worker internal naming.
authorThomas Munro <tmunro@postgresql.org>
Sat, 12 Jul 2025 01:43:27 +0000 (13:43 +1200)
committerThomas Munro <tmunro@postgresql.org>
Sat, 12 Jul 2025 02:44:09 +0000 (14:44 +1200)
Adopt PgAioXXX convention for pgaio module type names.  Rename a
function that didn't use a pgaio_worker_ submodule prefix.  Rename the
internal submit function's arguments to match the indirectly relevant
function pointer declaration and nearby examples.  Rename the array of
handle IDs in PgAioSubmissionQueue to sqes, a term of art seen in the
systems it emulates, also clarifying that they're not IO handle
pointers as the old name might imply.

No change in behavior, just type, variable and function name cleanup.

Back-patch to 18.

Discussion: https://postgr.es/m/CA%2BhUKG%2BwbaZZ9Nwc_bTopm4f-7vDmCwLk80uKDHj9mq%2BUp0E%2Bg%40mail.gmail.com

src/backend/storage/aio/method_worker.c
src/tools/pgindent/typedefs.list

index 6454987e9f5b92d4ca676e2638aded0eb5dda596..1ce1bcd57a00a0bc0f6bc2f9075bf902b63315a9 100644 (file)
 #define IO_WORKER_WAKEUP_FANOUT 2
 
 
-typedef struct AioWorkerSubmissionQueue
+typedef struct PgAioWorkerSubmissionQueue
 {
    uint32      size;
    uint32      mask;
    uint32      head;
    uint32      tail;
-   uint32      ios[FLEXIBLE_ARRAY_MEMBER];
-} AioWorkerSubmissionQueue;
+   uint32      sqes[FLEXIBLE_ARRAY_MEMBER];
+} PgAioWorkerSubmissionQueue;
 
-typedef struct AioWorkerSlot
+typedef struct PgAioWorkerSlot
 {
    Latch      *latch;
    bool        in_use;
-} AioWorkerSlot;
+} PgAioWorkerSlot;
 
-typedef struct AioWorkerControl
+typedef struct PgAioWorkerControl
 {
    uint64      idle_worker_mask;
-   AioWorkerSlot workers[FLEXIBLE_ARRAY_MEMBER];
-} AioWorkerControl;
+   PgAioWorkerSlot workers[FLEXIBLE_ARRAY_MEMBER];
+} PgAioWorkerControl;
 
 
 static size_t pgaio_worker_shmem_size(void);
@@ -96,8 +96,8 @@ int           io_workers = 3;
 
 static int io_worker_queue_size = 64;
 static int MyIoWorkerId;
-static AioWorkerSubmissionQueue *io_worker_submission_queue;
-static AioWorkerControl *io_worker_control;
+static PgAioWorkerSubmissionQueue *io_worker_submission_queue;
+static PgAioWorkerControl *io_worker_control;
 
 
 static size_t
@@ -106,15 +106,15 @@ pgaio_worker_queue_shmem_size(int *queue_size)
    /* Round size up to next power of two so we can make a mask. */
    *queue_size = pg_nextpower2_32(io_worker_queue_size);
 
-   return offsetof(AioWorkerSubmissionQueue, ios) +
+   return offsetof(PgAioWorkerSubmissionQueue, sqes) +
        sizeof(uint32) * *queue_size;
 }
 
 static size_t
 pgaio_worker_control_shmem_size(void)
 {
-   return offsetof(AioWorkerControl, workers) +
-       sizeof(AioWorkerSlot) * MAX_IO_WORKERS;
+   return offsetof(PgAioWorkerControl, workers) +
+       sizeof(PgAioWorkerSlot) * MAX_IO_WORKERS;
 }
 
 static size_t
@@ -162,7 +162,7 @@ pgaio_worker_shmem_init(bool first_time)
 }
 
 static int
-pgaio_choose_idle_worker(void)
+pgaio_worker_choose_idle(void)
 {
    int         worker;
 
@@ -180,7 +180,7 @@ pgaio_choose_idle_worker(void)
 static bool
 pgaio_worker_submission_queue_insert(PgAioHandle *ioh)
 {
-   AioWorkerSubmissionQueue *queue;
+   PgAioWorkerSubmissionQueue *queue;
    uint32      new_head;
 
    queue = io_worker_submission_queue;
@@ -192,7 +192,7 @@ pgaio_worker_submission_queue_insert(PgAioHandle *ioh)
        return false;           /* full */
    }
 
-   queue->ios[queue->head] = pgaio_io_get_id(ioh);
+   queue->sqes[queue->head] = pgaio_io_get_id(ioh);
    queue->head = new_head;
 
    return true;
@@ -201,14 +201,14 @@ pgaio_worker_submission_queue_insert(PgAioHandle *ioh)
 static uint32
 pgaio_worker_submission_queue_consume(void)
 {
-   AioWorkerSubmissionQueue *queue;
+   PgAioWorkerSubmissionQueue *queue;
    uint32      result;
 
    queue = io_worker_submission_queue;
    if (queue->tail == queue->head)
        return UINT32_MAX;      /* empty */
 
-   result = queue->ios[queue->tail];
+   result = queue->sqes[queue->tail];
    queue->tail = (queue->tail + 1) & (queue->size - 1);
 
    return result;
@@ -241,37 +241,37 @@ pgaio_worker_needs_synchronous_execution(PgAioHandle *ioh)
 }
 
 static void
-pgaio_worker_submit_internal(int nios, PgAioHandle *ios[])
+pgaio_worker_submit_internal(int num_staged_ios, PgAioHandle **staged_ios)
 {
    PgAioHandle *synchronous_ios[PGAIO_SUBMIT_BATCH_SIZE];
    int         nsync = 0;
    Latch      *wakeup = NULL;
    int         worker;
 
-   Assert(nios <= PGAIO_SUBMIT_BATCH_SIZE);
+   Assert(num_staged_ios <= PGAIO_SUBMIT_BATCH_SIZE);
 
    LWLockAcquire(AioWorkerSubmissionQueueLock, LW_EXCLUSIVE);
-   for (int i = 0; i < nios; ++i)
+   for (int i = 0; i < num_staged_ios; ++i)
    {
-       Assert(!pgaio_worker_needs_synchronous_execution(ios[i]));
-       if (!pgaio_worker_submission_queue_insert(ios[i]))
+       Assert(!pgaio_worker_needs_synchronous_execution(staged_ios[i]));
+       if (!pgaio_worker_submission_queue_insert(staged_ios[i]))
        {
            /*
             * We'll do it synchronously, but only after we've sent as many as
             * we can to workers, to maximize concurrency.
             */
-           synchronous_ios[nsync++] = ios[i];
+           synchronous_ios[nsync++] = staged_ios[i];
            continue;
        }
 
        if (wakeup == NULL)
        {
            /* Choose an idle worker to wake up if we haven't already. */
-           worker = pgaio_choose_idle_worker();
+           worker = pgaio_worker_choose_idle();
            if (worker >= 0)
                wakeup = io_worker_control->workers[worker].latch;
 
-           pgaio_debug_io(DEBUG4, ios[i],
+           pgaio_debug_io(DEBUG4, staged_ios[i],
                           "choosing worker %d",
                           worker);
        }
@@ -490,7 +490,7 @@ IoWorkerMain(const void *startup_data, size_t startup_data_len)
                           IO_WORKER_WAKEUP_FANOUT);
            for (int i = 0; i < nwakeups; ++i)
            {
-               if ((worker = pgaio_choose_idle_worker()) < 0)
+               if ((worker = pgaio_worker_choose_idle()) < 0)
                    break;
                latches[nlatches++] = io_worker_control->workers[worker].latch;
            }
index 8319203857175d127535ddee5c827d306acbe13d..ff050e93a50770446567afa9be05605dbd15a18a 100644 (file)
@@ -55,9 +55,6 @@ AggStrategy
 AggTransInfo
 Aggref
 AggregateInstrumentation
-AioWorkerControl
-AioWorkerSlot
-AioWorkerSubmissionQueue
 AlenState
 Alias
 AllocBlock
@@ -2184,6 +2181,9 @@ PgAioTargetInfo
 PgAioUringCaps
 PgAioUringContext
 PgAioWaitRef
+PgAioWorkerControl
+PgAioWorkerSlot
+PgAioWorkerSubmissionQueue
 PgArchData
 PgBackendGSSStatus
 PgBackendSSLStatus