Handle-with-cache.c | 2024 |

UserProfile* get_user_profile_handle(int user_id) { pthread_mutex_lock(&cache_lock); // Check cache CacheEntry *entry = g_hash_table_lookup(handle_cache, &user_id); if (entry) { // Cache hit entry->ref_count++; entry->last_access = time(NULL); pthread_mutex_unlock(&cache_lock); printf("Cache hit for user %d\n", user_id); return entry->profile; }

// Cache entry wrapper typedef struct { UserProfile *profile; time_t last_access; unsigned int ref_count; // Reference counting for safety } CacheEntry; handle-with-cache.c

A common optimization is or using a per-key mutex: Repeatedly opening, reading, or computing the same resource

The module handle-with-cache.c exemplifies a classic design pattern: the . A "handle" is an opaque pointer or identifier to a resource, and the cache stores recently accessed handles to avoid redundant initialization or I/O operations. a network socket

pthread_mutex_lock(&cache_lock); // Double-check: another thread might have inserted it while we were loading entry = g_hash_table_lookup(handle_cache, &user_id); if (entry) { // Discard our loaded profile and use the cached one free_user_profile(profile); entry->ref_count++; pthread_mutex_unlock(&cache_lock); return entry->profile; }

In systems programming, efficiency is paramount. Repeatedly opening, reading, or computing the same resource (a file, a network socket, a database row, or a complex calculation result) is wasteful. This is where caching becomes indispensable.