InterviewStack.io LogoInterviewStack.io

Backend Engineering & Performance Topics

Backend system optimization, performance tuning, memory management, and engineering proficiency. Covers system-level performance, remote support tools, and infrastructure optimization.

Performance Fundamentals and Troubleshooting

Core skills for identifying, diagnosing, and resolving general performance problems across applications and systems. Topics include establishing baselines and metrics, using monitoring and profiling tools to determine whether issues are CPU bound, memory bound, input output bound, or network bound, and applying systematic troubleshooting workflows. Candidates should be able to prioritize fixes, recommend temporary mitigations and long term solutions, and explain when to escalate to specialists. This canonical topic covers general performance awareness, common diagnostic tools, and basic remediation approaches for slow systems and resource exhaustion.

46 questions

Server Side Asynchronous Programming

Asynchronous and concurrent programming as applied to backend systems, including event loop models, thread pools, futures and promises, asynchronous I O, streaming, and reactive frameworks. Covers Node dot js event loop and streaming APIs, Java threading models and reactive libraries such as Project Reactor or RxJava, Python asyncio and multiprocessing versus multithreading trade offs, handling blocking operations, backpressure and flow control, and patterns to structure scalable non blocking servers. Candidates should demonstrate the ability to reason about throughput, latency, resource contention, and appropriate concurrency models for server workloads.

40 questions

Performance Profiling and Optimization

Comprehensive skills and methodology for profiling, diagnosing, and optimizing runtime performance across services, applications, and platforms. Involves measuring baseline performance using monitoring and profiling tools, capturing central processing unit, memory, input output, and network metrics, and interpreting flame graphs and execution traces to find hotspots. Requires a reproducible measure first approach to isolate root causes, distinguish central processing unit time from graphical processing unit time, and separate application bottlenecks from system level issues. Covers platform specific profilers and techniques such as frame time budgeting for interactive applications, synthetic benchmarks and production trace replay, and instrumentation with metrics, logs, and distributed traces. Candidates should be familiar with common root causes including lock contention, garbage collection pauses, disk saturation, cache misses, and inefficient algorithms, and be able to prioritize changes by expected impact. Optimization techniques included are algorithmic improvements, parallelization and concurrency control, memory management and allocation strategies, caching and batching, hardware acceleration, and focused micro optimizations. Also includes validating improvements through before and after measurements, regression and degradation analysis, reasoning about trade offs between performance, maintainability, and complexity, and creating reproducible profiling hooks and tests.

56 questions

System Resource Management and Monitoring

Monitor and manage operating system and hardware level resources to ensure application performance and stability. Topics include central processing unit utilization and context switching, system load trends, memory usage including heap and stack behavior, paging and swapping effects, disk input output operations and free space, and network bandwidth utilization and packet loss. Know diagnostic tools and commands for observing these signals, recognize patterns of resource contention and exhaustion such as out of memory and high input output wait, and understand mitigation techniques including tuning, resource limits, throttling, caching, capacity planning, and vertical or horizontal scaling.

42 questions

Performance Optimization and Latency Engineering

Covers systematic approaches to measuring and improving system performance and latency at architecture and code levels. Topics include profiling and tracing to find where time is actually spent, forming and testing hypotheses, optimizing critical paths, and validating improvements with measurable metrics. Candidates should be able to distinguish central processing unit bound work from input output bound work, analyze latency versus throughput trade offs, evaluate where caching and content delivery networks help or hurt, recognize database and network constraints, and propose strategies such as query optimization, asynchronous processing patterns, resource pooling, and load balancing. Also includes performance testing methodologies, reasoning about trade offs and risks, and describing end to end optimisation projects and their business impact.

40 questions

Performance Debugging and Latency Investigation

Finding the root cause of latency spikes: checking CPU/memory/disk/network utilization, profiling applications, querying slow logs, and identifying bottlenecks. Understanding the difference between resource exhaustion and an algorithmic problem. Using monitoring and tracing tools to narrow down where time is spent.

42 questions

Caching and Performance Optimization

Covers design and implementation of multi layer caching and end to end performance strategies for web and backend systems. Topics include client side techniques such as browser caching, service worker strategies, code splitting, and lazy loading for components images and data; edge and distribution techniques such as content delivery network design and caching of static assets; and server side and data layer caching using in memory stores such as Redis and Memcached, query result caching, and database caching patterns. Includes cache invalidation and coherence strategies such as time to live, least recently used eviction, cache aside, write through and write behind, and prevention of cache stampedes. Covers when to introduce caching and when not to, performance and consistency trade offs, connection pooling, monitoring and metrics, establishing performance budgets, and operational considerations such as cache warm up and invalidation during deploys. Also addresses higher level concerns including search engine optimization implications and server side rendering trade offs, and how performance decisions map to user experience and business metrics at senior levels.

46 questions

Performance Engineering and Cost Optimization

Engineering practices and trade offs for meeting performance objectives while controlling operational cost. Topics include setting latency and throughput targets and latency budgets; benchmarking profiling and tuning across application database and infrastructure layers; memory compute serialization and batching optimizations; asynchronous processing and workload shaping; capacity estimation and right sizing for compute and storage to reduce cost; understanding cost drivers in cloud environments including network egress and storage tiering; trade offs between real time and batch processing; and monitoring to detect and prevent performance regressions. Candidates should describe measurement driven approaches to optimization and be able to justify trade offs between cost complexity and user experience.

44 questions

Advanced Linux Performance and Services

Advanced administration focused on service lifecycle, process management, and system performance. Topics include deep systemd service management and unit file authoring, dependency ordering and service recovery, process lifecycle and signal handling, cgroups and resource controls, tuning kernel parameters, diagnosing CPU and memory pressure, understanding page cache and swap behavior, out of memory scenarios, I O performance analysis, interpreting load average, and using performance and sampling tools such as top, htop, pidstat, iostat, vmstat, sar, and perf for identifying bottlenecks and implementing mitigations.

46 questions
Page 1/2