Doguhan Ilter

Software Developer

System Design Notes

I will dive into the core concepts of system design, covering architecture patterns, scalability, and important considerations for building highly available and efficient systems.

1. Peer-2-Peer Architecture

A Peer-2-Peer (P2P) system is one where each node in the network can act both as a client and a server. In this decentralized architecture, peers communicate directly with one another without the need for a central server. Each peer has an equal responsibility in the network, and there's no central authority to control the entire system.

2. Master-Slave Architecture

Master-Slave is a more centralized approach compared to P2P. In this architecture, a single master node controls the system and directs tasks to multiple slave nodes. The master node has control over data or task flow, while the slave nodes are responsible for executing the tasks assigned by the master.

3. Single Point of Failure (SPOF)

A Single Point of Failure (SPOF) refers to a component in a system that, if it fails, causes the entire system to fail. SPOFs are a significant vulnerability in system design, and mitigating them is crucial to ensuring high availability and reliability. To avoid SPOFs, systems must incorporate redundancy and failover mechanisms that allow other components to take over in case of failure.

4. Scalability: Vertical vs Horizontal Scaling

Scalability refers to a system’s ability to handle increased load by adding resources. There are two major types of scalability: vertical and horizontal scaling.

Vertical Scaling

Vertical scaling (also known as "scaling up") involves adding more power to an existing server. This can include increasing the CPU, adding more RAM, or expanding storage capacity.

Horizontal Scaling

Horizontal scaling (or "scaling out") involves adding more machines or nodes to a system. It’s a more flexible approach to scalability, as you can continually add more servers to increase capacity without being limited by the performance of a single server.

5. Monitoring Tools

Monitoring tools are essential for tracking the health, performance, and availability of your system. By collecting metrics and logs, you can detect anomalies, bottlenecks, and potential failures before they impact the system.

6. Database Sharding

Database sharding is a method used to distribute data across multiple machines (or database instances). Instead of storing all data on one server, data is partitioned into smaller "shards," and each shard is stored on a different machine. This technique improves performance and scalability, especially for large datasets.

7. Rate Limiter

A rate limiter is a tool that controls the number of requests a user can make to a system within a specific time frame. Rate limiting helps prevent abuse and ensures fair usage by limiting the number of requests any single user or client can make in a given period.

8. CAP Theorem

The CAP theorem states that in any distributed data store, you can achieve at most two of the following three properties:

9. Microservices Architecture

Microservices architecture is a style where a large application is divided into smaller, loosely coupled services that communicate over a network. Each service is responsible for a specific business functionality, and they can be developed, deployed, and scaled independently.

10. Event-Driven Architecture

Event-driven architecture (EDA) is a design paradigm where events (state changes or updates) trigger system actions. Components react to events by processing them asynchronously, which allows for decoupling of producers and consumers of events.

11. Load Balancing Algorithms

Load balancing refers to the practice of distributing incoming network traffic across multiple servers to ensure no single server becomes overwhelmed. Different algorithms can be used to determine how traffic should be distributed.

12. Distributed Caching

Distributed caching is a technique where cached data is spread across multiple servers, enabling higher performance and scalability compared to traditional single-server caching. It helps reduce the load on databases by serving frequently accessed data from the cache.

13. Service Mesh

A service mesh is a dedicated infrastructure layer that helps manage and secure microservices communication. It abstracts the complexity of managing services and provides features like load balancing, service discovery, and secure communication between services.

14. Containerization & Orchestration

Containerization involves packaging applications and their dependencies into containers, allowing them to run consistently across different environments. Orchestration tools like Kubernetes manage and automate the deployment, scaling, and operation of containers in large-scale environments.

15. CAP Theorem & Consistency Models

While the CAP theorem (Consistency, Availability, Partition Tolerance) describes the trade-offs in distributed systems, understanding different consistency models is crucial for backend developers. Different databases or systems offer varying levels of consistency guarantees, which affect the design of the system.

16. Distributed Tracing

Distributed tracing is a technique used to track requests as they flow through various microservices in a system. It helps in debugging performance issues and monitoring system health by visualizing the path of a request.

17. Synchronous vs Asynchronous Communication

In distributed systems, communication between components can be either synchronous or asynchronous. Understanding the difference is crucial for designing systems that perform well under various conditions.

Systems must prioritize two of these guarantees, as it is impossible to provide all three simultaneously. Understanding this trade-off is essential when designing distributed systems.