diff --git a/blog/Spark-Architecture/index.md b/blog/Spark-Architecture/index.md
deleted file mode 100644
index 92338b86..00000000
--- a/blog/Spark-Architecture/index.md
+++ /dev/null
@@ -1,309 +0,0 @@
----
-title: "Spark Architecture Explained"
-authors: [Aditya-Singh-Rathore]
-tags: [Apache Spark, Spark Architecture, Big Data, Distributed Computing, Data Engineering]
-date: 2025-08-18
-
-description: Apache Spark is a fast, open-source big data framework that leverages in-memory computing for high performance. Its architecture powers scalable distributed processing across clusters, making it essential for analytics and machine learning.
-
-draft: false
-canonical_url:
-# meta:
-# - name: "robots"
-# content: "index, follow"
-# - property: "og:title"
-# content: "What is Google DeepMind AI?"
-# - property: "og:description"
-# content: "DeepMind is an auxiliary of Google that centers around man-made brainpower. All the more explicitly, it utilizes a part of AI called AI"
-# - property: "og:type"
-# content: "article"
-# - property: "og:url"
-# content: "/blog/getting-started-with-mern"
-# - property: "og:image"
-# content: "/assets/images/mern-8a27add30515e58f789f89a4c9072818.jpg"
-# - name: "twitter:card"
-# content: "summary_large_image"
-# - name: "twitter:title"
-# content: "A Comprehensive Guide to Get You Started with MERN Stack"
-# - name: "twitter:description"
-# content: "DeepMind is an auxiliary of Google that centers around man-made brainpower. All the more explicitly, it utilizes a part of AI called AI"
-# - name: "twitter:image"
-# content: "assets/images/mern-8a27add30515e58f789f89a4c9072818.jpg"
-
----
-
-# Understanding Apache Spark Architecture: A Deep Dive into Distributed Computing
-
-Hey there, fellow data enthusiasts! 👋
-
-I remember the first time I encountered a Spark architecture diagram. It looked like a complex web of boxes and arrows that seemed to communicate in some secret distributed computing language. But once I understood what each component actually does and how they work together, everything clicked into place.
-
-Today, I want to walk you through Spark's architecture in a way that I wish someone had explained it to me back then - focusing on the core components and how this beautiful system actually works under the hood.
-
-## What is Apache Spark?
-
-Before diving into the architecture, let's establish what we're dealing with. Apache Spark is an open-source, distributed computing framework designed to process massive datasets across clusters of computers. Think of it as a coordinator that can take your data processing job and intelligently distribute it across multiple machines to get the work done faster.
-
-The key insight that makes Spark special? It keeps data in memory between operations whenever possible, which is why it can be dramatically faster than traditional batch processing systems.
-
-## The Big Picture: High-Level Architecture
-
-
-
-
-When you look at Spark's architecture, you're essentially looking at a well-orchestrated system with three main types of components working together:
-
-1. **Driver Program** - The mastermind that coordinates everything
-2. **Cluster Manager** - The resource allocator
-3. **Executors** - The workers that do the actual processing
-
-Let's break down each of these and understand how they collaborate.
-
-## Core Components Deep Dive
-
-### 1. The Driver Program: Your Application's Brain
-
-The Driver Program is where your Spark application begins and ends. When you write a Spark program and run it, you're essentially creating a driver program. Here's what makes it the brain of the operation:
-
-**What the Driver Does:**
-- Contains your main() function and defines RDDs and operations on them
-- Converts your high-level operations into a DAG (Directed Acyclic Graph) of tasks
-- Schedules tasks across the cluster
-- Coordinates with the cluster manager to get resources
-- Collects results from executors and returns final results
-
-**Think of it this way:** If your Spark application were a restaurant, the Driver would be the head chef who takes orders (your code), breaks them down into specific cooking tasks, assigns those tasks to kitchen staff (executors), and ensures everything comes together for the final dish.
-
-The driver runs in its own JVM process and maintains all the metadata about your Spark application throughout its lifetime.
-
-### 2. Cluster Manager: The Resource Referee
-
-The Cluster Manager sits between your driver and the actual compute resources. Its job is to allocate and manage resources across the cluster. Spark is flexible and works with several cluster managers:
-
-**Standalone Cluster Manager:**
-- Spark's built-in cluster manager
-- Simple to set up and understand
-- Great for dedicated Spark clusters
-
-**Apache YARN (Yet Another Resource Negotiator):**
-- Hadoop's resource manager
-- Perfect if you're in a Hadoop ecosystem
-- Allows resource sharing between Spark and other Hadoop applications
-
-**Apache Mesos:**
-- A general-purpose cluster manager
-- Can handle multiple frameworks beyond just Spark
-- Good for mixed workload environments
-
-**Kubernetes:**
-- The modern container orchestration platform
-- Increasingly popular for new deployments
-- Excellent for cloud-native environments
-
-**The key point:** The cluster manager's job is resource allocation - it doesn't care what your application does, just how much CPU and memory it needs.
-
-### 3. Executors: The Workhorses
-
-Executors are the processes that actually run your tasks and store data for your application. Each executor runs in its own JVM process and can run multiple tasks concurrently using threads.
-
-**What Executors Do:**
-- Execute tasks sent from the driver
-- Store computation results in memory or disk storage
-- Provide in-memory storage for cached RDDs/DataFrames
-- Report heartbeat and task status back to the driver
-
-**Key Characteristics:**
-- Each executor has a fixed number of cores and amount of memory
-- Executors are launched at the start of a Spark application and run for the entire lifetime
-- If an executor fails, Spark can launch new ones and recompute lost data
-
-Think of executors as skilled workers in our restaurant analogy - they can handle multiple cooking tasks simultaneously and have their own workspace (memory) to store ingredients and intermediate results.
-
-## How These Components Work Together: The Execution Flow
-
-Now that we know the players, let's see how they orchestrate a typical Spark application:
-
-### Step 1: Application Submission
-When you submit a Spark application, the driver program starts up and contacts the cluster manager requesting resources for executors.
-
-### Step 2: Resource Allocation
-The cluster manager examines available resources and launches executor processes on worker nodes across the cluster.
-
-### Step 3: Task Planning
-The driver analyzes your code and creates a logical execution plan. It breaks down operations into stages and tasks that can be executed in parallel.
-
-### Step 4: Task Distribution
-The driver sends tasks to executors. Each task operates on a partition of data, and multiple tasks can run in parallel across different executors.
-
-### Step 5: Execution and Communication
-Executors run the tasks, storing intermediate results and communicating progress back to the driver. The driver coordinates everything and handles any failures.
-
-### Step 6: Result Collection
-Once all tasks complete, the driver collects results and returns the final output to your application.
-
-## Understanding RDDs: The Foundation
-
-At the heart of Spark's architecture lies the concept of Resilient Distributed Datasets (RDDs). Understanding RDDs is crucial to understanding how Spark actually works.
-
-**What makes RDDs special:**
-
-**Resilient:** RDDs can automatically recover from node failures. Spark remembers how each RDD was created (its lineage) and can rebuild lost partitions.
-
-**Distributed:** RDD data is automatically partitioned and distributed across multiple nodes in the cluster.
-
-**Dataset:** At the end of the day, it's still just a collection of your data - but with superpowers.
-
-### RDD Operations: Transformations vs Actions
-
-RDDs support two types of operations, and understanding the difference is crucial:
-
-**Transformations** (Lazy):
-```scala
-val filtered = data.filter(x => x > 10)
-val mapped = filtered.map(x => x * 2)
-val grouped = mapped.groupByKey()
-```
-These operations don't actually execute immediately. Spark just builds up a computation graph.
-
-**Actions** (Eager):
-```scala
-val results = grouped.collect() // Brings data to driver
-val count = filtered.count() // Returns number of elements
-grouped.saveAsTextFile("hdfs://...") // Saves to storage
-```
-Actions trigger the actual execution of all the transformations in the lineage.
-
-This lazy evaluation allows Spark to optimize the entire computation pipeline before executing anything.
-
-## The DAG: Spark's Optimization Engine
-
-One of Spark's most elegant features is how it converts your operations into a Directed Acyclic Graph (DAG) for optimal execution.
-
-### How DAG Optimization Works
-
-When you chain multiple transformations together, Spark doesn't execute them immediately. Instead, it builds a DAG that represents the computation. This allows for powerful optimizations:
-
-**Pipelining:** Multiple transformations that don't require data shuffling can be combined into a single stage and executed together.
-
-**Stage Boundaries:** Spark creates stage boundaries at operations that require data shuffling (like `groupByKey`, `join`, or `repartition`).
-
-### Stages and Tasks Breakdown
-
-**Stage:** A set of tasks that can all be executed without data shuffling. All tasks in a stage can run in parallel.
-
-**Task:** The smallest unit of work in Spark. Each task processes one partition of data.
-
-**Wide vs Narrow Dependencies:**
-- **Narrow Dependencies:** Each partition of child RDD depends on a constant number of parent partitions (like `map`, `filter`)
-- **Wide Dependencies:** Each partition of child RDD may depend on multiple parent partitions (like `groupByKey`, `join`)
-
-Wide dependencies create stage boundaries because they require shuffling data across the network.
-
-## Memory Management: Where the Magic Happens
-
-Spark's memory management is what gives it the speed advantage over traditional batch processing systems. Here's how it works:
-
-### Memory Regions
-
-Spark divides executor memory into several regions:
-
-**Storage Memory (60% by default):**
-- Used for caching RDDs/DataFrames
-- LRU eviction when space is needed
-- Can borrow from execution memory when available
-
-**Execution Memory (20% by default):**
-- Used for computation in shuffles, joins, sorts, aggregations
-- Can borrow from storage memory when needed
-
-**User Memory (20% by default):**
-- For user data structures and internal metadata
-- Not managed by Spark
-
-**Reserved Memory (300MB by default):**
-- System reserved memory for Spark's internal objects
-
-The beautiful thing about this system is that storage and execution memory can dynamically borrow from each other based on current needs.
-
-## The Unified Stack: Multiple APIs, One Engine
-
-What makes Spark truly powerful is that it provides multiple high-level APIs that all run on the same core engine:
-
-### Spark Core
-The foundation that provides:
-- Basic I/O functionality
-- Task scheduling and memory management
-- Fault tolerance
-- RDD abstraction
-
-### Spark SQL
-- SQL queries on structured data
-- DataFrame and Dataset APIs
-- Catalyst query optimizer
-- Integration with various data sources
-
-### Spark Streaming
-- Real-time stream processing
-- Micro-batch processing model
-- Integration with streaming sources like Kafka
-
-### MLlib
-- Distributed machine learning algorithms
-- Feature transformation utilities
-- Model evaluation and tuning
-
-### GraphX
-- Graph processing and analysis
-- Built-in graph algorithms
-- Graph-parallel computation
-
-The key insight: all of these APIs compile down to the same core RDD operations, so they all benefit from Spark's optimization engine and can interoperate seamlessly.
-
-## Putting It All Together
-
-Now that we've covered all the components, let's see how they work together in a real example:
-
-```scala
-// This creates RDDs but doesn't execute anything yet
-val textFile = spark.textFile("hdfs://large-file.txt")
-val words = textFile.flatMap(line => line.split(" "))
-val wordCounts = words.map(word => (word, 1))
-val aggregated = wordCounts.reduceByKey(_ + _)
-
-// This action triggers execution of the entire pipeline
-val results = aggregated.collect()
-```
-
-**What happens behind the scenes:**
-1. Driver creates a DAG with two stages (split by the `reduceByKey` shuffle)
-2. Driver requests executors from cluster manager
-3. Stage 1 tasks (read, flatMap, map) execute on partitions across executors
-4. Data gets shuffled for the `reduceByKey` operation
-5. Stage 2 tasks perform the aggregation
-6. Results get collected back to the driver
-
-## Why This Architecture Matters
-
-Understanding Spark's architecture isn't just academic knowledge - it's the key to working effectively with big data:
-
-**Fault Tolerance:** The RDD lineage graph means Spark can recompute lost data automatically without manual intervention.
-
-**Scalability:** The driver/executor model scales horizontally - just add more worker nodes to handle bigger datasets.
-
-**Efficiency:** Lazy evaluation and DAG optimization mean Spark can optimize entire computation pipelines before executing anything.
-
-**Flexibility:** The unified stack means you can mix SQL, streaming, and machine learning in the same application without data movement penalties.
-
-## Conclusion: The Beauty of Distributed Computing
-
-Spark's architecture represents one of the most elegant solutions to distributed computing that I've encountered. By clearly separating concerns - coordination (driver), resource management (cluster manager), and execution (executors) - Spark creates a system that's both powerful and understandable.
-
-The magic isn't in any single component, but in how they all work together. The driver's intelligence in creating optimal execution plans, the cluster manager's efficiency in resource allocation, and the executors' reliability in task execution combine to create something greater than the sum of its parts.
-
-Whether you're processing terabytes of log data, training machine learning models, or running real-time analytics, understanding this architecture will help you reason about performance, debug issues, and design better data processing solutions.
-
----
-
-*The next time you see a Spark architecture diagram, I hope you'll see what I see now - not a confusing web of boxes and arrows, but an elegant dance of distributed computing components working in perfect harmony. Happy Sparking! 🚀*
-
-
{content}
+ + {/* Card Meta */} +