• About Us
  • Disclaimer
  • Contact Us
  • Privacy Policy
Wednesday, April 29, 2026
mGrowTech
No Result
View All Result
  • Technology And Software
    • Account Based Marketing
    • Channel Marketing
    • Marketing Automation
      • Al, Analytics and Automation
      • Ad Management
  • Digital Marketing
    • Social Media Management
    • Google Marketing
  • Direct Marketing
    • Brand Management
    • Marketing Attribution and Consulting
  • Mobile Marketing
  • Event Management
  • PR Solutions
  • Technology And Software
    • Account Based Marketing
    • Channel Marketing
    • Marketing Automation
      • Al, Analytics and Automation
      • Ad Management
  • Digital Marketing
    • Social Media Management
    • Google Marketing
  • Direct Marketing
    • Brand Management
    • Marketing Attribution and Consulting
  • Mobile Marketing
  • Event Management
  • PR Solutions
No Result
View All Result
mGrowTech
No Result
View All Result
Home Al, Analytics and Automation

Mixture of Experts Architecture in Transformer Models

Josh by Josh
July 20, 2025
in Al, Analytics and Automation
0
Mixture of Experts Architecture in Transformer Models


import torch

import torch.nn as nn

import torch.nn.functional as F

 

class Expert(nn.Module):

    def __init__(self, dim, intermediate_dim):

        super().__init__()

        self.gate_proj = nn.Linear(dim, intermediate_dim)

        self.up_proj = nn.Linear(dim, intermediate_dim)

        self.down_proj = nn.Linear(intermediate_dim, dim)

        self.act = nn.SiLU()

 

    def forward(self, x):

        gate = self.gate_proj(x)

        up = self.up_proj(x)

        swish = self.act(gate)

        output = self.down_proj(swish * up)

        return output

 

class MoELayer(nn.Module):

    def __init__(self, dim, intermediate_dim, num_experts, top_k=2):

        super().__init__()

        self.num_experts = num_experts

        self.top_k = top_k

        self.dim = dim

        # Create expert networks

        self.experts = nn.ModuleList([

            Expert(dim, intermediate_dim) for _ in range(num_experts)

        ])

        self.router = nn.Linear(dim, num_experts)

 

    def forward(self, hidden_states):

        batch_size, seq_len, hidden_dim = hidden_states.shape

        

        # Reshape for expert processing, the compute routing probabilities

        hidden_states_reshaped = hidden_states.view(–1, hidden_dim)

        router_logits = self.router(hidden_states_reshaped)  # (batch_size * seq_len, num_experts)

        routing_probs = F.softmax(router_logits, dim=–1)

 

        # Select top-k experts, and scale the probabilities to sum to 1

        # output shape: (batch_size * seq_len, k)

        top_k_probs, top_k_indices = torch.topk(routing_probs, self.top_k, dim=–1)

        top_k_probs = top_k_probs / top_k_probs.sum(dim=–1, keepdim=True)

 

        # Process through selected experts

        output = []

        for i in range(self.top_k):

            expert_idx = top_k_indices[:, i]

            expert_probs = top_k_probs[:, i]

            # Process each vector in the batch and sequence with the selected expert

            expert_output = torch.stack([

                self.experts[j](hidden_states_reshaped[j])

                for j in expert_idx

            ], dim=0)

            # Weighted sum by routing probability

            output.sum(expert_probs.unsqueeze(–1) * expert_output)

 

        # Reshape back to original shape

        output = sum(output).view(batch_size, seq_len, hidden_dim)

        return output

 

class MoETransformerLayer(nn.Module):

    def __init__(self, dim, intermediate_dim, num_experts, top_k=2, num_heads=8):

        super().__init__()

        self.attention = nn.MultiheadAttention(dim, num_heads, batch_first=True)

        self.moe = MoELayer(dim, intermediate_dim, num_experts, top_k)

        self.norm1 = nn.RMSNorm(dim)

        self.norm2 = nn.RMSNorm(dim)

 

    def forward(self, x):

        # Attention sublayer

        input_x = x

        x = self.norm1(x)

        attn_output, _ = self.attention(x, x, x)

        input_x = input_x + attn_output

 

        # MoE sublayer

        x = self.norm2(input_x)

        moe_output = self.moe(x)

        return input_x + moe_output



Source_link

READ ALSO

The MIT-IBM Computing Research Lab launches to shape the future of AI and quantum computing | MIT News

Meta FAIR Releases NeuralSet: A Python Package for Neuro-AI That Supports fMRI, M/EEG, Spikes, and HuggingFace Embeddings

Related Posts

The MIT-IBM Computing Research Lab launches to shape the future of AI and quantum computing | MIT News
Al, Analytics and Automation

The MIT-IBM Computing Research Lab launches to shape the future of AI and quantum computing | MIT News

April 29, 2026
Meta FAIR Releases NeuralSet: A Python Package for Neuro-AI That Supports fMRI, M/EEG, Spikes, and HuggingFace Embeddings
Al, Analytics and Automation

Meta FAIR Releases NeuralSet: A Python Package for Neuro-AI That Supports fMRI, M/EEG, Spikes, and HuggingFace Embeddings

April 29, 2026
Enabling privacy-preserving AI training on everyday devices | MIT News
Al, Analytics and Automation

Enabling privacy-preserving AI training on everyday devices | MIT News

April 29, 2026
OpenAI Releases Privacy Filter: A 1.5B-Parameter Open-Source PII Redaction Model with 50M Active Parameters
Al, Analytics and Automation

OpenAI Releases Privacy Filter: A 1.5B-Parameter Open-Source PII Redaction Model with 50M Active Parameters

April 29, 2026
Top 10 Physical AI Models Powering Real-World Robots in 2026
Al, Analytics and Automation

Top 10 Physical AI Models Powering Real-World Robots in 2026

April 28, 2026
Build a Reinforcement Learning Powered Agent that Learns to Retrieve Relevant Long-Term Memories for Accurate LLM Question Answering
Al, Analytics and Automation

Build a Reinforcement Learning Powered Agent that Learns to Retrieve Relevant Long-Term Memories for Accurate LLM Question Answering

April 28, 2026
Next Post
At Least 750 US Hospitals Faced Disruptions During Last Year’s CrowdStrike Outage, Study Finds

At Least 750 US Hospitals Faced Disruptions During Last Year’s CrowdStrike Outage, Study Finds

POPULAR NEWS

Trump ends trade talks with Canada over a digital services tax

Trump ends trade talks with Canada over a digital services tax

June 28, 2025
Communication Effectiveness Skills For Business Leaders

Communication Effectiveness Skills For Business Leaders

June 10, 2025
15 Trending Songs on TikTok in 2025 (+ How to Use Them)

15 Trending Songs on TikTok in 2025 (+ How to Use Them)

June 18, 2025
App Development Cost in Singapore: Pricing Breakdown & Insights

App Development Cost in Singapore: Pricing Breakdown & Insights

June 22, 2025
Comparing the Top 7 Large Language Models LLMs/Systems for Coding in 2025

Comparing the Top 7 Large Language Models LLMs/Systems for Coding in 2025

November 4, 2025

EDITOR'S PICK

2025 guide for all major networks

2025 guide for all major networks

August 27, 2025
AOL’s dial-up internet still exists, but not for much longer

AOL’s dial-up internet still exists, but not for much longer

August 11, 2025
How to turn your old iPad into a digital picture frame

How to turn your old iPad into a digital picture frame

June 5, 2025
How To Launch, Grow, and Scale a Community That Supports Your Brand [MozCon 2025 Speaker Series]

How To Launch, Grow, and Scale a Community That Supports Your Brand [MozCon 2025 Speaker Series]

June 3, 2025

About

We bring you the best Premium WordPress Themes that perfect for news, magazine, personal blog, etc. Check our landing page for details.

Follow us

Categories

  • Account Based Marketing
  • Ad Management
  • Al, Analytics and Automation
  • Brand Management
  • Channel Marketing
  • Digital Marketing
  • Direct Marketing
  • Event Management
  • Google Marketing
  • Marketing Attribution and Consulting
  • Marketing Automation
  • Mobile Marketing
  • PR Solutions
  • Social Media Management
  • Technology And Software
  • Uncategorized

Recent Posts

  • Moburst’s Monthly Marketing Roundup #30
  • BrandSmart 2026: How to Market Better, not faster, in the Age of Acceleration
  • The retrieval rebuild: Why hybrid retrieval intent tripled as enterprise RAG programs hit the scale wall
  • The MIT-IBM Computing Research Lab launches to shape the future of AI and quantum computing | MIT News
  • About Us
  • Disclaimer
  • Contact Us
  • Privacy Policy
No Result
View All Result
  • Technology And Software
    • Account Based Marketing
    • Channel Marketing
    • Marketing Automation
      • Al, Analytics and Automation
      • Ad Management
  • Digital Marketing
    • Social Media Management
    • Google Marketing
  • Direct Marketing
    • Brand Management
    • Marketing Attribution and Consulting
  • Mobile Marketing
  • Event Management
  • PR Solutions